From edb1c92eda535c1d5d7cfa6e449b4cba64e263c1 Mon Sep 17 00:00:00 2001 From: Hongji Zhu Date: Wed, 17 Apr 2024 21:57:48 +0800 Subject: [PATCH] Update README_en.md --- README_en.md | 24 +++++++++++++++++++++++- 1 file changed, 23 insertions(+), 1 deletion(-) diff --git a/README_en.md b/README_en.md index 2571fca..e79aaeb 100644 --- a/README_en.md +++ b/README_en.md @@ -43,6 +43,7 @@ - [Multi-turn Conversation](#multi-turn-conversation) - [Inference on Mac](#inference-on-mac) - [Deployment on Mobile Phone](#deployment-on-mobile-phone) + - [WebUI Demo](#webui-demo) - [TODO](#todo) - [Citation](#citation) @@ -594,10 +595,31 @@ PYTORCH_ENABLE_MPS_FALLBACK=1 python test.py ### Deployment on Mobile Phone Currently MiniCPM-V 2.0 can be deployed on mobile phones with Android and Harmony operating systems. 🚀 Try it out [here](https://github.com/OpenBMB/mlc-MiniCPM). +### WebUI Demo + +
+Click to see how to deploy WebUI demo on different devices + +```shell +pip install -r requirements.txt +``` + +```shell +# For Nvidia GPUs support BF16 (like A100, H100, RTX3090), run: +python web_demo.py --device cuda --dtype bf16 + +# For Nvidia GPUs do NOT support BF16 (like V100, T4, RTX2080), run: +python web_demo.py --device cuda --dtype fp16 + +# For Mac with MPS (Apple silicon or AMD GPUs), run: +PYTORCH_ENABLE_MPS_FALLBACK=1 python web_demo.py --device mps --dtype fp16 +``` +
+ + ## TODO - [ ] Fine-tuning support -- [ ] Local Web-UI deployment - [ ] Code release for real-time interactive assistant ## Model License