mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-04 09:49:20 +08:00
Update README
This commit is contained in:
@@ -106,7 +106,6 @@
|
||||
- [Multimodal Live Streaming](#multimodal-live-streaming)
|
||||
- [Inference on Multiple GPUs](#inference-on-multiple-gpus)
|
||||
- [Inference on Mac](#inference-on-mac)
|
||||
- [Deployment on Mobile Phone](#deployment-on-mobile-phone)
|
||||
- [Efficient Inference with llama.cpp, ollama, vLLM](#efficient-inference-with-llamacpp-ollama-vllm)
|
||||
- [Fine-tuning](#fine-tuning)
|
||||
- [FAQs](#faqs)
|
||||
@@ -2372,8 +2371,6 @@ PYTORCH_ENABLE_MPS_FALLBACK=1 python test.py
|
||||
```
|
||||
</details>
|
||||
|
||||
### Deployment on Mobile Phone
|
||||
MiniCPM-V 2.0 can be deployed on mobile phones with Android operating systems. 🚀 Click [MiniCPM-V 2.0](https://github.com/OpenBMB/mlc-MiniCPM) to install apk.
|
||||
|
||||
### Efficient Inference with llama.cpp, ollama, vLLM
|
||||
|
||||
|
||||
@@ -90,7 +90,6 @@
|
||||
- [多模态流式交互](#多模态流式交互)
|
||||
- [多卡推理](#多卡推理)
|
||||
- [Mac 推理](#mac-推理)
|
||||
- [手机端部署](#手机端部署)
|
||||
- [基于 llama.cpp、ollama、vLLM 的高效推理](#基于-llamacppollamavllm-的高效推理)
|
||||
- [微调](#微调)
|
||||
- [FAQs](#faqs)
|
||||
@@ -2353,10 +2352,6 @@ PYTORCH_ENABLE_MPS_FALLBACK=1 python test.py
|
||||
</details>
|
||||
|
||||
|
||||
### 手机端部署
|
||||
MiniCPM-V 2.0 可运行在Android手机上,点击[MiniCPM-V 2.0](https://github.com/OpenBMB/mlc-MiniCPM)安装apk使用;
|
||||
|
||||
|
||||
### 基于 llama.cpp、ollama、vLLM 的高效推理
|
||||
|
||||
llama.cpp 用法请参考[我们的fork llama.cpp](https://github.com/OpenBMB/llama.cpp/tree/minicpmv-main/examples/llava/README-minicpmv2.6.md), 在iPad上可以支持 16~18 token/s 的流畅推理(测试环境:iPad Pro + M4)。
|
||||
|
||||
@@ -292,3 +292,8 @@ We deploy MiniCPM-V 2.0 on end devices. The demo video is the raw screen recordi
|
||||
|:-----------|:--:|:-----------:|:-------------------|:---------------:|
|
||||
| MiniCPM-V 2.0 | GPU | 8 GB | Light version, balance the performance the computation cost. | [🤗](https://huggingface.co/openbmb/MiniCPM-V-2) [<img src="../assets/modelscope_logo.png" width="20px"></img>](https://modelscope.cn/models/OpenBMB/MiniCPM-V-2) |
|
||||
| MiniCPM-V 1.0 | GPU | 7 GB | Lightest version, achieving the fastest inference. | [🤗](https://huggingface.co/openbmb/MiniCPM-V) [<img src="../assets/modelscope_logo.png" width="20px"></img>](https://modelscope.cn/models/OpenBMB/MiniCPM-V) |
|
||||
|
||||
|
||||
### Deployment on Mobile Phone
|
||||
|
||||
MiniCPM-V 2.0 can be deployed on mobile phones with Android operating systems. 🚀 Click [MiniCPM-V 2.0](https://github.com/OpenBMB/mlc-MiniCPM) to install apk.
|
||||
Reference in New Issue
Block a user