mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-04 17:59:18 +08:00
Update README, vllm officially supports MiniCPM-V serials
This commit is contained in:
10
README_en.md
10
README_en.md
@@ -614,24 +614,22 @@ MiniCPM-Llama3-V 2.5 can run with llama.cpp now! See our fork of [llama.cpp](htt
|
|||||||
### Inference with vLLM<a id="vllm"></a>
|
### Inference with vLLM<a id="vllm"></a>
|
||||||
|
|
||||||
<details>
|
<details>
|
||||||
<summary>Click to see how to inference MiniCPM-V 2.0 and MiniCPM-Llama3-V 2.5 with vLLM </summary>
|
<summary> vLLM now officially supports MiniCPM-V 2.0 and MiniCPM-Llama3-V 2.5, Click to see. </summary>
|
||||||
Because our pull request to vLLM is still waiting for reviewing, we fork this repository to build and test our vLLM demo. Here are the steps:
|
|
||||||
|
|
||||||
1. Clone our version of vLLM:
|
1. Clone the official vLLM:
|
||||||
```shell
|
```shell
|
||||||
git clone https://github.com/OpenBMB/vllm.git
|
git clone https://github.com/vllm-project/vllm.git
|
||||||
```
|
```
|
||||||
2. Install vLLM:
|
2. Install vLLM:
|
||||||
```shell
|
```shell
|
||||||
cd vllm
|
cd vllm
|
||||||
git checkout minicpmv
|
|
||||||
pip install -e .
|
pip install -e .
|
||||||
```
|
```
|
||||||
3. Install timm:
|
3. Install timm:
|
||||||
```shell
|
```shell
|
||||||
pip install timm==0.9.10
|
pip install timm==0.9.10
|
||||||
```
|
```
|
||||||
4. Run our demo:
|
4. Run the example:
|
||||||
```shell
|
```shell
|
||||||
python examples/minicpmv_example.py
|
python examples/minicpmv_example.py
|
||||||
```
|
```
|
||||||
|
|||||||
@@ -644,17 +644,15 @@ MiniCPM-Llama3-V 2.5 现在支持llama.cpp啦! 用法请参考我们的fork [lla
|
|||||||
|
|
||||||
### vLLM 部署 <a id='vllm'></a>
|
### vLLM 部署 <a id='vllm'></a>
|
||||||
<details>
|
<details>
|
||||||
<summary>点击查看 MiniCPM-V 2.0 和 MiniCPM-Llama3-V 2.5 利用vLLM 部署运行的方法</summary>
|
<summary>点击查看, vLLM 现已官方支持MiniCPM-V 2.0 和 MiniCPM-Llama3-V 2.5</summary>
|
||||||
由于我们对 vLLM 提交的 PR 还在 review 中,因此目前我们 fork 了一个 vLLM 仓库以供测试使用。
|
|
||||||
|
|
||||||
1. 首先克隆我们 fork 的 vLLM 库:
|
1. 首先克隆官方的 vLLM 库:
|
||||||
```shell
|
```shell
|
||||||
git clone https://github.com/OpenBMB/vllm.git
|
git clone https://github.com/vllm-project/vllm.git
|
||||||
```
|
```
|
||||||
2. 安装 vLLM 库:
|
2. 安装 vLLM 库:
|
||||||
```shell
|
```shell
|
||||||
cd vllm
|
cd vllm
|
||||||
git checkout minicpmv
|
|
||||||
pip install -e .
|
pip install -e .
|
||||||
```
|
```
|
||||||
3. 安装 timm 库:
|
3. 安装 timm 库:
|
||||||
|
|||||||
Reference in New Issue
Block a user