Update README.md

This commit is contained in:
Hongji Zhu
2024-07-25 15:55:42 +08:00
committed by GitHub
parent efae873a22
commit 80ac0eb1aa

View File

@@ -614,24 +614,22 @@ MiniCPM-Llama3-V 2.5 can run with llama.cpp now! See our fork of [llama.cpp](htt
### Inference with vLLM<a id="vllm"></a> ### Inference with vLLM<a id="vllm"></a>
<details> <details>
<summary>Click to see how to inference MiniCPM-V 2.0 and MiniCPM-Llama3-V 2.5 with vLLM </summary> <summary> vLLM now officially supports MiniCPM-V 2.0 and MiniCPM-Llama3-V 2.5, Click to see. </summary>
Because our pull request to vLLM is still waiting for reviewing, we fork this repository to build and test our vLLM demo. Here are the steps:
1. Clone our version of vLLM: 1. Clone the official vLLM:
```shell ```shell
git clone https://github.com/OpenBMB/vllm.git git clone https://github.com/vllm-project/vllm.git
``` ```
2. Install vLLM: 2. Install vLLM:
```shell ```shell
cd vllm cd vllm
git checkout minicpmv
pip install -e . pip install -e .
``` ```
3. Install timm: 3. Install timm:
```shell ```shell
pip install timm==0.9.10 pip install timm==0.9.10
``` ```
4. Run our demo: 4. Run the example:
```shell ```shell
python examples/minicpmv_example.py python examples/minicpmv_example.py
``` ```