mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-05 18:29:18 +08:00
@@ -614,7 +614,7 @@ MiniCPM-Llama3-V 2.5 can run with llama.cpp now! See our fork of [llama.cpp](htt
|
|||||||
### Inference with vLLM<a id="vllm"></a>
|
### Inference with vLLM<a id="vllm"></a>
|
||||||
|
|
||||||
<details>
|
<details>
|
||||||
<summary>Click to see how to inference MiniCPM-V 2.0 with vLLM (MiniCPM-Llama3-V 2.5 coming soon) </summary>
|
<summary>Click to see how to inference MiniCPM-V 2.0 and MiniCPM-Llama3-V 2.5 with vLLM </summary>
|
||||||
Because our pull request to vLLM is still waiting for reviewing, we fork this repository to build and test our vLLM demo. Here are the steps:
|
Because our pull request to vLLM is still waiting for reviewing, we fork this repository to build and test our vLLM demo. Here are the steps:
|
||||||
|
|
||||||
1. Clone our version of vLLM:
|
1. Clone our version of vLLM:
|
||||||
@@ -624,6 +624,7 @@ git clone https://github.com/OpenBMB/vllm.git
|
|||||||
2. Install vLLM:
|
2. Install vLLM:
|
||||||
```shell
|
```shell
|
||||||
cd vllm
|
cd vllm
|
||||||
|
git checkout minicpmv
|
||||||
pip install -e .
|
pip install -e .
|
||||||
```
|
```
|
||||||
3. Install timm:
|
3. Install timm:
|
||||||
|
|||||||
Reference in New Issue
Block a user