mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-04 09:49:20 +08:00
update readme
This commit is contained in:
@@ -25,7 +25,7 @@
|
||||
|
||||
## News <!-- omit in toc -->
|
||||
|
||||
* [2024.05.23] We've released a comprehensive comparison between Phi-3-vision-128k-instruct and MiniCPM-Llama3-V 2.5, including benchmarks evaluations, and multilingual capabilities 🌟📊🌍. Click [here](#evaluation) to view more details.
|
||||
* [2024.05.23] We've released a comprehensive comparison between Phi-3-vision-128k-instruct and MiniCPM-Llama3-V 2.5, including benchmarks evaluations, and multilingual capabilities 🌟📊🌍. Click [here](./docs/compare_with_phi-3_vision.md) to view more details.
|
||||
* [2024.05.20] We open-soure MiniCPM-Llama3-V 2.5, it has improved OCR capability and supports 30+ languages, representing the first edge-side MLLM achieving GPT-4V level performance! We provide [efficient inference](#deployment-on-mobile-phone) and [simple fine-tuning](./finetune/readme.md). Try it now!
|
||||
* [2024.04.23] MiniCPM-V-2.0 supports vLLM now! Click [here](#vllm) to view more details.
|
||||
* [2024.04.18] We create a HuggingFace Space to host the demo of MiniCPM-V 2.0 at [here](https://huggingface.co/spaces/openbmb/MiniCPM-V-2)!
|
||||
|
||||
@@ -28,7 +28,7 @@
|
||||
|
||||
## 更新日志 <!-- omit in toc -->
|
||||
|
||||
* [2024.05.23] 我们添加了Phi-3-vision-128k-instruct与MiniCPM-Llama3-V 2.5的全面对比,包括基准测试评估和多语言能力 🌟📊🌍。点击[这里](#性能评估)查看详细信息。
|
||||
* [2024.05.23] 我们添加了Phi-3-vision-128k-instruct与MiniCPM-Llama3-V 2.5的全面对比,包括基准测试评估和多语言能力 🌟📊🌍。点击[这里](./docs/compare_with_phi-3_vision.md)查看详细信息。
|
||||
<!-- * [2024.05.22] 我们进一步提升了端侧推理速度!实现了 6-8 tokens/s 的流畅体验,欢迎试用! -->
|
||||
* [2024.05.20] 我们开源了 MiniCPM-Llama3-V 2.5,增强了 OCR 能力,支持 30 多种语言,并首次在端侧实现了 GPT-4V 级的多模态能力!我们提供了[高效推理](#手机端部署)和[简易微调](./finetune/readme.md)的支持,欢迎试用!
|
||||
* [2024.04.23] 我们增加了对 [vLLM](#vllm) 的支持,欢迎体验!
|
||||
|
||||
@@ -1,8 +1,8 @@
|
||||
## Phi-3-vision-128K-Instruct vs MiniCPM-Llama3-V 2.5
|
||||
|
||||
Comparison results of Phi-3-vision-128K-Instruct and MiniCPM-Llama3-V 2.5, regarding the model size, hardware requirements, and performances on multiple popular benchmarks.
|
||||
Comparison results of Phi-3-vision-128K-Instruct and MiniCPM-Llama3-V 2.5, regarding the model size, hardware requirements, and performances.
|
||||
|
||||
我们提供了从模型参数、硬件需求、全面性能指标等方面对比 Phi-3-vision-128K-Instruct 和 MiniCPM-Llama3-V 2.5 的结果。
|
||||
我们提供了从模型参数、硬件需求、性能指标等方面对比 Phi-3-vision-128K-Instruct 和 MiniCPM-Llama3-V 2.5 的结果。
|
||||
|
||||
## Hardeware Requirements (硬件需求)
|
||||
|
||||
@@ -18,7 +18,9 @@ With in4 quantization, MiniCPM-Llama3-V 2.5 delivers smooth inference with only
|
||||
|
||||
## Model Size and Peformance (模型参数和性能)
|
||||
|
||||
In most benchmarks, MiniCPM-Llama3-V 2.5 achieves **better performance** compared with Phi-3-vision-128K-Instruct.
|
||||
|
||||
在大多数评测集上, MiniCPM-Llama3-V 2.5 相比于 Phi-3-vision-128K-Instruct 都展现出了**更优的性能表现**.
|
||||
|
||||
| | Phi-3-vision-128K-Instruct | MiniCPM-Llama3-V 2.5|
|
||||
|:-|:----------:|:-------------------:|
|
||||
@@ -28,4 +30,20 @@ With in4 quantization, MiniCPM-Llama3-V 2.5 delivers smooth inference with only
|
||||
| RealworldQA | 58.8 | **63.5**|
|
||||
| TextVQA | 72.2 | **76.6** |
|
||||
| ScienceQA| **90.8** | 89.0 |
|
||||
| POPE | 83.4 | **87.2** |
|
||||
| POPE | 83.4 | **87.2** |
|
||||
|
||||
|
||||
## Multilingual Capabilities
|
||||
|
||||
|
||||
MiniCPM-Llama3-V 2.5 exhibits **stronger multilingual** capabilities compared with Phi-3-vision-128K-Instruct on LLaVA Bench.
|
||||
|
||||
MiniCPM-Llama3-V 2.5 在对话和推理评测榜单 LLaVA Bench 上展现出了比 Phi-3-vision-128K-Instruct **更强的多语言的性能**。
|
||||
|
||||
<div align="center">
|
||||
<img src="../assets/llavabench_compare_3.png" width="85%" />
|
||||
<br>
|
||||
Evaluation results of LLaVABench in multiple languages
|
||||
<br>
|
||||
多语言LLaVA Bench评测结果
|
||||
</div>
|
||||
|
||||
Reference in New Issue
Block a user