Update readme, support vLLM
Browse files
README.md
CHANGED
|
@@ -14,13 +14,11 @@ datasets:
|
|
| 14 |
|
| 15 |
## News <!-- omit in toc -->
|
| 16 |
|
|
|
|
| 17 |
* [2024.04.18] We create a HuggingFace Space to host the demo of MiniCPM-V 2.0 at [here](https://huggingface.co/spaces/openbmb/MiniCPM-V-2)!
|
| 18 |
-
* [2024.04.17] MiniCPM-V
|
| 19 |
-
* [2024.04.15] MiniCPM-V
|
| 20 |
* [2024.04.12] We open-source MiniCPM-V-2.0, which achieves comparable performance with Gemini Pro in understanding scene text and outperforms strong Qwen-VL-Chat 9.6B and Yi-VL 34B on <a href="https://rank.opencompass.org.cn/leaderboard-multimodal">OpenCompass</a>, a comprehensive evaluation over 11 popular benchmarks. Click <a href="https://openbmb.vercel.app/minicpm-v-2">here</a> to view the MiniCPM-V 2.0 technical blog.
|
| 21 |
-
* [2024.03.14] MiniCPM-V now supports [fine-tuning](https://github.com/modelscope/swift/blob/main/docs/source/Multi-Modal/minicpm-v最佳实践.md) with the SWIFT framework. Thanks to [Jintao](https://github.com/Jintao-Huang) for the contribution!
|
| 22 |
-
* [2024.03.01] MiniCPM-V now can be deployed on Mac!
|
| 23 |
-
* [2024.02.01] We open-source MiniCPM-V and OmniLMM-12B, which support efficient end-side deployment and powerful multimodal capabilities correspondingly.
|
| 24 |
|
| 25 |
## MiniCPM-V 2.0
|
| 26 |
|
|
@@ -86,6 +84,31 @@ Click here to try out the Demo of [MiniCPM-V 2.0](http://120.92.209.146:80).
|
|
| 86 |
## Deployment on Mobile Phone
|
| 87 |
MiniCPM-V 2.0 can be deployed on mobile phones with Android and Harmony operating systems. 🚀 Try it out [here](https://github.com/OpenBMB/mlc-MiniCPM).
|
| 88 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 89 |
|
| 90 |
## Usage
|
| 91 |
Inference using Huggingface transformers on Nivdia GPUs or Mac with MPS (Apple silicon or AMD GPUs). Requirements tested on python 3.10:
|
|
|
|
| 14 |
|
| 15 |
## News <!-- omit in toc -->
|
| 16 |
|
| 17 |
+
* [2024.04.23] MiniCPM-V 2.0 supports [vLLM](#vllm) now!
|
| 18 |
* [2024.04.18] We create a HuggingFace Space to host the demo of MiniCPM-V 2.0 at [here](https://huggingface.co/spaces/openbmb/MiniCPM-V-2)!
|
| 19 |
+
* [2024.04.17] MiniCPM-V 2.0 supports deploying [WebUI Demo](https://github.com/OpenBMB/MiniCPM-V/blob/8a1f766b85595a8095651eed9a44a83a965b305b/README_en.md#minicpm-v-) now!
|
| 20 |
+
* [2024.04.15] MiniCPM-V 2.0 supports [fine-tuning](https://github.com/modelscope/swift/blob/main/docs/source/Multi-Modal/minicpm-v-2最佳实践.md) with the SWIFT framework!
|
| 21 |
* [2024.04.12] We open-source MiniCPM-V-2.0, which achieves comparable performance with Gemini Pro in understanding scene text and outperforms strong Qwen-VL-Chat 9.6B and Yi-VL 34B on <a href="https://rank.opencompass.org.cn/leaderboard-multimodal">OpenCompass</a>, a comprehensive evaluation over 11 popular benchmarks. Click <a href="https://openbmb.vercel.app/minicpm-v-2">here</a> to view the MiniCPM-V 2.0 technical blog.
|
|
|
|
|
|
|
|
|
|
| 22 |
|
| 23 |
## MiniCPM-V 2.0
|
| 24 |
|
|
|
|
| 84 |
## Deployment on Mobile Phone
|
| 85 |
MiniCPM-V 2.0 can be deployed on mobile phones with Android and Harmony operating systems. 🚀 Try it out [here](https://github.com/OpenBMB/mlc-MiniCPM).
|
| 86 |
|
| 87 |
+
## Inference with vLLM<a id="vllm"></a>
|
| 88 |
+
|
| 89 |
+
<details>
|
| 90 |
+
<summary>Click to see how to inference with vLLM </summary>
|
| 91 |
+
Because our pull request to vLLM is still waiting for reviewing, we fork this repository to build and test our vLLM demo. Here are the steps:
|
| 92 |
+
|
| 93 |
+
1. Clone our version of vLLM:
|
| 94 |
+
```shell
|
| 95 |
+
git clone https://github.com/OpenBMB/vllm.git
|
| 96 |
+
```
|
| 97 |
+
2. Install vLLM:
|
| 98 |
+
```shell
|
| 99 |
+
cd vllm
|
| 100 |
+
pip install -e .
|
| 101 |
+
```
|
| 102 |
+
3. Install timm:
|
| 103 |
+
```shell
|
| 104 |
+
pip install timm=0.9.10
|
| 105 |
+
```
|
| 106 |
+
4. Run our demo:
|
| 107 |
+
```shell
|
| 108 |
+
python examples/minicpmv_example.py
|
| 109 |
+
```
|
| 110 |
+
</details>
|
| 111 |
+
|
| 112 |
|
| 113 |
## Usage
|
| 114 |
Inference using Huggingface transformers on Nivdia GPUs or Mac with MPS (Apple silicon or AMD GPUs). Requirements tested on python 3.10:
|