From 86b90c9d0d92e89480e4cd9c9a7391fe090e1f27 Mon Sep 17 00:00:00 2001 From: SillyXu Date: Mon, 5 Feb 2024 20:05:26 +0900 Subject: [PATCH] Update README-en.md --- README-en.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README-en.md b/README-en.md index 5ea91d6..9db7b07 100644 --- a/README-en.md +++ b/README-en.md @@ -345,7 +345,7 @@ print(res) #### Performance * We did not conduct in-depth optimization and system testing on the mobile inference model, only verifying the feasibility of MiniCPM using mobile phone chips for inference. -* There have been no previous attempts to deploy multimodal models on mobile phones. We have verified the feasibility of deploying MiniCPM-V on mobile phones based on MLC-LLM this time, and it can input and output normally. However, there also exist a problem of long image processing time, which needs further optimization :) +* Besides us, there are also some [efforts](https://github.com/ggerganov/llama.cpp/blob/master/examples/llava/MobileVLM-README.md) to deploy multimodal models on mobile phones based on llama.cpp. We have verified the feasibility of deploying MiniCPM-V on mobile phones based on MLC-LLM this time, and it can input and output normally. However, there also exist a problem of long image processing time, which needs further optimization :) * **We welcome more developers to continuously improve the inference performance of LLMs on mobile phones and update the test results below.** | Mobile Phones | OS | Processor | Memory(GB) | Inference Throughput(token/s) |