mirror of
https://github.com/RYDE-WORK/MiniCPM.git
synced 2026-02-05 14:43:30 +08:00
commit
0920a16e1c
@ -64,7 +64,7 @@ We release all model parameters for research and limited commercial use.
|
|||||||
|-------------|------------|-----------|-----------|
|
|-------------|------------|-----------|-----------|
|
||||||
|[Transformers](#Huggingface)|[Transformers](#6)|[MLC](#MLC)|[GPTQ](#gptq)|
|
|[Transformers](#Huggingface)|[Transformers](#6)|[MLC](#MLC)|[GPTQ](#gptq)|
|
||||||
|[vLLM](#vLLM)|[mlx_finetune](#mlx_finetune)|[llama.cpp](#llama.cpp)|[AWQ](#awq)|
|
|[vLLM](#vLLM)|[mlx_finetune](#mlx_finetune)|[llama.cpp](#llama.cpp)|[AWQ](#awq)|
|
||||||
|[llama.cpp](#llama.cpp)|[llama_factory](./finetune/llama_factory_example/README.md)||[bnb](#bnb)|
|
|[llama.cpp](#llama.cpp)|[LLaMA-Factory](./finetune/llama_factory_example/README.md)||[bnb](#bnb)|
|
||||||
|[ollama](#ollama)|||[quantize_test](#quantize_test)|
|
|[ollama](#ollama)|||[quantize_test](#quantize_test)|
|
||||||
|[fastllm](#fastllm)||||
|
|[fastllm](#fastllm)||||
|
||||||
|[mlx_lm](#mlx)||||
|
|[mlx_lm](#mlx)||||
|
||||||
@ -416,7 +416,7 @@ python bnb_quantize.py
|
|||||||
<p id="Community"></p>
|
<p id="Community"></p>
|
||||||
|
|
||||||
## Community
|
## Community
|
||||||
|
- [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory.git): [MiniCPM fine-tuning one-click solution](https://modelbest.feishu.cn/wiki/AIU3wbREcirOm9kkvd7cxujFnMb?from=from_copylink)
|
||||||
- [ChatLLM](https://github.com/foldl/chatllm.cpp): [Run MiniCPM on CPU](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16/discussions/2#65c59c4f27b8c11e43fc8796)
|
- [ChatLLM](https://github.com/foldl/chatllm.cpp): [Run MiniCPM on CPU](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16/discussions/2#65c59c4f27b8c11e43fc8796)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@ -64,7 +64,7 @@ MiniCPM 是面壁智能与清华大学自然语言处理实验室共同开源的
|
|||||||
|-------------|------------|-----------|-----------|
|
|-------------|------------|-----------|-----------|
|
||||||
|[Transformers](#Huggingface模型)|[Transformers](#transformer_finetune)|[MLC部署](#MLC)|[GPTQ](#gptq)|
|
|[Transformers](#Huggingface模型)|[Transformers](#transformer_finetune)|[MLC部署](#MLC)|[GPTQ](#gptq)|
|
||||||
|[vLLM](#vllm-推理)|[mlx_finetune](#mlx)|[llama.cpp](#llama.cpp)|[AWQ](#awq)|
|
|[vLLM](#vllm-推理)|[mlx_finetune](#mlx)|[llama.cpp](#llama.cpp)|[AWQ](#awq)|
|
||||||
|[llama.cpp](#llama.cpp)|[llama_factory](./finetune/llama_factory_example/README.md)||[bnb](#bnb)|
|
|[llama.cpp](#llama.cpp)|[LLaMA-Factory](./finetune/llama_factory_example/README.md)||[bnb](#bnb)|
|
||||||
|[ollama](#ollama)|||[量化测试](#quantize_test)|
|
|[ollama](#ollama)|||[量化测试](#quantize_test)|
|
||||||
|[fastllm](#fastllm)||||
|
|[fastllm](#fastllm)||||
|
||||||
|[mlx_lm](#mlx_lm)||||
|
|[mlx_lm](#mlx_lm)||||
|
||||||
@ -424,7 +424,7 @@ python bnb_quantize.py
|
|||||||
<p id="community"></p>
|
<p id="community"></p>
|
||||||
|
|
||||||
## 开源社区
|
## 开源社区
|
||||||
|
- [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory.git):[MiniCPM微调一键式解决方案](https://modelbest.feishu.cn/wiki/AIU3wbREcirOm9kkvd7cxujFnMb?from=from_copylink)
|
||||||
- [ChatLLM框架](https://github.com/foldl/chatllm.cpp):[在CPU上跑MiniCPM](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16/discussions/2#65c59c4f27b8c11e43fc8796)
|
- [ChatLLM框架](https://github.com/foldl/chatllm.cpp):[在CPU上跑MiniCPM](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16/discussions/2#65c59c4f27b8c11e43fc8796)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user