diff --git a/README.md b/README.md index edd6156..9974a9c 100644 --- a/README.md +++ b/README.md @@ -23,14 +23,13 @@ Our vision for KTransformers is to serve as a flexible platform for experimentin

🔥 Updates

-* **Feb 10, 2025**: Support Deepseek-R1 and V3 on single (24GB VRAM)/multi gpu and 382G DRAM, up to 3~28x speedup. The detailed tutorial is [here](./doc/en/DeepseekR1_V3_tutorial.md). -* **Aug 28, 2024**: Support 1M context under the InternLM2.5-7B-Chat-1M model, utilizing 24GB of VRAM and 150GB of DRAM. The detailed tutorial is [here](./doc/en/long_context_tutorial.md). +* **Feb 10, 2025**: Support Deepseek-R1 and V3 on single (24GB VRAM)/multi gpu and 382G DRAM, up to 3~28x speedup. For detailed show case and reproduction tutorial, see [here](./doc/en/DeepseekR1_V3_tutorial.md). * **Aug 28, 2024**: Decrease DeepseekV2's required VRAM from 21G to 11G. -* **Aug 15, 2024**: Update detailed [TUTORIAL](doc/en/injection_tutorial.md) for injection and multi-GPU. +* **Aug 15, 2024**: Update detailed [tutorial](doc/en/injection_tutorial.md) for injection and multi-GPU. * **Aug 14, 2024**: Support llamfile as linear backend. * **Aug 12, 2024**: Support multiple GPU; Support new model: mixtral 8\*7B and 8\*22B; Support q2k, q3k, q5k dequant on gpu. * **Aug 9, 2024**: Support windows native. - +

🌟 Show Cases

@@ -105,11 +104,6 @@ Getting started with KTransformers is simple! Follow the steps below to set up a To install KTransformers, follow the official [Installation Guide](https://kvcache-ai.github.io/ktransformers/). -Alternatively, you can install it directly via pip: - -```bash -pip install ktransformers -```

📃 Brief Injection Tutorial

At the heart of KTransformers is a user-friendly, template-based injection framework. diff --git a/doc/SUMMARY.md b/doc/SUMMARY.md index 4686212..c2461fc 100644 --- a/doc/SUMMARY.md +++ b/doc/SUMMARY.md @@ -6,10 +6,10 @@ # Tutorial - [Deepseek-R1/V3 Show Case](en/DeepseekR1_V3_tutorial.md) -- [Why Ktransformers So Fast](en/deepseek-v2-injection.md) +- [Why KTransformers So Fast](en/deepseek-v2-injection.md) - [Injection Tutorial](en/injection_tutorial.md) - [Multi-GPU Tutorial](en/multi-gpu-tutorial.md) -# Server +# Server(Temperary Deprected) - [Server](en/api/server/server.md) - [Website](en/api/server/website.md) - [Tabby](en/api/server/tabby.md) diff --git a/doc/en/install.md b/doc/en/install.md index 8c516a0..a191d22 100644 --- a/doc/en/install.md +++ b/doc/en/install.md @@ -1,6 +1,6 @@ # How to Run DeepSeek-R1 -In this document, we will show you how to run the local_chat.py script to test the DeepSeek-R1's performance. There are two versions: +In this document, we will show you how to install and run KTransformers on your local machine. There are two versions: * V0.2 is the current main branch. * V0.3 is a preview version only provides binary distribution for now. * To reproduce our DeepSeek-R1/V3 results, please refer to [Deepseek-R1/V3 Tutorial](./DeepseekR1_V3_tutorial.md) for more detail settings after installation.