mirror of
https://github.com/RYDE-WORK/llama.cpp.git
synced 2026-01-20 05:33:37 +08:00
* Bump model_template to 16384 bytes to support larger chat templates. * Use `model->gguf_kv` for efficiency.
* Bump model_template to 16384 bytes to support larger chat templates. * Use `model->gguf_kv` for efficiency.