mirror of
https://github.com/RYDE-WORK/llama.cpp.git
synced 2026-01-19 21:23:26 +08:00
* server: add option to output probabilities for completion * server: fix issue when handling probability output for incomplete tokens for multibyte character generation * server: fix llama_sample_top_k order * examples/common.h: put all bool variables in gpt_params together