mirror of
https://github.com/RYDE-WORK/Langchain-Chatchat.git
synced 2026-01-19 13:23:16 +08:00
更新两行注释
This commit is contained in:
parent
f9a5ae99ee
commit
852ca5f849
@ -40,8 +40,6 @@ FSCHAT_MODEL_WORKERS = {
|
||||
"device": LLM_DEVICE,
|
||||
# False,'vllm',使用的推理加速框架,使用vllm如果出现HuggingFace通信问题,参见doc/FAQ
|
||||
# vllm对一些模型支持还不成熟,暂时默认关闭
|
||||
# fschat=0.2.33的代码有bug, 如需使用,源码修改fastchat.server.vllm_worker,
|
||||
# 将103行中sampling_params = SamplingParams的参数stop=list(stop)修改为stop= [i for i in stop if i!=""]
|
||||
"infer_turbo": False,
|
||||
|
||||
# model_worker多卡加载需要配置的参数
|
||||
|
||||
@ -164,11 +164,11 @@ def dialogue_page(api: ApiRequest, is_lite: bool = False):
|
||||
available_models = []
|
||||
config_models = api.list_config_models()
|
||||
if not is_lite:
|
||||
for k, v in config_models.get("local", {}).items(): # 列出配置了有效本地路径的模型
|
||||
for k, v in config_models.get("local", {}).items():
|
||||
if (v.get("model_path_exists")
|
||||
and k not in running_models):
|
||||
available_models.append(k)
|
||||
for k, v in config_models.get("online", {}).items(): # 列出ONLINE_MODELS中可直接访问且在LLM_MODELS中配置的模型
|
||||
for k, v in config_models.get("online", {}).items():
|
||||
if not v.get("provider") and k not in running_models and k in LLM_MODELS:
|
||||
available_models.append(k)
|
||||
llm_models = running_models + available_models
|
||||
|
||||
@ -437,9 +437,6 @@ class ApiRequest:
|
||||
"prompt_name": prompt_name,
|
||||
}
|
||||
|
||||
# print(f"received input message:")
|
||||
# pprint(data)
|
||||
|
||||
response = self.post(
|
||||
"/chat/file_chat",
|
||||
json=data,
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user