mirror of
https://github.com/RYDE-WORK/Langchain-Chatchat.git
synced 2026-02-06 06:49:48 +08:00
配置中心agent信息 工具配置
This commit is contained in:
parent
04590a2fd5
commit
52a119fd49
@ -50,6 +50,8 @@ def basic(**kwargs):
|
|||||||
@click.option("--model_providers_cfg_path_config", help="模型平台配置文件路径")
|
@click.option("--model_providers_cfg_path_config", help="模型平台配置文件路径")
|
||||||
@click.option("--model_providers_cfg_host", help="模型平台配置服务host")
|
@click.option("--model_providers_cfg_host", help="模型平台配置服务host")
|
||||||
@click.option("--model_providers_cfg_port", type=int, help="模型平台配置服务port")
|
@click.option("--model_providers_cfg_port", type=int, help="模型平台配置服务port")
|
||||||
|
@click.option("--set_model_platforms", type=str, help="模型平台配置")
|
||||||
|
@click.option("--set_tool_config", type=str, help="工具配置项 ")
|
||||||
@click.option("--clear", is_flag=True, help="清除配置")
|
@click.option("--clear", is_flag=True, help="清除配置")
|
||||||
@click.option("--show", is_flag=True, help="显示配置")
|
@click.option("--show", is_flag=True, help="显示配置")
|
||||||
def model(**kwargs):
|
def model(**kwargs):
|
||||||
@ -85,6 +87,13 @@ def model(**kwargs):
|
|||||||
if kwargs["model_providers_cfg_port"]:
|
if kwargs["model_providers_cfg_port"]:
|
||||||
config_model_workspace.set_model_providers_cfg_port(model_providers_cfg_port=kwargs["model_providers_cfg_port"])
|
config_model_workspace.set_model_providers_cfg_port(model_providers_cfg_port=kwargs["model_providers_cfg_port"])
|
||||||
|
|
||||||
|
if kwargs["set_model_platforms"]:
|
||||||
|
model_platforms_dict = json.loads(kwargs["set_model_platforms"])
|
||||||
|
config_model_workspace.set_model_platforms(model_platforms=model_platforms_dict)
|
||||||
|
if kwargs["set_tool_config"]:
|
||||||
|
tool_config_dict = json.loads(kwargs["set_tool_config"])
|
||||||
|
config_model_workspace.set_tool_config(tool_config=tool_config_dict)
|
||||||
|
|
||||||
if kwargs["clear"]:
|
if kwargs["clear"]:
|
||||||
config_model_workspace.clear()
|
config_model_workspace.clear()
|
||||||
if kwargs["show"]:
|
if kwargs["show"]:
|
||||||
|
|||||||
@ -86,53 +86,6 @@ class ConfigModelFactory(core_config.ConfigFactory[ConfigModel]):
|
|||||||
|
|
||||||
self.MODEL_PROVIDERS_CFG_PORT = 20000
|
self.MODEL_PROVIDERS_CFG_PORT = 20000
|
||||||
|
|
||||||
self._init_llm_work_config()
|
|
||||||
|
|
||||||
def _init_llm_work_config(self):
|
|
||||||
"""初始化知识库runtime的一些配置"""
|
|
||||||
|
|
||||||
self.LLM_MODEL_CONFIG = {
|
|
||||||
# 意图识别不需要输出,模型后台知道就行
|
|
||||||
"preprocess_model": {
|
|
||||||
self.DEFAULT_LLM_MODEL: {
|
|
||||||
"temperature": 0.05,
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"history_len": 100,
|
|
||||||
"prompt_name": "default",
|
|
||||||
"callbacks": False
|
|
||||||
},
|
|
||||||
},
|
|
||||||
"llm_model": {
|
|
||||||
self.DEFAULT_LLM_MODEL: {
|
|
||||||
"temperature": 0.9,
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"history_len": 10,
|
|
||||||
"prompt_name": "default",
|
|
||||||
"callbacks": True
|
|
||||||
},
|
|
||||||
},
|
|
||||||
"action_model": {
|
|
||||||
self.DEFAULT_LLM_MODEL: {
|
|
||||||
"temperature": 0.01,
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"prompt_name": "ChatGLM3",
|
|
||||||
"callbacks": True
|
|
||||||
},
|
|
||||||
},
|
|
||||||
"postprocess_model": {
|
|
||||||
self.DEFAULT_LLM_MODEL: {
|
|
||||||
"temperature": 0.01,
|
|
||||||
"max_tokens": 4096,
|
|
||||||
"prompt_name": "default",
|
|
||||||
"callbacks": True
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"image_model": {
|
|
||||||
"sd-turbo": {
|
|
||||||
"size": "256*256",
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
# 可以通过 model_providers 提供转换不同平台的接口为openai endpoint的能力,启动后下面变量会自动增加相应的平台
|
# 可以通过 model_providers 提供转换不同平台的接口为openai endpoint的能力,启动后下面变量会自动增加相应的平台
|
||||||
# ### 如果您已经有了一个openai endpoint的能力的地址,可以在这里直接配置
|
# ### 如果您已经有了一个openai endpoint的能力的地址,可以在这里直接配置
|
||||||
@ -314,6 +267,53 @@ class ConfigModelFactory(core_config.ConfigFactory[ConfigModel]):
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
self._init_llm_work_config()
|
||||||
|
|
||||||
|
def _init_llm_work_config(self):
|
||||||
|
"""初始化知识库runtime的一些配置"""
|
||||||
|
|
||||||
|
self.LLM_MODEL_CONFIG = {
|
||||||
|
# 意图识别不需要输出,模型后台知道就行
|
||||||
|
"preprocess_model": {
|
||||||
|
self.DEFAULT_LLM_MODEL: {
|
||||||
|
"temperature": 0.05,
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"history_len": 100,
|
||||||
|
"prompt_name": "default",
|
||||||
|
"callbacks": False
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"llm_model": {
|
||||||
|
self.DEFAULT_LLM_MODEL: {
|
||||||
|
"temperature": 0.9,
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"history_len": 10,
|
||||||
|
"prompt_name": "default",
|
||||||
|
"callbacks": True
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"action_model": {
|
||||||
|
self.DEFAULT_LLM_MODEL: {
|
||||||
|
"temperature": 0.01,
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"prompt_name": "ChatGLM3",
|
||||||
|
"callbacks": True
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"postprocess_model": {
|
||||||
|
self.DEFAULT_LLM_MODEL: {
|
||||||
|
"temperature": 0.01,
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"prompt_name": "default",
|
||||||
|
"callbacks": True
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"image_model": {
|
||||||
|
"sd-turbo": {
|
||||||
|
"size": "256*256",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
def default_llm_model(self, llm_model: str):
|
def default_llm_model(self, llm_model: str):
|
||||||
self.DEFAULT_LLM_MODEL = llm_model
|
self.DEFAULT_LLM_MODEL = llm_model
|
||||||
@ -345,6 +345,12 @@ class ConfigModelFactory(core_config.ConfigFactory[ConfigModel]):
|
|||||||
def model_providers_cfg_port(self, model_providers_cfg_port: int):
|
def model_providers_cfg_port(self, model_providers_cfg_port: int):
|
||||||
self.MODEL_PROVIDERS_CFG_PORT = model_providers_cfg_port
|
self.MODEL_PROVIDERS_CFG_PORT = model_providers_cfg_port
|
||||||
|
|
||||||
|
def model_platforms(self, model_platforms: List[Dict[str, Any]]):
|
||||||
|
self.MODEL_PLATFORMS = model_platforms
|
||||||
|
|
||||||
|
def tool_config(self, tool_config: Dict[str, Any]):
|
||||||
|
self.TOOL_CONFIG = tool_config
|
||||||
|
|
||||||
def get_config(self) -> ConfigModel:
|
def get_config(self) -> ConfigModel:
|
||||||
config = ConfigModel()
|
config = ConfigModel()
|
||||||
config.DEFAULT_LLM_MODEL = self.DEFAULT_LLM_MODEL
|
config.DEFAULT_LLM_MODEL = self.DEFAULT_LLM_MODEL
|
||||||
@ -396,6 +402,10 @@ class ConfigModelWorkSpace(core_config.ConfigWorkSpace[ConfigModelFactory, Confi
|
|||||||
_config_factory.model_providers_cfg_host(config_json.get("MODEL_PROVIDERS_CFG_HOST"))
|
_config_factory.model_providers_cfg_host(config_json.get("MODEL_PROVIDERS_CFG_HOST"))
|
||||||
if config_json.get("MODEL_PROVIDERS_CFG_PORT"):
|
if config_json.get("MODEL_PROVIDERS_CFG_PORT"):
|
||||||
_config_factory.model_providers_cfg_port(config_json.get("MODEL_PROVIDERS_CFG_PORT"))
|
_config_factory.model_providers_cfg_port(config_json.get("MODEL_PROVIDERS_CFG_PORT"))
|
||||||
|
if config_json.get("MODEL_PLATFORMS"):
|
||||||
|
_config_factory.model_platforms(config_json.get("MODEL_PLATFORMS"))
|
||||||
|
if config_json.get("TOOL_CONFIG"):
|
||||||
|
_config_factory.tool_config(config_json.get("TOOL_CONFIG"))
|
||||||
|
|
||||||
return _config_factory
|
return _config_factory
|
||||||
|
|
||||||
@ -446,5 +456,13 @@ class ConfigModelWorkSpace(core_config.ConfigWorkSpace[ConfigModelFactory, Confi
|
|||||||
self._config_factory.model_providers_cfg_port(model_providers_cfg_port)
|
self._config_factory.model_providers_cfg_port(model_providers_cfg_port)
|
||||||
self.store_config()
|
self.store_config()
|
||||||
|
|
||||||
|
def set_model_platforms(self, model_platforms: List[Dict[str, Any]]):
|
||||||
|
self._config_factory.model_platforms(model_platforms=model_platforms)
|
||||||
|
self.store_config()
|
||||||
|
|
||||||
|
def set_tool_config(self, tool_config: Dict[str, Any]):
|
||||||
|
self._config_factory.tool_config(tool_config=tool_config)
|
||||||
|
self.store_config()
|
||||||
|
|
||||||
|
|
||||||
config_model_workspace: ConfigModelWorkSpace = ConfigModelWorkSpace()
|
config_model_workspace: ConfigModelWorkSpace = ConfigModelWorkSpace()
|
||||||
Loading…
x
Reference in New Issue
Block a user