From 9bfdc6febbec5be5af2ac1f0754cf0ce95cfa162 Mon Sep 17 00:00:00 2001 From: glide-the <2533736852@qq.com> Date: Wed, 12 Jun 2024 14:14:26 +0800 Subject: [PATCH] =?UTF-8?q?provider=5Fmodel=5Frecord=20model=5Fcredentials?= =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=97=AE=E9=A2=98?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../xinference/llm/_position.yaml | 1 - .../xinference/llm/chatglm3-6b.yaml | 43 ------------------- .../model_providers/core/provider_manager.py | 2 +- .../xinference/model_providers.yaml | 1 + 4 files changed, 2 insertions(+), 45 deletions(-) delete mode 100644 libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/_position.yaml delete mode 100644 libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/chatglm3-6b.yaml diff --git a/libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/_position.yaml b/libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/_position.yaml deleted file mode 100644 index 36640c5e..00000000 --- a/libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/_position.yaml +++ /dev/null @@ -1 +0,0 @@ -- chatglm3-6b diff --git a/libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/chatglm3-6b.yaml b/libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/chatglm3-6b.yaml deleted file mode 100644 index 371b0126..00000000 --- a/libs/model-providers/model_providers/core/model_runtime/model_providers/xinference/llm/chatglm3-6b.yaml +++ /dev/null @@ -1,43 +0,0 @@ -model: chatglm3-6b -label: - zh_Hans: chatglm3-6b - en_US: chatglm3-6b -model_type: llm -features: - - multi-tool-call - - agent-thought - - stream-tool-call -model_properties: - mode: chat - context_size: 4096 -parameter_rules: - - name: temperature - use_template: temperature - - name: top_p - use_template: top_p - - name: presence_penalty - use_template: presence_penalty - - name: frequency_penalty - use_template: frequency_penalty - - name: max_tokens - use_template: max_tokens - default: 512 - min: 1 - max: 4096 - - name: response_format - label: - zh_Hans: 回复格式 - en_US: response_format - type: string - help: - zh_Hans: 指定模型必须输出的格式 - en_US: specifying the format that the model must output - required: false - options: - - text - - json_object -pricing: - input: '0.001' - output: '0.002' - unit: '0.001' - currency: USD diff --git a/libs/model-providers/model_providers/core/provider_manager.py b/libs/model-providers/model_providers/core/provider_manager.py index 1cde79d2..a4c2af32 100644 --- a/libs/model-providers/model_providers/core/provider_manager.py +++ b/libs/model-providers/model_providers/core/provider_manager.py @@ -239,7 +239,7 @@ class ProviderManager: if not provider_model_record.get("model_credentials"): continue - provider_model_credentials = {} + provider_model_credentials = provider_model_record.get("model_credentials") for variable in model_credential_variables: if variable in provider_model_record.get("model_credentials"): try: diff --git a/libs/model-providers/tests/unit_tests/xinference/model_providers.yaml b/libs/model-providers/tests/unit_tests/xinference/model_providers.yaml index a7ee556d..f93d47e0 100644 --- a/libs/model-providers/tests/unit_tests/xinference/model_providers.yaml +++ b/libs/model-providers/tests/unit_tests/xinference/model_providers.yaml @@ -6,6 +6,7 @@ xinference: model_credentials: server_url: 'http://127.0.0.1:9997/' model_uid: 'chatglm3-6b' + completion_type: 'chat' - model: 'bge-m3' model_type: 'embeddings' completion_type: 'embeddings'