diff --git a/frontend/src/app/settings/llm/ChatChat/index.tsx b/frontend/src/app/settings/llm/ChatChat/index.tsx index 1849cd65..5d7c82e8 100644 --- a/frontend/src/app/settings/llm/ChatChat/index.tsx +++ b/frontend/src/app/settings/llm/ChatChat/index.tsx @@ -18,10 +18,6 @@ const ChatChatProvider = memo(() => { const { t } = useTranslation('setting'); const theme = useTheme(); - console.log('----ttt---1-', t('llm.ChatChat.endpoint.title')) - console.log('----ttt---2-', t('llm.Ollama.endpoint.title')) - console.log('----ttt---3-', t) - return ( (({ model, size = 12 }) => { return ; if (model.includes('mistral') || model.includes('mixtral')) return ; if (model.includes('pplx')) return ; + if (model.includes('Spark')) return ; + if (model.includes('ERNIE')) return ; }); export default ModelIcon; diff --git a/frontend/src/config/modelProviders/chatchat.ts b/frontend/src/config/modelProviders/chatchat.ts index 83af503f..5c736fb8 100644 --- a/frontend/src/config/modelProviders/chatchat.ts +++ b/frontend/src/config/modelProviders/chatchat.ts @@ -4,15 +4,78 @@ const ChatChat: ModelProviderCard = { id: 'chatchat', chatModels: [ { - id: 'chatglm_pro', - tokens: 128_000, - displayName: 'chatglm_pro' + id: 'chatglm3-6b', + tokens: 4000, + displayName: 'chatglm3-6b', + functionCall: true, }, { - id: 'gpt-4-turbo-2024-04-09', - tokens: 128_000, - displayName: 'gpt-4-turbo-2024-04-09', - vision: true, + id: 'chatglm_turbo', + tokens: 4000, + displayName: 'chatglm_turbo', + }, + { + id: 'chatglm_std', + tokens: 4000, + displayName: 'chatglm_std', + }, + { + id: 'chatglm_lite', + tokens: 4000, + displayName: 'chatglm_lite', + }, + { + id: 'qwen-turbo', + tokens: 4000, + displayName: 'qwen-turbo', + functionCall: true, + }, + { + id: 'qwen-plus', + tokens: 4000, + displayName: 'qwen-plus', + }, + { + id: 'qwen-max', + tokens: 4000, + displayName: 'qwen-max', + }, + { + id: 'qwen:7b', + tokens: 4000, + displayName: 'qwen:7b', + functionCall: true, + }, + { + id: 'qwen:14b', + tokens: 4000, + displayName: 'qwen:14b', + functionCall: true, + }, + { + id: 'qwen-max-longcontext', + tokens: 4000, + displayName: 'qwen-max-longcontext', + }, + { + id: 'ERNIE-Bot', + tokens: 4000, + displayName: 'ERNIE-Bot', + }, + { + id: 'ERNIE-Bot-turbo', + tokens: 4000, + displayName: 'ERNIE-Bot-turbo', + }, + { + id: 'ERNIE-Bot-4', + tokens: 4000, + displayName: 'ERNIE-Bot-4', + }, + { + id: 'SparkDesk', + tokens: 4000, + displayName: 'SparkDesk', } ] } diff --git a/frontend/src/features/ChatInput/ActionBar/Tools/index.tsx b/frontend/src/features/ChatInput/ActionBar/Tools/index.tsx index 68939be9..9a4c898f 100644 --- a/frontend/src/features/ChatInput/ActionBar/Tools/index.tsx +++ b/frontend/src/features/ChatInput/ActionBar/Tools/index.tsx @@ -35,9 +35,6 @@ const Tools = memo(() => { const list = useToolStore(pluginSelectors.installedPluginMetaList, isEqual); const builtinList = useToolStore(builtinToolSelectors.metaList, isEqual); - console.log('builtinList--', builtinList) - console.log('list--', list) - const enablePluginCount = useSessionStore( (s) => agentSelectors diff --git a/frontend/src/libs/agent-runtime/chatchat/index.ts b/frontend/src/libs/agent-runtime/chatchat/index.ts index fb605fb3..9a62f142 100644 --- a/frontend/src/libs/agent-runtime/chatchat/index.ts +++ b/frontend/src/libs/agent-runtime/chatchat/index.ts @@ -27,7 +27,6 @@ export class LobeChatChatAI implements LobeRuntimeAI { async chat(payload: ChatStreamPayload, options?: ChatCompetitionOptions) { - console.log('payload---', payload) try { const response = await this.client.chat.completions.create( payload as unknown as (OpenAI.ChatCompletionCreateParamsStreaming | OpenAI.ChatCompletionCreateParamsNonStreaming),