From 9e100957ebb45e6374387497d815d621e5318135 Mon Sep 17 00:00:00 2001 From: Archer <545436317@qq.com> Date: Sat, 1 Feb 2025 18:40:44 +0800 Subject: [PATCH] fix: stream response (#3682) * perf: supplement assistant empty response * check array * fix: stream response * fix: model config cannot set to null --- .../core/ai/config/provider/OpenAI.json | 40 +++++++++++++++---- .../core/workflow/dispatch/chat/oneapi.ts | 15 ++++--- .../account/model/ModelConfigTable.tsx | 4 +- 3 files changed, 42 insertions(+), 17 deletions(-) diff --git a/packages/service/core/ai/config/provider/OpenAI.json b/packages/service/core/ai/config/provider/OpenAI.json index b45942518..a852cf22d 100644 --- a/packages/service/core/ai/config/provider/OpenAI.json +++ b/packages/service/core/ai/config/provider/OpenAI.json @@ -45,14 +45,14 @@ "type": "llm" }, { - "model": "o1-mini", - "name": "o1-mini", - "maxContext": 128000, - "maxResponse": 4000, + "model": "o3-mini", + "name": "o3-mini", + "maxContext": 200000, + "maxResponse": 100000, "quoteMaxToken": 120000, "maxTemperature": null, "vision": false, - "toolChoice": false, + "toolChoice": true, "functionCall": false, "defaultSystemChatPrompt": "", "datasetProcess": true, @@ -71,8 +71,8 @@ "type": "llm" }, { - "model": "o1-preview", - "name": "o1-preview", + "model": "o1-mini", + "name": "o1-mini", "maxContext": 128000, "maxResponse": 4000, "quoteMaxToken": 120000, @@ -103,6 +103,32 @@ "maxResponse": 8000, "quoteMaxToken": 120000, "maxTemperature": null, + "vision": true, + "toolChoice": false, + "functionCall": false, + "defaultSystemChatPrompt": "", + "datasetProcess": true, + "usedInClassify": true, + "customCQPrompt": "", + "usedInExtractFields": true, + "usedInQueryExtension": true, + "customExtractPrompt": "", + "usedInToolCall": true, + "defaultConfig": { + "stream": false + }, + "fieldMap": { + "max_tokens": "max_completion_tokens" + }, + "type": "llm" + }, + { + "model": "o1-preview", + "name": "o1-preview", + "maxContext": 128000, + "maxResponse": 4000, + "quoteMaxToken": 120000, + "maxTemperature": null, "vision": false, "toolChoice": false, "functionCall": false, diff --git a/packages/service/core/workflow/dispatch/chat/oneapi.ts b/packages/service/core/workflow/dispatch/chat/oneapi.ts index 93f23651f..0d3482b68 100644 --- a/packages/service/core/workflow/dispatch/chat/oneapi.ts +++ b/packages/service/core/workflow/dispatch/chat/oneapi.ts @@ -208,14 +208,13 @@ export const dispatchChatCompletion = async (props: ChatProps): Promise { if (!e) { - setValue('defaultConfig', undefined); + setValue('defaultConfig', {}); return; } try { @@ -1014,7 +1014,7 @@ const ModelEditModal = ({ resize onChange={(e) => { if (!e) { - setValue('defaultConfig', undefined); + setValue('defaultConfig', {}); return; } try {