model perf (#3657)

* fix: model

* dataset quote

* perf: model config

* model tag

* doubao model config

* perf: config model

* feat: model test
This commit is contained in:
Archer
2025-01-24 14:10:14 +08:00
committed by archer
parent f2be9ae32d
commit e48df175d7
171 changed files with 1902 additions and 3126 deletions

View File

@@ -1,5 +1,6 @@
import type { NextApiResponse } from 'next';
import { getAIApi } from '../config';
import { getTTSModel } from '../model';
export async function text2Speech({
res,
@@ -18,15 +19,26 @@ export async function text2Speech({
voice: string;
speed?: number;
}) {
const modelData = getTTSModel(model)!;
const ai = getAIApi();
const response = await ai.audio.speech.create({
model,
// @ts-ignore
voice,
input,
response_format: 'mp3',
speed
});
const response = await ai.audio.speech.create(
{
model,
// @ts-ignore
voice,
input,
response_format: 'mp3',
speed
},
modelData.requestUrl && modelData.requestAuth
? {
path: modelData.requestUrl,
headers: {
Authorization: `Bearer ${modelData.requestAuth}`
}
}
: {}
);
const readableStream = response.body as unknown as NodeJS.ReadableStream;
readableStream.pipe(res);

View File

@@ -14,8 +14,7 @@ export const openaiBaseUrl = process.env.OPENAI_BASE_URL || 'https://api.openai.
export const getAIApi = (props?: { userKey?: OpenaiAccountType; timeout?: number }) => {
const { userKey, timeout } = props || {};
const baseUrl =
userKey?.baseUrl || global?.systemEnv?.oneapiUrl || process.env.ONEAPI_URL || openaiBaseUrl;
const baseUrl = userKey?.baseUrl || global?.systemEnv?.oneapiUrl || openaiBaseUrl;
const apiKey = userKey?.key || global?.systemEnv?.chatApiKey || process.env.CHAT_API_KEY || '';
return new OpenAI({
@@ -30,8 +29,7 @@ export const getAIApi = (props?: { userKey?: OpenaiAccountType; timeout?: number
export const getAxiosConfig = (props?: { userKey?: OpenaiAccountType }) => {
const { userKey } = props || {};
const baseUrl =
userKey?.baseUrl || global?.systemEnv?.oneapiUrl || process.env.ONEAPI_URL || openaiBaseUrl;
const baseUrl = userKey?.baseUrl || global?.systemEnv?.oneapiUrl || openaiBaseUrl;
const apiKey = userKey?.key || global?.systemEnv?.chatApiKey || process.env.CHAT_API_KEY || '';
return {

View File

@@ -0,0 +1,10 @@
{
"provider": "Siliconflow",
"model": "BAAI/bge-m3",
"name": "BAAI/bge-m3",
"defaultToken": 512,
"maxToken": 8000,
"charsPointsPrice": 0
}

View File

@@ -4,7 +4,7 @@
"name": "text-embedding-3-large",
"defaultToken": 512,
"maxToken": 3000,
"maxToken": 8000,
"charsPointsPrice": 0,

View File

@@ -4,7 +4,7 @@
"name": "text-embedding-3-small",
"defaultToken": 512,
"maxToken": 3000,
"maxToken": 8000,
"charsPointsPrice": 0
}

View File

@@ -4,7 +4,7 @@
"name": "text-embedding-ada-002",
"defaultToken": 512,
"maxToken": 3000,
"maxToken": 8000,
"charsPointsPrice": 0
}

View File

@@ -3,26 +3,23 @@
"model": "Doubao-lite-128k",
"name": "Doubao-lite-128k",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 120000,
"maxTemperature": 1,
"vision": false,
"toolChoice": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,16 +3,13 @@
"model": "Doubao-lite-32k",
"name": "Doubao-lite-32k",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 30000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
"toolChoice": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
@@ -22,7 +19,7 @@
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -0,0 +1,26 @@
{
"provider": "Doubao",
"model": "Doubao-lite-4k",
"name": "Doubao-lite-4k",
"maxContext": 4000,
"maxResponse": 4000,
"quoteMaxToken": 4000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -3,26 +3,23 @@
"model": "Doubao-pro-128k",
"name": "Doubao-pro-128k",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 120000,
"maxTemperature": 1,
"vision": false,
"toolChoice": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,16 +3,13 @@
"model": "Doubao-pro-32k",
"name": "Doubao-pro-32k",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 30000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
"toolChoice": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
@@ -22,7 +19,7 @@
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -0,0 +1,26 @@
{
"provider": "Doubao",
"model": "Doubao-pro-4k",
"name": "Doubao-pro-4k",
"maxContext": 4000,
"maxResponse": 4000,
"quoteMaxToken": 4000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -3,12 +3,9 @@
"model": "Doubao-vision-lite-32k",
"name": "Doubao-vision-lite-32k",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 30000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": true,
@@ -22,7 +19,7 @@
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,12 +3,9 @@
"model": "Doubao-vision-pro-32k",
"name": "Doubao-vision-pro-32k",
"censor": false,
"charsPointsPrice": 2,
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 30000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": true,
@@ -22,7 +19,7 @@
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,9 +3,6 @@
"model": "ERNIE-4.0-8K",
"name": "ERNIE-4.0-8K",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 8000,
"maxResponse": 2048,
"quoteMaxToken": 5000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "ERNIE-4.0-Turbo-8K",
"name": "ERNIE-4.0-Turbo-8K",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 8000,
"maxResponse": 2048,
"quoteMaxToken": 5000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "ERNIE-Lite-8K",
"name": "ERNIE-lite-8k",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 8000,
"maxResponse": 2048,
"quoteMaxToken": 6000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -1,10 +1,7 @@
{
"provider": "Ernie",
"model": "ERNIE-Speed-128K",
"name": "ERNIE-Speed-128K(测试)",
"censor": false,
"charsPointsPrice": 0,
"name": "ERNIE-Speed-128K",
"maxContext": 128000,
"maxResponse": 4096,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "MiniMax-Text-01",
"name": "MiniMax-Text-01",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 1000000,
"maxResponse": 1000000,
"quoteMaxToken": 100000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -0,0 +1,20 @@
{
"provider": "SparkDesk",
"model": "lite",
"name": "SparkDesk-lite",
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,26 @@
{
"provider": "SparkDesk",
"model": "max-32k",
"name": "SparkDesk-max-32k",
"maxContext": 32000,
"maxResponse": 8000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
"toolChoice": false,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -0,0 +1,20 @@
{
"provider": "SparkDesk",
"model": "generalv3.5",
"name": "SparkDesk-max",
"maxContext": 32000,
"maxResponse": 8000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "SparkDesk",
"model": "pro-128k",
"name": "SparkDesk-Pro-128k",
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 128000,
"maxTemperature": 1,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -1,15 +1,13 @@
{
"provider": "SparkDesk",
"model": "SparkDesk-v3.5",
"name": "SparkDesk-v3.5",
"model": "generalv3",
"name": "SparkDesk-Pro",
"maxContext": 8000,
"maxResponse": 8000,
"quoteMaxToken": 6000,
"quoteMaxToken": 8000,
"maxTemperature": 1,
"charsPointsPrice": 0.3,
"censor": false,
"vision": false,
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,

View File

@@ -1,14 +1,11 @@
{
"provider": "SparkDesk",
"model": "SparkDesk-v4.0",
"name": "SparkDesk-v4.0",
"censor": false,
"charsPointsPrice": 0,
"model": "4.0Ultra",
"name": "SparkDesk-v4.0 Ultra",
"maxContext": 8000,
"maxResponse": 8000,
"quoteMaxToken": 6000,
"quoteMaxToken": 8000,
"maxTemperature": 1,
"vision": false,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "abab6.5s-chat",
"name": "MiniMax-abab6.5s",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 245000,
"maxResponse": 10000,
"quoteMaxToken": 240000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "claude-3-5-haiku-20241022",
"name": "claude-3-5-haiku-20241022",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 200000,
"maxResponse": 8000,
"quoteMaxToken": 100000,

View File

@@ -3,9 +3,6 @@
"model": "claude-3-5-sonnet-20240620",
"name": "Claude-3-5-sonnet-20240620",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 200000,
"maxResponse": 8000,
"quoteMaxToken": 100000,

View File

@@ -3,9 +3,6 @@
"model": "claude-3-5-sonnet-20241022",
"name": "Claude-3-5-sonnet-20241022",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 200000,
"maxResponse": 8000,
"quoteMaxToken": 100000,

View File

@@ -3,9 +3,6 @@
"model": "claude-3-opus-20240229",
"name": "claude-3-opus-20240229",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 200000,
"maxResponse": 4096,
"quoteMaxToken": 100000,

View File

@@ -3,9 +3,6 @@
"model": "deepseek-chat",
"name": "Deepseek-chat",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 64000,
"maxResponse": 4096,
"quoteMaxToken": 60000,

View File

@@ -3,9 +3,6 @@
"model": "deepseek-reasoner",
"name": "Deepseek-reasoner",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 64000,
"maxResponse": 4096,
"quoteMaxToken": 60000,

View File

@@ -3,9 +3,6 @@
"model": "gemini-1.5-flash",
"name": "Gemini-1.5-flash",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 1000000,
"maxResponse": 8000,
"quoteMaxToken": 60000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "gemini-1.5-pro",
"name": "Gemini-1.5-pro",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 2000000,
"maxResponse": 8000,
"quoteMaxToken": 60000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "gemini-2.0-flash-exp",
"name": "Gemini-2.0-flash-exp",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 1000000,
"maxResponse": 8000,
"quoteMaxToken": 60000,

View File

@@ -3,9 +3,6 @@
"model": "gemini-2.0-flash-thinking-exp-01-21",
"name": "Gemini-2.0-flash-thinking-exp",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 1000000,
"maxResponse": 8000,
"quoteMaxToken": 60000,

View File

@@ -3,9 +3,6 @@
"model": "gemini-exp-1206",
"name": "gemini-exp-1206",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 120000,

View File

@@ -3,9 +3,6 @@
"model": "glm-4-air",
"name": "glm-4-air",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 120000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "glm-4-flash",
"name": "glm-4-flash",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 120000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "glm-4-long",
"name": "glm-4-long",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 1000000,
"maxResponse": 4000,
"quoteMaxToken": 900000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "glm-4-plus",
"name": "GLM-4-plus",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 120000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "glm-4v-flash",
"name": "glm-4v-flash",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 8000,
"maxResponse": 1000,
"quoteMaxToken": 6000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "glm-4v-plus",
"name": "GLM-4v-plus",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 8000,
"maxResponse": 1000,
"quoteMaxToken": 6000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -1,10 +1,7 @@
{
"provider": "OpenAI",
"model": "gpt-3.5-turbo",
"name": "gpt-3.5-turbo(abandon)",
"censor": true,
"charsPointsPrice": 1,
"name": "gpt-3.5-turbo",
"maxContext": 16000,
"maxResponse": 4000,
@@ -16,15 +13,11 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {
"max_tokens": null
}
"usedInToolCall": true
}

View File

@@ -1,12 +1,9 @@
{
"provider": "OpenAI",
"model": "gpt-4-turbo",
"name": "gpt-4-turbo(abandon)",
"name": "gpt-4-turbo",
"censor": true,
"charsPointsPrice": 15,
"maxContext": 125000,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 60000,
"maxTemperature": 1.2,
@@ -16,15 +13,11 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {
"max_tokens": null
}
"usedInToolCall": true
}

View File

@@ -3,10 +3,7 @@
"model": "gpt-4o-mini",
"name": "GPT-4o-mini",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 125000,
"maxContext": 128000,
"maxResponse": 16000,
"quoteMaxToken": 60000,
"maxTemperature": 1.2,

View File

@@ -3,10 +3,7 @@
"model": "gpt-4o",
"name": "GPT-4o",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 125000,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 60000,
"maxTemperature": 1.2,

View File

@@ -3,10 +3,7 @@
"model": "llama-3.1-8b-instant",
"name": "Groq-llama-3.1-8b-instant",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 125000,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 60000,
"maxTemperature": 1.2,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,10 +3,7 @@
"model": "llama-3.3-70b-versatile",
"name": "Groq-llama-3.3-70b-versatile",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 125000,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 60000,
"maxTemperature": 1.2,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "hunyuan-large",
"name": "hunyuan-large",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 28000,
"maxResponse": 4000,
"quoteMaxToken": 20000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "hunyuan-lite",
"name": "hunyuan-lite",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 250000,
"maxResponse": 6000,
"quoteMaxToken": 100000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -1,10 +1,7 @@
{
"provider": "Hunyuan",
"model": "hunyuan-pro-32k(测试)",
"name": "hunyuan-pro-32k(测试)",
"censor": false,
"charsPointsPrice": 0,
"model": "hunyuan-pro",
"name": "hunyuan-pro",
"maxContext": 28000,
"maxResponse": 4000,

View File

@@ -3,10 +3,7 @@
"model": "hunyuan-standard",
"name": "hunyuan-standard",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 30000,
"maxContext": 32000,
"maxResponse": 2000,
"quoteMaxToken": 20000,
"maxTemperature": 1,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "hunyuan-turbo-vision",
"name": "hunyuan-turbo-vision",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 6000,
"maxResponse": 2000,
"quoteMaxToken": 6000,
@@ -16,13 +13,13 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"usedInClassify": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": false,
"usedInQueryExtension": false,
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,9 +3,6 @@
"model": "hunyuan-turbo",
"name": "hunyuan-turbo",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 28000,
"maxResponse": 4000,
"quoteMaxToken": 20000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "hunyuan-vision",
"name": "hunyuan-vision",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 6000,
"maxResponse": 2000,
"quoteMaxToken": 4000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -0,0 +1,26 @@
{
"provider": "Intern",
"model": "internlm2-pro-chat",
"name": "internlm2-pro-chat",
"maxContext": 32000,
"maxResponse": 8000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -0,0 +1,26 @@
{
"provider": "Intern",
"model": "internlm3-8b-instruct",
"name": "internlm3-8b-instruct",
"maxContext": 32000,
"maxResponse": 8000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -3,9 +3,6 @@
"model": "ministral-3b-latest",
"name": "Ministral-3b-latest",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 130000,
"maxResponse": 8000,
"quoteMaxToken": 60000,

View File

@@ -3,9 +3,6 @@
"model": "ministral-8b-latest",
"name": "Ministral-8b-latest",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 130000,
"maxResponse": 8000,
"quoteMaxToken": 60000,

View File

@@ -3,9 +3,6 @@
"model": "mistral-large-latest",
"name": "Mistral-large-latest",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 130000,
"maxResponse": 8000,
"quoteMaxToken": 60000,

View File

@@ -3,12 +3,9 @@
"model": "mistral-small-latest",
"name": "Mistral-small-latest",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 30000,
"quoteMaxToken": 32000,
"maxTemperature": 1.2,
"vision": false,

View File

@@ -1,10 +1,7 @@
{
"provider": "Moonshot",
"model": "moonshot-v1-128k",
"name": "月之暗面-128k",
"censor": false,
"charsPointsPrice": 0,
"name": "moonshot-v1-128k",
"maxContext": 128000,
"maxResponse": 4000,
@@ -22,7 +19,7 @@
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -1,14 +1,11 @@
{
"provider": "Moonshot",
"model": "moonshot-v1-32k",
"name": "月之暗面-32k",
"censor": false,
"charsPointsPrice": 0,
"name": "moonshot-v1-32k",
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 30000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": false,
@@ -22,7 +19,7 @@
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -1,10 +1,7 @@
{
"provider": "Moonshot",
"model": "moonshot-v1-8k",
"name": "月之暗面-8k",
"censor": false,
"charsPointsPrice": 0,
"name": "moonshot-v1-8k",
"maxContext": 8000,
"maxResponse": 4000,
@@ -16,13 +13,13 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,10 +3,7 @@
"model": "o1-mini",
"name": "o1-mini",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 125000,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 120000,
"maxTemperature": 1.2,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,10 +3,7 @@
"model": "o1-preview",
"name": "o1-preview",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 125000,
"maxContext": 128000,
"maxResponse": 4000,
"quoteMaxToken": 120000,
"maxTemperature": 1.2,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "o1",
"name": "o1",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 195000,
"maxResponse": 8000,
"quoteMaxToken": 120000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "qwen-coder-turbo",
"name": "qwen-coder-turbo",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 50000,
@@ -16,13 +13,13 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"usedInClassify": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": false,
"usedInQueryExtension": false,
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,9 +3,6 @@
"model": "qwen-max",
"name": "Qwen-max",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 8000,
"maxResponse": 4000,
"quoteMaxToken": 6000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "qwen-plus",
"name": "Qwen-plus",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 64000,
"maxResponse": 8000,
"quoteMaxToken": 60000,

View File

@@ -3,9 +3,6 @@
"model": "qwen-turbo",
"name": "Qwen-turbo",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 100000,

View File

@@ -3,9 +3,6 @@
"model": "qwen-vl-max",
"name": "qwen-vl-max",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 32000,
"maxResponse": 2000,
"quoteMaxToken": 20000,
@@ -16,13 +13,13 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"usedInClassify": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": false,
"usedInQueryExtension": false,
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false,
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}

View File

@@ -3,9 +3,6 @@
"model": "qwen-vl-plus",
"name": "qwen-vl-plus",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 32000,
"maxResponse": 2000,
"quoteMaxToken": 20000,
@@ -16,11 +13,11 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"usedInClassify": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": false,
"usedInQueryExtension": false,
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": false
"usedInToolCall": true
}

View File

@@ -1,10 +1,7 @@
{
"provider": "Qwen",
"model": "qwen2.5-14b-instruct",
"name": "qwen2.5-14b-128k",
"censor": false,
"charsPointsPrice": 0,
"name": "qwen2.5-14b-instruct",
"maxContext": 128000,
"maxResponse": 8000,

View File

@@ -1,10 +1,7 @@
{
"provider": "Qwen",
"model": "qwen2.5-32b-instruct",
"name": "qwen2.5-32b-128k",
"censor": false,
"charsPointsPrice": 0,
"name": "qwen2.5-32b-instruct",
"maxContext": 128000,
"maxResponse": 8000,

View File

@@ -1,10 +1,7 @@
{
"provider": "Qwen",
"model": "qwen2.5-72b-instruct",
"name": "Qwen2.5-72B-128k",
"censor": false,
"charsPointsPrice": 0,
"name": "Qwen2.5-72B-instruct",
"maxContext": 128000,
"maxResponse": 8000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -0,0 +1,26 @@
{
"provider": "Qwen",
"model": "qwen2.5-7b-instruct",
"name": "qwen2.5-7b-instruct",
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 50000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -0,0 +1,22 @@
{
"provider": "Siliconflow",
"model": "Qwen/Qwen2-VL-72B-Instruct",
"name": "Qwen/Qwen2-VL-72B-Instruct",
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"charsPointsPrice": 0,
"censor": false,
"vision": true,
"datasetProcess": false,
"usedInClassify": false,
"usedInExtractFields": false,
"usedInToolCall": false,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": "",
"defaultConfig": {}
}

View File

@@ -0,0 +1,26 @@
{
"provider": "Siliconflow",
"model": "deepseek-ai/DeepSeek-V2.5",
"name": "deepseek-ai/DeepSeek-V2.5",
"maxContext": 32000,
"maxResponse": 4000,
"quoteMaxToken": 32000,
"maxTemperature": 1,
"vision": true,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -0,0 +1,26 @@
{
"provider": "Siliconflow",
"model": "Qwen/Qwen2.5-72B-Instruct",
"name": "Qwen/Qwen2.5-72B-Instruct",
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 50000,
"maxTemperature": 1,
"vision": false,
"toolChoice": true,
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,
"usedInQueryExtension": true,
"customExtractPrompt": "",
"usedInToolCall": true,
"defaultConfig": {},
"fieldMap": {}
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1-128k",
"name": "step-1-128k",
"maxContext": 128000,
"maxResponse": 8000,
"quoteMaxToken": 128000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1-256k",
"name": "step-1-256k",
"maxContext": 256000,
"maxResponse": 8000,
"quoteMaxToken": 256000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1-32k",
"name": "step-1-32k",
"maxContext": 32000,
"maxResponse": 8000,
"quoteMaxToken": 32000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1-8k",
"name": "step-1-8k",
"maxContext": 8000,
"maxResponse": 8000,
"quoteMaxToken": 8000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1-flash",
"name": "step-1-flash",
"maxContext": 8000,
"maxResponse": 4000,
"quoteMaxToken": 6000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1o-vision-32k",
"name": "step-1o-vision-32k",
"maxContext": 32000,
"quoteMaxToken": 32000,
"maxResponse": 8000,
"maxTemperature": 2,
"vision": true,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1v-32k",
"name": "step-1v-32k",
"maxContext": 32000,
"quoteMaxToken": 32000,
"maxResponse": 8000,
"maxTemperature": 2,
"vision": true,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-1v-8k",
"name": "step-1v-8k",
"maxContext": 8000,
"maxResponse": 8000,
"quoteMaxToken": 8000,
"maxTemperature": 2,
"vision": true,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-2-16k-exp",
"name": "step-2-16k-exp",
"maxContext": 16000,
"maxResponse": 4000,
"quoteMaxToken": 4000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-2-16k",
"name": "step-2-16k",
"maxContext": 16000,
"maxResponse": 4000,
"quoteMaxToken": 4000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -0,0 +1,20 @@
{
"provider": "StepFun",
"model": "step-2-mini",
"name": "step-2-mini",
"maxContext": 8000,
"maxResponse": 4000,
"quoteMaxToken": 6000,
"maxTemperature": 2,
"vision": false,
"datasetProcess": true,
"usedInClassify": true,
"usedInExtractFields": true,
"usedInToolCall": true,
"usedInQueryExtension": true,
"toolChoice": false,
"functionCall": false,
"customCQPrompt": "",
"customExtractPrompt": "",
"defaultSystemChatPrompt": ""
}

View File

@@ -3,9 +3,6 @@
"model": "yi-lightning",
"name": "yi-lightning",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 16000,
"maxResponse": 4000,
"quoteMaxToken": 12000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -3,9 +3,6 @@
"model": "yi-vision-v2",
"name": "yi-vision-v2",
"censor": false,
"charsPointsPrice": 0,
"maxContext": 16000,
"maxResponse": 4000,
"quoteMaxToken": 12000,
@@ -16,7 +13,7 @@
"functionCall": false,
"defaultSystemChatPrompt": "",
"datasetProcess": false,
"datasetProcess": true,
"usedInClassify": true,
"customCQPrompt": "",
"usedInExtractFields": true,

View File

@@ -0,0 +1,84 @@
{
"provider": "StepFun",
"model": "step-tts-mini",
"name": "step-tts-mini",
"charsPointsPrice": 0,
"voices": [
{
"label": "cixingnansheng",
"value": "cixingnansheng"
},
{
"label": "zhengpaiqingnian",
"value": "zhengpaiqingnian"
},
{
"label": "yuanqinansheng",
"value": "yuanqinansheng"
},
{
"label": "qingniandaxuesheng",
"value": "qingniandaxuesheng"
},
{
"label": "boyinnansheng",
"value": "boyinnansheng"
},
{
"label": "ruyananshi",
"value": "ruyananshi"
},
{
"label": "shenchennanyin",
"value": "shenchennanyin"
},
{
"label": "qinqienvsheng",
"value": "qinqienvsheng"
},
{
"label": "wenrounvsheng",
"value": "wenrounvsheng"
},
{
"label": "jilingshaonv",
"value": "jilingshaonv"
},
{
"label": "yuanqishaonv",
"value": "yuanqishaonv"
},
{
"label": "ruanmengnvsheng",
"value": "ruanmengnvsheng"
},
{
"label": "youyanvsheng",
"value": "youyanvsheng"
},
{
"label": "lengyanyujie",
"value": "lengyanyujie"
},
{
"label": "shuangkuaijiejie",
"value": "shuangkuaijiejie"
},
{
"label": "wenjingxuejie",
"value": "wenjingxuejie"
},
{
"label": "linjiajiejie",
"value": "linjiajiejie"
},
{
"label": "linjiameimei",
"value": "linjiameimei"
},
{
"label": "zhixingjiejie",
"value": "zhixingjiejie"
}
]
}

View File

@@ -14,6 +14,11 @@ import { debounce } from 'lodash';
type FolderBaseType = `${ModelTypeEnum}`;
/*
TODO: 分优先级读取:
1. 有外部挂载目录,则读取外部的
2. 没有外部挂载目录,则读取本地的。然后试图拉取云端的进行覆盖。
*/
export const loadSystemModels = async (init = false) => {
const getModelNameList = (base: FolderBaseType) => {
const currentFileUrl = new URL(import.meta.url);
@@ -46,9 +51,7 @@ export const loadSystemModels = async (init = false) => {
}
};
if (!init && global.systemModelList && global.systemModelList.length > 0) return;
const dbModels = await MongoSystemModel.find({}).lean();
if (!init && global.systemModelList) return;
global.systemModelList = [];
global.systemActiveModelList = [];
@@ -58,50 +61,58 @@ export const loadSystemModels = async (init = false) => {
global.sttModelMap = new Map<string, STTModelType>();
global.reRankModelMap = new Map<string, ReRankModelItemType>();
const baseList: FolderBaseType[] = [
ModelTypeEnum.llm,
ModelTypeEnum.embedding,
ModelTypeEnum.tts,
ModelTypeEnum.stt,
ModelTypeEnum.rerank
];
try {
const dbModels = await MongoSystemModel.find({}).lean();
// System model
await Promise.all(
baseList.map(async (base) => {
const modelList = getModelNameList(base);
const nameList = modelList.map((name) => `${base}/${name}`);
const baseList: FolderBaseType[] = [
ModelTypeEnum.llm,
ModelTypeEnum.embedding,
ModelTypeEnum.tts,
ModelTypeEnum.stt,
ModelTypeEnum.rerank
];
await Promise.all(
nameList.map(async (name) => {
const fileContent = (await import(`./${name}`))?.default as SystemModelItemType;
// System model
await Promise.all(
baseList.map(async (base) => {
const modelList = getModelNameList(base);
const nameList = modelList.map((name) => `${base}/${name}`);
const dbModel = dbModels.find((item) => item.model === fileContent.model);
await Promise.all(
nameList.map(async (name) => {
const fileContent = (await import(`./${name}`))?.default as SystemModelItemType;
const model: any = {
...fileContent,
...dbModel?.metadata,
type: dbModel?.metadata?.type || base,
isCustom: false
};
const dbModel = dbModels.find((item) => item.model === fileContent.model);
pushModel(model);
})
);
})
);
const model: any = {
...fileContent,
...dbModel?.metadata,
type: dbModel?.metadata?.type || base,
isCustom: false
};
// Custom model
dbModels.forEach((dbModel) => {
if (global.systemModelList.find((item) => item.model === dbModel.model)) return;
pushModel(model);
})
);
})
);
pushModel({
...dbModel.metadata,
isCustom: true
// Custom model
dbModels.forEach((dbModel) => {
if (global.systemModelList.find((item) => item.model === dbModel.model)) return;
pushModel({
...dbModel.metadata,
isCustom: true
});
});
});
console.log('Load models success', JSON.stringify(global.systemActiveModelList, null, 2));
console.log('Load models success', JSON.stringify(global.systemActiveModelList, null, 2));
} catch (error) {
console.error('Load models error', error);
// @ts-ignore
global.systemModelList = undefined;
}
};
export const watchSystemModelUpdate = () => {

View File

@@ -1,6 +1,7 @@
import { addLog } from '../../../common/system/log';
import { POST } from '../../../common/api/serverRequest';
import { getFirstReRankModel } from '../model';
import { getAxiosConfig } from '../config';
type PostReRankResponse = {
id: string;
@@ -24,9 +25,11 @@ export function reRankRecall({
return Promise.reject('no rerank model');
}
const { baseUrl, authorization } = getAxiosConfig({});
let start = Date.now();
return POST<PostReRankResponse>(
model.requestUrl,
model.requestUrl ? model.requestUrl : `${baseUrl}/v1/rerank`,
{
model: model.model,
query,
@@ -34,7 +37,7 @@ export function reRankRecall({
},
{
headers: {
Authorization: `Bearer ${model.requestAuth}`
Authorization: model.requestAuth ? model.requestAuth : authorization
},
timeout: 30000
}

View File

@@ -33,7 +33,18 @@ const DatasetDataTextSchema = new Schema({
});
try {
DatasetDataTextSchema.index({ teamId: 1, datasetId: 1, fullTextToken: 'text' });
DatasetDataTextSchema.index(
{ teamId: 1, datasetId: 1, fullTextToken: 'text' },
{
name: 'teamId_1_datasetId_1_fullTextToken_text',
default_language: 'none',
collation: {
locale: 'simple', // 使用简单匹配规则
strength: 2, // 忽略大小写
caseLevel: false // 进一步确保大小写不敏感
}
}
);
DatasetDataTextSchema.index({ dataId: 1 }, { unique: true });
} catch (error) {
console.log(error);

View File

@@ -281,11 +281,12 @@ async function filterDatasetQuote({
}) {
function getValue(item: SearchDataResponseItemType, index: number) {
return replaceVariable(quoteTemplate || Prompt_QuoteTemplateList[0].value, {
id: item.id,
q: item.q,
a: item.a,
updateTime: formatTime2YMDHM(item.updateTime),
source: item.sourceName,
sourceId: String(item.sourceId || 'UnKnow'),
sourceId: String(item.sourceId || ''),
index: index + 1
});
}

View File

@@ -279,6 +279,7 @@ export async function authDatasetData({
return {
...result,
datasetData: data
datasetData: data,
collection: result.collection
};
}