perf: token slice

This commit is contained in:
archer
2023-05-04 18:02:41 +08:00
parent 2d74fa8e10
commit a34a2b622c
8 changed files with 31 additions and 37 deletions

View File

@@ -45,18 +45,13 @@ export const lafClaudChat = async ({
}
);
let responseText = '';
let totalTokens = 0;
if (!stream) {
responseText = lafResponse.data?.text || '';
}
const responseText = stream ? '' : lafResponse.data?.text || '';
return {
streamResponse: lafResponse,
responseMessages: messages.concat({ obj: ChatRoleEnum.AI, value: responseText }),
responseText,
totalTokens
totalTokens: 0
};
};
@@ -83,18 +78,15 @@ export const lafClaudStreamResponse = async ({
} catch (error) {
console.log('pipe error', error);
}
// count tokens
const finishMessages = prompts.concat({
obj: ChatRoleEnum.AI,
value: responseContent
});
const totalTokens = modelToolMap[ClaudeEnum.Claude].countTokens({
messages: finishMessages
});
return {
responseContent,
totalTokens,
totalTokens: 0,
finishMessages
};
} catch (error) {