This commit is contained in:
Archer
2023-10-11 17:18:43 +08:00
committed by GitHub
parent d0041a98b4
commit bcf9491999
51 changed files with 852 additions and 460 deletions

View File

@@ -1,9 +1,6 @@
### Fast GPT V4.4.6
### Fast GPT V4.4.7
1. 高级编排新增模块 - 应用调用
2. 新增 - 必要连接校验
3. 新增 - 下一步指引选项,可以通过模型生成 3 个预测问题。
4. 新增 - 分享链接 hook 身份校验。
5. [使用文档](https://doc.fastgpt.run/docs/intro/)
6. [点击查看高级编排介绍文档](https://doc.fastgpt.run/docs/workflow)
7. [点击查看商业版](https://doc.fastgpt.run/docs/commercial/)
1. 优化数据集管理,区分手动录入和标注,可追数据至某个文件,保留链接读取的原始链接。
2. [使用文档](https://doc.fastgpt.run/docs/intro/)
3. [点击查看高级编排介绍文档](https://doc.fastgpt.run/docs/workflow)
4. [点击查看商业版](https://doc.fastgpt.run/docs/commercial/)

View File

@@ -0,0 +1,187 @@
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE svg PUBLIC "-//W3C//DTD SVG 1.1//EN" "http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd">
<svg version="1.1" id="Layer_1" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" x="0px" y="0px" width="140px" height="140px" viewBox="0 0 140 140" enable-background="new 0 0 140 140" xml:space="preserve"> <image id="image0" width="140" height="140" x="0" y="0"
href="data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAIwAAACMCAYAAACuwEE+AAAABGdBTUEAALGPC/xhBQAAACBjSFJN
AAB6JgAAgIQAAPoAAACA6AAAdTAAAOpgAAA6mAAAF3CculE8AAAABmJLR0QA/wD/AP+gvaeTAAAA
CXBIWXMAAAsTAAALEwEAmpwYAAAngklEQVR42u2deXxVxfn/33POXXNvdrIRIAsh7LJbEAEBWRUU
VxSxbkXU4tdS61YVUWtb90q1CtqqlbpQpQgqAhXckH0R2QmQQBayr3c/Z35/JEBCbpIbEHLp777/
4cU5c+Y+Z84nM8/MPDMDIUKECBEiRIgQIUKECBEiRIgQIUKECBEiRIgQIUKECBEiRIgQIUL876C2
tQFnmxdeeMFqUpQJfXr1GjD8kksKN23aVPNz5Lt69WpDdXn50At69rymS+fOpieffvrookWLZFu/
b4gzYM6cOZbM9PSls++9Vz419wnZt2fP3GGDB/c/43xnzYrokZHx0a3Tp+uvv/aavGX6dL1bevo/
Z82aZW7rdw5x+ohOSUkvvvrKK/I4hw8flj0yM4/OuvXWuNPNVMo5Sq9umYveffttWZ9X570i0zt2
nNfWLx3iNBk+ZMigyRMnenRdb/BhVyxfLpNiY9+SUorTyXdwv36TZtx++4lMdSllic8rV+cXyLGT
Jml9e/Ua0dbvHuI06JySsmj7tm3yVHy6LsdMutwxevTozNbmOWPGDGOHhIQt+fn5UkopHVKTL+ml
cqJ2VIZ9ukRaXpsv28XFrQ5EjKtXr7a0dRmFqGPMmDFpkydO9J0qFk1K+Re9VA5et1Qmt2//Zmvz
HdCz5yX33HmnJqWUxbpP3qEXyNHyqBwtj8rM3RskCz+U5gmX+YYMHDiouXyklMqkSZOuaOtyOh2U
tjbgbJC1e/fNN99yS6Me4EbpZJlwYBvUBz2j09TL7rorujX5FpaUXP2rmTMVH5I/Usoh4Ttxr12H
GIQi8F08TD2al3dnC1mJgwcP9mvrcjod/ucEM2fOHJPRaLzpiilTGlz3IJlPJTqAohB/xw22LRt/
uLk1+SYlJl7To2dP1kkX24SnwX2r3YItIgw9JZUym23KjBkzwprKa9KkScmO6qKbp02b1qGty6u1
/M8JZs3KlRdcPHx4msFgaHB9nXSSXa9GiJkyHmmzzH5Oz48PJN/lS5f2G3LxxYlGo5GPqPKbpl1y
NFII3H37R3+z9ptGzdILL7xgHTJkyGgLhz//5t9XpvmqdvzQtWvXJ+++++6Jp+uEn2v+5wSTc/jw
uOum3tCgOZLAYhqO1yl2G5YuaZ02HTzcUvMBQElR0SWjRo+mUPrYI7x+08Qm1bZw+oCB4lhV9VUA
UkpVyqxOevmS8dOmpHw093cjvkztYOt5NL+GtI5hHVZ9dNNjLz511QJZld+urcsuEAxnnkVwMf6y
yyyDLxrS4FqR9LHrlCYEIOrSYRxeseaO1VI+N1IIVzPZCofTOfqSUaP4SrrQm6gLImLtKIpAS2qP
FhE5uSx/dpVe8NT16K50oUYq8ZHxXDrqIkaMnMCI8b9l6cd/ITbaANXftddd3/8KeKaty68l/qcE
M3PmzPjIiIhh4eHhDa5/ixPNT/rwoYPIeeiPHTf+qnA4sKKpfGfMmGHdvnlTn/DwcDZR0uTvK6qC
LcpGVWk13ti4lPX0fWhsTLoqDPGgWIBapZmA226fQUzSYPBlg1YlEeacti6/QPifEkxMeHgnqy1s
uBANq4D1+K88TMmJ+CorRU5NzR3UE4yUUpVlX3VAO9BH+sp+cayoZvwDz3SJ8wnYL73N2hAVF0FV
aTW+9C5i5ZZSddzEkY0TSQ+jhyYhSueD51AWxrg/qnH3L4SH27oIW+S8F8xxZ1EIIRVVTRw3bnyD
+1VSZxd+PrLPR+5zr2O/sB95e/aOdbsLeqsVi7srWvkoPf/JUUhnijC0MwlrN3bnHKVbdwPX33Eb
NfOfAtG0f2qPrO0c6Z078+26H/BMvAoTAtCQ3nxwbAbHFk9qpPsH3WN/Q+XiJSJ+sgN+19ZFGRDn
tWBmTZsW0btr13marrtnzJhxz4fvv3/BkKFDG3zNbLy4hJ9JZIOBqvVb6fjg3exbtSayqk/Gxhh8
JsL6CMWcCcYkUGo//p6D88jo0oX3t64jUWm+MxMWaQVAj0+gfNW3HKvezYJn3iLCcIh7b0v7RjVZ
v1RNCZ+IdrP2CiHOu9nt81own3377X1XXX31qJycHOeKL74Y6Xa5+gwe0tDh3U/TTUjXj15Her3k
vrSATaYk87i4X/tNl52dzegxY3hw3DDeacYeo9TIsLgY2L6QkZEVDL0wnbVLn6W6UuA0Jsv+45d+
uXPv/jrH9t62Lr7T4rwSzKFDhyzz589PP7hvX2xMTIy2dMmSqbN/+9sF8+fPv+C7b765ICYyMtNi
tTZ4Zh+eJvM7+vTLRE8YhZSSHzUH4xSb33R5ubls27qVJ7d8Q8pzj564bpI+2vvK6e7JZ5D7MBne
QtzCwMpwI+/kxzL7nY2Ep7TnqWEj5Nhx43z/eu+Dq2696abPdCmtqR06FEy69trcgQMHNu8UBRnn
hWAeue++pBVr1tx3xeWX3dapc1psfGKCqHA56JieKqddf9VNnRJ1aTGbd+Tm5nbq37s3M2bOZPbv
an2Cfc3UMGq4HVdWNvZ+vdi0cwda35GoNGxypJQUFRaiax6S25kZ7DpIF28hPT25pHpLKFbt7DMm
8I01k79EXUqlYmXLU09jW76KUl1ikDrrvevEksWLjR3TUwc4dO9Wk8kk1m7drH+85D9H+/bs+cbI
sWNff+mll0rbupwDIehHF8eOHHvRwUP7Pp79xO8TL792CtawkyPuUtc5kp3DC088zerPvjw2fuLE
hCef+QOz7rqbxUuX4kMyWebhaeItC//xIb7ySmwXdMe7cx+rZj1CO6GC9CG1cvDmorsOs+n7JVzQ
M4ESCdlRHfnJlMxOU3tyDDHUKGbkKcV4YPxNrHv7fWbddRerNq1n0phLue/Rh+iYmoJQTo6VVlVW
8u93F/K3Z186lJnRZdKXq1fvbOvybomgrmFGDRvWv6i04PNFa5ZHdkjpRCE6n0gHU4SVMARCUeiU
lsrLby/gP+9/lLDorXc5dPAQkZGRABRJrUmxKEiio6yEVeTSs2sY4QdzkGXvI33HQK8BYQFjPBWV
ZtZuN7Juj4U3w80k3NnyJLOS0I683FzKysp48cXnmHTtVQ3uVyNZJl2Mjwjn1l/fxcWjRqbNuObG
VZePHTty2YoVe9q63JsjaGN6Z8yY0W7Lli0r3/50UWJK53QA3pMOvhRuVCnoKYwn0goh6N67Fz5N
44FZ93Hvb37DxvXrefHlF3H2SiMxykCGt5CB7mxGuPYxuWY7N1atZ3q7Y4ztrNI+UqF8Xw7e3sNJ
DRuGZh2JsA1HsfUnp8DIf1dvQQoTu2Jt2Pv18muva28WRx97Hil1ZEQkS373KLfP/jVXTL22UdoP
pZOlwoUmoa8wEhvXjgGDL7S//tK8YVNvvPHdzZs3B61fE5Q1jJRSZKalzf3dU4+lpWd2qb0G7Kjz
RwSwT/pYjJP/E3YsdU3CtbfcxBeLl/DRwlfJSHYyY1wnPnn+9/zu6UmUKDaOGSLJNsSw0ZxKoRrB
/hVf4nBpxN8xmax13xP+9d9ILapCURSQEkVVSUpKIjEpCaPFghpmbdLmgjfe451pv2LRRx/xr+ws
hvbqzpU3Xn/ivgPJq7Ka6wk7Mepcf3yo94B+3PPQb/u88dzLDwBPtPU3aIqg9GFGDh3aE5O67f2V
nxlU9WQl+L304EZysTDziKzgqND4g4ygszip++9Xf81bzz7Pkk/f44//+IgPdAcJM6fjd8BD1l0V
gsqvfyAhLY3/dOiHUjcw5/V4+PrrNdjCbHQcciG3KoVNDtrlv/J3/pRyAZdPnsykCROY/YfHuWDA
yZCXndLHU6KSLtLAoyKcr6SbKBSGCNOJND6vj8kXjahISEy+YNmyZUE5VRCMs9Vi3759v7n39w+d
EEuJ1NksvahAel2lGI6guzTQQTRsVftdOAinW8ejR/Ph/Ldpd8OVNDk6JsQJAUSMGIKzUyLFQkcI
gRACk9nMmLHjuOjii3EpNDvC227KBP7y6l8BKK+ooEef3g3upwuVVKliAlQE3TFiALZLL5V1FhqM
Bu595MHIHVu23N/WH6Epgq5JunL06JhCR821Q0eNwInkn9LBGuGuDXyqo4s0cL8IJwLBWukhHZWk
OuFYw6w43U7effttvNdOQIkMB5+Gt7AYrboG1R6GMb4dGPy/+rfSiYqgAA0TkISBbpioqmeBXuPA
V1iM1HQMsdGoUREYOyaRfUFn3v7737Haw6gfj5OHRp7U+ZOIpAydJ2QlB+rF5qjABGnhOmFl5MRx
RMZET502cuSjCxcurGzr73EqQSeYLbt2XT3r97+L8ArBn2Ule+oV7HH2Cx//lk4mYOZvoppe0shD
1M5Qi7pa4+8LFtB76d/Z9vjzOPdmYclMxxhhx1dVg3PfQcJ6ZJL8wF0IS8OlRG8olY1qJAUIlwqu
vQc5+sd5KCYTptRkhKriKyjCdTSfuOsn03HObB6/6EoGde3aME9ZQ5bwMY8oPpDOBmIB0IBlwkWe
1LjfHM6Uade3W/TWu+OARW39PU4lqAQzZ84c5R/z5984YcoVLJNOv2I5zm689MKAD4ir37JKqKmu
5vLLJvHBbx4n/uF7sPbp0ej5mnVb2H/7/XR+9Q+oURH1H2+EDhz57zcUffQpaS/OQY2JaphA0yh6
7xMO3TeHuIfvofLtxQ1ux6GwFzgsNX5qZiBxi/CyUrq57OopYv6L86YTEkzzfLV0aUJyaqdBtqRE
llLebFo3kv7CxOMynMx6XWy3201VRSXZeUfp9u4rOFVBxZdrKF22Cq2iCktmGu3/71fYBveno93G
0T+80mC43x+6w0nBW/8ic+FfQVFwbNtJwYJ/oVdVY+mcQvzN1xD3y2uJHDGE3EefpSo/r8Hzdwgb
w6SZDGHARfPzjZ/i5IW0FBKT2w99aOpU8//Nm+du6+9Sn6ByeguKi/sPHzM67Ce8OFuYyA1DYAR6
CGMD1efmHCElJYV58xeg+7xk3fUQnrxCUp55iIx3XiZ28jgO3Pkg0uPF2qsrnvxjoDf/WzVbdhBz
+aW1Ytm+i4I33iP1mQfJeOdl4m++htyX36RwwUJMqR1IemQWpUXFFBUcO/G8BUEfYcSCwNpC6G6J
0Nmp6lx48ZCYBStXnvGy3p+boBKM2+sd8othQ/mRlset2jcx5vj9V2u45dbbkAaVrAefIXHGNOJu
vQ7Hzr0cvPsRhKoQMXQQ7oPZAAihQAt/9ejyxJB+0XsfkzL3forfW8yRx59HjY4k7eUn8eQVUv3D
Ziw9uhB/721s+G5to2xUICWASn0XPoaPGU1hQcHIFhOfY4JKMKrBMLB3/74cxtdi2oGYGl2TUvLl
f5YyfuJEHKVlaEhsA/sg3W7yXn6T1Ocfw9IjE+feAxhiY3DvP4ShXTQozReDrX8vSj/7CqTEkpFG
9dYdxN/zS+Kun8zh3z8LAtrfdweF79S6HAm3TeXD9973m1fvAARzFI1uvXuiaVqftv4mpxI0gpFS
iqiYqExbZARVATRHfer5LcfR3F6mT5tOXFwcNZWVKLG1UfzCYMAQGUHun17l4MyHCB/cH2EycuTp
l+nw8K9pCcVuI3riKPJfWkDCrddT8ulKjjz6LPl/e5fIIQPq0oThq6xdfmJIaEfOhBHsLK1olNdg
YaqLwGuaEnSSkpOxhFkz58yZEzTfCILI6b3zzjut0e3atReqilfS7Bj0EGnC7mcQLdZsZ9r06QDs
bR9DzY+7kA4nIsxK+t/+iFZRhTAaqd64jf233Efqn3+PIb7h6g7P4SPorlo/UxgMmDungBC0u+EK
it5ZRNavH6XTnN+ghFkRJiOKrXb2vGrtJux96+aZhCD2lmt4et1e/jXiwgavEoVCf2lknWg6TseF
xGAyktShQ8e1a9dagZ9lT5ufg6ARzJGsrA4JSYlmARiaEUsYgitpPKfzw5pv2fHdBh55tLbHs8Lk
IeXpBzgw8yGs6Smo9jC0ympcObnYB/ah6/uvNRqDkU4XWwZdhq+kDADFbGLQge8xJieCEMTdch2R
l1xE7vOvoztdWFI6IAwGPPnHkKpCytMPnsjLaDay/d8fsi3SQr++FzT4nauxsglPkw3v8cY2PTMj
ovhQThQhwTQmNz8/vku/3nWDZILiJkRzlbQSJxrX0p8v+oTZ/zcbgELpY7fwYu2RSZd3XkavcaA7
XKj2MITV0uQQv7Ba6LXsHfZMvRtfVTWdn3u0Viz1MKV2IPWlJ0DT0Ktq0L1e1IhwhLmxT2UZPoR5
b73L3+c93+B6R6EySVpZLJx+7YitC+NKSEoy5Ow9EAPktvX3OU7QCMZoMiVFx8YggM4YOORnJdFA
aWKCaLxLhorg1Xmv1c4yA5txn+z3CIFit6HYbQSC/Rf9GLB7DdLrRQm3N51QVVGiIpp1Aq29u7HC
Z+GY002CtWFtNkVY2St97PKzirJTXQ8wKiZaON3u2HP2EQIgaBwqo6omRkTVBj4N9uMW9pVG7hV2
v53poiP5J8QC8LV0ciYIi7l5sQRIWIQVt5Qs3PJjo3sm4LfCTjfZ8G9WAS6sa5Ts4XYc1dWt2mHi
bBM0gsnPLTDY7LUfqYcwMlKaMUiIkQo3yjDuF+F+OtKg+Xz8+tY70PXayUEXkn3i9OKPvEfzKV+2
iqK/vUvJv/6DY+tO8Jx+LFNYhBUhJR//9gG83sb52BA8LCK4UloJlwJVwhhpIbMuXMNstaDrwtTa
3z2bBE2T5NO9GOtmeBVghrBxOzYUmg/a2bdrD5ldMk/UMNnSS5XQW/w9AKTEfSiHik9XEr9lD0Pt
0Vw8aBC9+l7CsaIi1i1Zy7o/v8ne5FjksEFEjh7aqprHaDaiGA0ciYoh78hRUtLTGqUxA1OFleuw
ogH1BwtU1YDJbIhpy+9yKkEjGIBj7oZLWgOJH/3+qzUMGz78xP9342lp3BY0jfLla/D8czFTklKZ
dsUVdJlyC/WX2HZKTmZQ377MAhxOJ9+sW8d7d81lY5SZdnfehLV3txZtUxSBalDwpndm43c/+BXM
ibQ0ru5LnC7cLpeLICJomqSoqCi50ldDdcufuwHbN23mwl/84sT/m1uHpJVVUPTGexSMnEq3f37O
xUY7T9x3H5np6YhmgqPCrFbGjxzJe8/8idlJGUTe/yxZ46ZR+vHnSGfT31MoAoNRRaaksHXj5la9
V6HLzSvbd6MoSlDFxASNYOx2u7ugqpIPpCPgZ6SU5Bw8THKH2o2cJLBT+hdMybv/JuHuJ1iQ2I3b
R4zmT889R2bfvhzJywv497w+Hxt37+HfixczIb4jT1aoGKffT+VX3/tNL4RANajose3YtWdv4O8F
PLFlF1UOB6rJ5An4wXNA0DRJPk3L16uqWS3cDJVmuouWTdN8GrqmYTLV+oVVUqdAadwdl14vaZ9/
zwfP1o6HLPvvV0RGRdGtR3c+W7WK4YMHc6y4mKLiYkoqyimvqR0ni7LZiY+JoX1CAokJCTgcDrr1
7oXFYkGqKpdfeiljR4xgyO/vh1FD/dqoGBQwmyksryBQ1uQX8Ul2LorLhc1oLG/rb1OfoBGM2Wwu
KS+rQAPepIY/EklL3QNd1zEaTrqJhfjw5+4Kg4GstASefvVVdCTLN63n63tmkIsPJTGOV3f9gDE5
AXP/jhjj+2GIrI3e81ZU4ck7hicnF8+a9XgPH8Wxaz8fLl2Ct7yClV9/TXZeHt4BvZq08XgEoMvX
8oQqQLnHy4Mbf0STEtXh0A1mc0lAD54jgkYwKhz1FtaWTa7Q+Kes4TZha7KH5AK2qBrFbhe6rqMo
Cnlo/j0gIYj7wwOszC9EGAyEz52JsJhJbcmm+HZYujR2VHWHE19BEQ/v3o+akkH80AubzEPqEjQN
p4RVeYUMS2iHWfXvCehS8tjmnRyqrm2W1coKzR4TE3jVdA4IGsEMHjGiOGv7Bi91PctVwo0u4ToR
RmSdbDTgkPTxPR6+xo3DILHcOBmHw4Hdbie3ubAIRWk0zH+6KGFWTOmdMKV3ajGt5tPB66W8/0Bu
WLOedmYTN3VO4arUZLpG2k8saSl0unlq224+OHTk5MNFRVUFBjUkGH8MGjTI8d6aFeXoehyKggT+
K9ysxUOKVAlDIQ8fhUJv0OzEzf4VUtbOGGdLb1CttJK6xOf1IcPC8F46FoBit4eXd+3nr7sPkBFh
p0uEnTK3h+2lFVTVa7aErmMoKyu89u67XHPnzm3rVzlB0Ajmuuuu0xIy0g9Ljzeu/iyyE9lsMHiK
NGCq+yvNF4H5CecKoQg6dEniwLbDje75pGRPRRV7Kvxv4SocDnDUZM+dOzeoXipoutUASlXNbm9+
YcDpY6TCU8RiRqAD5QQ4wnsO6dQ9mbiOrZ8/FKUlGBUlq63tP5WgEozQ9Z9qftwdUFozgseIIaGu
++1FNlhsFiwIAd0GdcZsbd2UkMjPw2A07m9r+08lqASjqOqO6u0tb5GiAPfKSHqL2qbLg+QVWUZl
kG4ZZ7IY6TowvXVlceAAcXFxG9ra9kZ2tbUB9YmMjd1fs+WnFtvsq6SNsaI2vkUCr8pyvhRnFtJw
tonrGEts+8AjFZTsw66+Awf+1NZ2N7KrrQ2oz9ixY/M8O/YW4tOaTDNEmpkhok50hj7QK/lMBD6d
0JZ07pOCUFruxgmPB2NJcdbChQurAsj2nBJUgnnppZecSk3Nek9uvt/7KdLAQyLmhNFbpIt3lKAr
0yYJj7aRmNLy6YFKcRFGp3MrLS6YOvcElWAADFJ+U71+a6Pr0VJhLrF1ETK180bPURbACqbgIrVX
B5QWahmxdw/hNtt/29pWfwSdYGxW2/LSL1Y36O4YgIeJpkNdj0gCb1JBUaCBUkFEWLi1RV9G3b/P
Gxsd/W1b2+qPoBNMevfuh1zbdh2jLuRSAWbJSPrXC/7Okh6+PE/8Fn906ta+yXvC68W0d2/RpGuu
OdKKLM8ZQSeY5cuXu7UjuT94c2rjVCZLGxPFybBICbxN5XnXFNUnsl0EVrv/M0KVvFy8JcXRn332
WUJb2+nXvrY2wB9hJssnxYu/YIA0MVNENpge2is9rBdBtQNGqxGKoH1n/3pQftyOHDHSeuzIkWlt
badf+9raAL9GqeqRjEqPfFTEYqgnFwksoir4ug6nQVyHGL9hodrgi/BeN5UKxXAdQTWVWktQCsbh
cFzym2GXivBTzKuUGus4v2uX44RFWLHazY2u63Fx6DYb7tSUnmPGjGm5D36OCTrBzJkzR4mMiJg8
dNiwRvdW48QdpMP/rUUIQWz7pleQaH37m/bu2DGure08laATTFZWljUtPb1z/dDL46zm/O0Z+eP4
oaL+0Dt3xuF0DmlFdueEoBPMj5s2JWdmZkae2noXSY29p7miMVgJj2l6vbeMi8NtMfckyI4nDpoA
quM4HI5BAwYNalRIe5vZHuNMUAAjArMEKwpWBEZE3aF7tZsvOtBxCYkHicbPN15vshgxh5lwOxqv
JJEGI1pS+549Xn3Ctguqz8KrnxZBJxin09mte4/G26Tu5MyX5wggTqqkY6QrRrpior0wEI6CWZwU
Sn21SkCjViwuJGVS5whe9uHlAF4O46Vc6Kctosh2ERTmFPu9p8fFR6Z0SO+2Czad7XIPlKATjGow
JCckNB6j2CHdp9XJtEpBL0xciIXBWEgUhlZlIwClrtaxAbFCJQMjx3crlEC+9LEBF9txswMP5a2Y
srBFNH3ghR4RYZBREVcQEkzTqKqaEh3d0BnUgZxWxuumSgNXYediYSXiLLpqAmgvDFyJnSux40ay
S7pZiZNvcba4fawlzNzkPWm1UeNyDjhrxp8GQeX0SimFqioRp57bWCY1HAF2pxVgugxnvkhgorCd
VbH4w4ygn7DwgIjmI5HEHTKi2U0FTGFNh25KqwWPrveRUgbNdwoaQ47bYzKaIo3Ghl3qY61wd/tK
M9PFuZaJf6wIpopwhsqmmx2TuZlK3mhE03ztnpXFgW2fdQ4IhnKtj9B03XDqkHlBKwTTFWPQvZS1
Ga9JNTQtGKGooEtTmdCCZsQ32MrWb9GWy8CdyHy0gNOeC3RgWzPTGara9CeQmoY0Gij16kGzqVDQ
Ccafp+JsRbdmNx68QTQ9WSB9FIpmRHxqP74+Xi9KmBVdiAiChGATjDSoqkfTGhZwawRQLDRKgqiW
2RXIjlhNJBAuJ8bIcJz4rC1lca4INsHoLper/NQNBAMo8hNowE8yePbg2UTzO47pWtPNrXA6UaMi
MSOC5vTfoBKMEEJKKQurqxquBAhr5YjdOoJjWzgfssURaq2ZJTWK04EaYceLEjQBhkElGACf13u0
rLyswTW1lYIJaGPEc0Cp1Jv3XwCvp2ktGB3VKGEW7ATPKr2gE4zb4zmcn9dwXZKtlYIpFBq5su3/
KLfhatGb8rj8z8ALKTGrEmE2g5RBszFi0AkmzGrd/dOOHQ2uRbdyoFMCa2n7P8otAUQHuh3+0win
E0tcNAhBvG4ImtV6QSeY5Pbtd+7Yvr2BJxgfwAaJpxLIxzqbuJBsCsAGZ3UTgikuwpqRitCl1sFk
Km4xo3NE0AkmKjGx4ODBg+VSnvRC4gPa4rkhu/BQ3YbbfxyQgc1aV5X5P9lGHD2KvW9PVKjpJyJD
NUxTLF261Hno4MED9cdiIoWKpZWBZw4h2SrbrpZZE0CTKHVJdbn/2CjlwD4ihgzAgMgbKIIn1DDo
BCOEkJUVFd/s/OnkThcKTR8K2hyr2igG2Ivk+wC69o4qJ7rmpz8nJebqCgyJ8VjhUJu8RBMEnWAA
jAbDmv+uXNngWmaLu/Y2ZoNwUdWKeaifi+3STZFoebS5stR/7SJcTsKTYkERqCjbzvkLNENQCiYj
LW3zd99+26Aa7oyx1fl4gRVtUMssDfDEvZLcMr/XlUOHiB5ee36CXYqtAWV2jghKwazZsOHYTzt2
bKisPDn8kHEaggFYieOcur5HpY/1ouXmSNd1yov8D6+o+/cSOeoihMQXpyghwbSEEEKWlpX9Z+OG
k1u8dcKI6TRWXGQJL1vluZsqWEp1QNE71WUO3C4/0wZSYi7IxZKRhkFSFIc7aM57hCAVDEByUtLy
he++e6JZihQKGbL14zES+Oc52l+zUGosDXAbkpL8Mr+z1EpZKdHp7RFmEwbY+lulU9uPQNa3r60N
aIop11+/Z82aNXt99XbH7i8sp5XXTuFh41muZSTwBhUBzWJJCccOF/m9p+zdS/SltcuETZLQLpqB
MnfuXJ/H7Xzr40WLTly7EPNp5SWB1yg/qwN5G6WLbwOcI6wur6Gm0n9aU+8exEweC1LKRF1dftYM
Pk2CVjAAXdMzFr2/cOGJKqaLMBF1mgH0uULjz7Ks/gHFPxvZ0stzlAUsxyZrl4ICYm0KSrgNA6Lk
EmNiUDm8EOSC+WrdutxNmzZ9eGB/7YbYRgQDTrOWAfhBuLhXFrFT/jyy8SL5XNZwryiiLMDFa7qm
k9+EYNQtm4iLqPXTrDobrhMieCLB6ghqwQBYzObXXnz++ROjYCM5s2jFLOHlPlHMA7KI1dJBBa1f
5lqOzueyhrtkIS+KcmpakUNRbikep38dmK6/mujLLwUgHPnBWSrSMyLoVj6eSv9f/GL9x4sWffPg
ww+PTElNpbcwEy4FVWewT4wEtgoPW/FgRZAiDfTBTCYmElGJQsGKggHwAFVoHENjH1424uKA8La4
otHv70rJkT3+z5hUKiqIXb8RJg1E6NIRrRhXtHXZ+yOotpJoim4pKYOHjhz53Zv/+IcK8LIsY9lZ
3EVT1PtX8vPt1lBeVMnmVTv8Zmj4eg29x/cnZsoE7Brf/UdNHi5E8O2eFPQ1DMCe7Ox1NZ9/vujH
7dunXtCnD5cSxmc4GpW7r6CI4o8/p2rdZtzZuSDAktKR8IsGEHv1RAxxgR1DI0/5t9VISfUPmyn5
dCU123ehVVZhjI3GmdgBtXNPtI4dGz1i3b+L6L/cD4AZ8VEwigXOkxoGoF+PHhmx8fFbPl+xIlwY
Ddwhj3G0boLPtXs/R56ZR+GHS5FNHMYpTEYSbriSjo/MwtwlrTU/HThSUrbsv2Q/9izVTR7jI5Dd
u6Ndez1aZlcA1OIiOvy0loz5z2KQsnqU09L1QVtc4Ocjn0OCZvlCSxQUFZW6qqvdVqt13ODBg/FK
ySathvwX3mD3DfdQvfUnjm8G7RdNp3r7Lgr+/gHGCDv2AX0ggIMiAkUrr+TArbM5/PhzeI4VNZtW
FBehfPctitOJ7NET9csv6DJzKqZOyYRJVvzVnDS/rcu7Sdvb2oDWMGPGDOPypUtXfr5ixQhbchLD
bp1G0aen4RsKQfzUyWQseA4l7MzXiHlyctl52c3U7NwXwE8L6kcT0qs3VquBAT8sAYHspKmX/cPY
/otzWa6tIei71fWZP3++NyYq6lczbr29ZOqYcQ3EIoTgrnvuQSgNX+mqa65pnJGUFL6/hH03zQLt
zFZJeguK+HHkdY3EYrFYGD9xYqP0jzz2GDEx9ZZK/7QDg+4DXcMsRfZ41bemrcu5Oc6bJuk4BcXF
pdLntRzMyhp56r2c7Gxqqk8GJQkh8Hq9lJaW+s3LsfcgUcMHY05r+Tjhpsh/aQHFixuP4Ou6TmRU
FHm5DSebd+7Ygcvlov7qTk9+IfYLetCpW+bjj6sdv2/rMm6O86qGOY7D4WhULdz8y18eE4rS4Prt
M2agKM28opQ4D2afkS2uHP++qa7ruJxOrps6tcH16upqXn/zzcbv9Mnylb9R279+7krx9DgvBRNu
s62FhlM3GV27Hq2prm7QFR04cCBlTdQuUNtzirxo0BnZYu/Xs8l7LpdLdu/Ro4FNuq5jNBrlqXvg
mDfvnBdMwd5NcV4K5paZM7+OjIp6VYiTG98dyclp37tPnwbLBLKysug/cGCT+cRcOgzd68VzMAdv
/jF8xaVo5RVolVVoldVo5ZVoJWV4jxXhOZKHa28Wjs07qFj1HSUffopj60/E3XAlltSOfvMfcckl
+pYtWxpc65KZSWFhoTju+AohtPDIyLe379mzrK3LNRDOq15SfaSUYtzo0YN27tjRCSAjI8MwdsIE
5bXX5nnMBqOhKsJm7jhuVOeBt0+/f8POHQ26QkKtdd3cecfwlZSiV1Sj1dSg1TiQbje6T0ciUVQV
xWhAWMNQbVYUWxhqhB1DTBSmhDjs/XtjSu2AdHuo+m4DruyG/soVl4zWv/j+G8Xj8Ui83mo9r3Dr
9b36/ZT16Zd527ZtE1632xsdG7mlfaf0rxYtWhQ8e5Q0w3krmECQUor7ZX73Ql2/ulww0gfdvYJ2
Upz9EW5Vx2WALAtifbxu+DzBp2+etDM3d+DAgUHf7DTH/7RgTmW1lIYNsihtq+7p5xGil4RuHmRH
IMUrpF2XmBEIKUTLEecSBNIHuA2SahWyjCg5iuTHCJ++fYwncsO+8PDSueI8PGewGf6/Eow/pJRi
DZjXVBwJq4wIs8Z79fhKQUS58BgrkWZFKJEuzWcUqiIFis+gKBX49EqzQTijPHolZlGaLLWa2aKj
K1jnf0KECBEiRIgQIUKECBEiRIgQIUKECBEiRIgQIUKECBEiRIgQIUKECOGf/wcZdEkygdX0dgAA
ACV0RVh0ZGF0ZTpjcmVhdGUAMjAyMy0xMC0xMVQwNzozODozOSswMDowMJUcyLsAAAAldEVYdGRh
dGU6bW9kaWZ5ADIwMjMtMTAtMTFUMDc6Mzg6MzkrMDA6MDDkQXAHAAAAKHRFWHRkYXRlOnRpbWVz
dGFtcAAyMDIzLTEwLTExVDA3OjM4OjQwKzAwOjAwLAkdEgAAAABJRU5ErkJggg==" />
</svg>

After

Width:  |  Height:  |  Size: 14 KiB

View File

@@ -131,6 +131,7 @@
"Rename Success": "Rename Success",
"Search": "Search",
"Status": "Status",
"Unknow": "Unknow",
"Update Successful": "Update Successful",
"export": ""
},

View File

@@ -131,6 +131,7 @@
"Rename Success": "重命名成功",
"Search": "搜索",
"Status": "状态",
"Unknow": "未知",
"Update Successful": "更新成功",
"export": ""
},

View File

@@ -2,9 +2,9 @@ import { GET, POST, DELETE, PUT } from './request';
import type { AppSchema } from '@/types/mongoSchema';
import type { AppListItemType, AppUpdateParams } from '@/types/app';
import { RequestPaging } from '../types/index';
import type { Props as CreateAppProps } from '@/pages/api/app/create';
import { addDays } from 'date-fns';
import { GetAppChatLogsParams } from './request/app';
import type { CreateAppParams } from '@/types/app';
/**
* 获取模型列表
@@ -14,7 +14,7 @@ export const getMyModels = () => GET<AppListItemType[]>('/app/myApps');
/**
* 创建一个模型
*/
export const postCreateApp = (data: CreateAppProps) => POST<string>('/app/create', data);
export const postCreateApp = (data: CreateAppParams) => POST<string>('/app/create', data);
/**
* 根据 ID 删除模型

View File

@@ -1,5 +1,5 @@
import React, { useCallback, useMemo, useState } from 'react';
import { ModalBody, Box, useTheme } from '@chakra-ui/react';
import { ModalBody, Box, useTheme, Flex, Progress } from '@chakra-ui/react';
import { getDatasetDataItemById } from '@/api/core/dataset/data';
import { useLoading } from '@/hooks/useLoading';
import { useToast } from '@/hooks/useToast';
@@ -8,22 +8,19 @@ import { QuoteItemType } from '@/types/chat';
import MyIcon from '@/components/Icon';
import InputDataModal, { RawFileText } from '@/pages/kb/detail/components/InputDataModal';
import MyModal from '../MyModal';
import type { PgDataItemType } from '@/types/core/dataset/data';
import { useTranslation } from 'react-i18next';
import { useRouter } from 'next/router';
type SearchType = PgDataItemType & {
kb_id?: string;
};
const QuoteModal = ({
onUpdateQuote,
rawSearch = [],
onClose
}: {
onUpdateQuote: (quoteId: string, sourceText?: string) => Promise<void>;
rawSearch: SearchType[];
rawSearch: QuoteItemType[];
onClose: () => void;
}) => {
const { t } = useTranslation();
const theme = useTheme();
const router = useRouter();
const { toast } = useToast();
@@ -36,7 +33,7 @@ const QuoteModal = ({
* click edit, get new kbDataItem
*/
const onclickEdit = useCallback(
async (item: SearchType) => {
async (item: QuoteItemType) => {
if (!item.id) return;
try {
setIsLoading(true);
@@ -95,9 +92,30 @@ const QuoteModal = ({
_hover={{ '& .edit': { display: 'flex' } }}
overflow={'hidden'}
>
{item.source && !isShare && (
<RawFileText filename={item.source} fileId={item.file_id} />
{!isShare && (
<Flex alignItems={'center'} mb={1}>
<RawFileText
filename={item.source || t('common.Unknow') || 'Unknow'}
fileId={item.file_id}
/>
<Box flex={'1'} />
{item.score && (
<>
<Progress
mx={2}
w={['60px', '100px']}
value={item.score * 100}
size="sm"
borderRadius={'20px'}
colorScheme="gray"
border={theme.borders.base}
/>
<Box>{item.score.toFixed(4)}</Box>
</>
)}
</Flex>
)}
<Box>{item.q}</Box>
<Box>{item.a}</Box>
{item.id && !isShare && (

View File

@@ -102,7 +102,7 @@ const Layout = ({ children }: { children: JSX.Element }) => {
</>
)}
</Box>
<Loading loading={loading} />
<Loading loading={loading} zIndex={9999} />
</>
);
};

View File

@@ -4,16 +4,18 @@ import { Spinner, Flex, Box } from '@chakra-ui/react';
const Loading = ({
fixed = true,
text = '',
bg = 'rgba(255,255,255,0.5)'
bg = 'rgba(255,255,255,0.5)',
zIndex = 1000
}: {
fixed?: boolean;
text?: string;
bg?: string;
zIndex?: number;
}) => {
return (
<Flex
position={fixed ? 'fixed' : 'absolute'}
zIndex={1000}
zIndex={zIndex}
bg={bg}
top={0}
left={0}

View File

@@ -59,5 +59,5 @@ export enum OutLinkTypeEnum {
apikey = 'apikey'
}
export const HUMAN_ICON = `/icon/human.png`;
export const HUMAN_ICON = `/icon/human.svg`;
export const LOGO_ICON = `/icon/logo.svg`;

View File

@@ -1,4 +1,4 @@
import { SystemInputEnum } from '../app';
import { AppTypeEnum, SystemInputEnum } from '../app';
import { TaskResponseKeyEnum } from '../chat';
import {
FlowModuleTypeEnum,
@@ -575,12 +575,17 @@ export const ModuleTemplatesFlat = [
];
// template
export const appTemplates: (AppItemType & { avatar: string; intro: string })[] = [
export const appTemplates: (AppItemType & {
avatar: string;
intro: string;
type: `${AppTypeEnum}`;
})[] = [
{
id: 'simpleChat',
avatar: '/imgs/module/AI.png',
name: '简单的对话',
intro: '一个极其简单的 AI 对话应用',
type: AppTypeEnum.basic,
modules: [
{
moduleId: 'userGuide',
@@ -797,6 +802,7 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
avatar: '/imgs/module/db.png',
name: '知识库 + 对话引导',
intro: '每次提问时进行一次知识库搜索,将搜索结果注入 LLM 模型进行参考回答',
type: AppTypeEnum.basic,
modules: [
{
moduleId: 'userGuide',
@@ -811,7 +817,7 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
key: 'welcomeText',
type: 'input',
label: '开场白',
value: '你好,我是 laf 助手,有什么可以帮助你的么?',
value: '你好,我是知识库助手,请不要忘记选择知识库噢~',
connected: true
}
],
@@ -1162,6 +1168,7 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
avatar: '/imgs/module/userGuide.png',
name: '对话引导 + 变量',
intro: '可以在对话开始发送一段提示,或者让用户填写一些内容,作为本次对话的变量',
type: AppTypeEnum.basic,
modules: [
{
moduleId: 'userGuide',
@@ -1174,27 +1181,15 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
inputs: [
{
key: 'welcomeText',
type: 'input',
type: 'hidden',
label: '开场白',
value: '你好,我可以为你翻译各种语言,请告诉我你需要翻译成什么语言?',
connected: true
}
],
outputs: []
},
{
moduleId: 'variable',
name: '全局变量',
flowType: 'variable',
position: {
x: 444.0369195277651,
y: 1008.5185781784537
},
inputs: [
},
{
key: 'variables',
type: 'systemInput',
label: '变量输入',
type: 'hidden',
label: '对话框变量',
value: [
{
id: '35c640eb-cf22-431f-bb57-3fc21643880e',
@@ -1227,6 +1222,13 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
}
],
connected: true
},
{
key: 'questionGuide',
type: 'switch',
label: '问题引导',
value: false,
connected: true
}
],
outputs: []
@@ -1275,7 +1277,7 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
key: 'maxContext',
type: 'numberInput',
label: '最长记录数',
value: 10,
value: 2,
min: 0,
max: 50,
connected: true
@@ -1317,7 +1319,6 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
type: 'custom',
label: '对话模型',
value: 'gpt-3.5-turbo-16k',
list: [],
connected: true
},
{
@@ -1346,7 +1347,7 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
label: '回复上限',
value: 8000,
min: 100,
max: 16000,
max: 4000,
step: 50,
markList: [
{
@@ -1354,8 +1355,8 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
value: 100
},
{
label: '16000',
value: 16000
label: '4000',
value: 4000
}
],
connected: true
@@ -1364,11 +1365,28 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
key: 'systemPrompt',
type: 'textarea',
label: '系统提示词',
max: 300,
valueType: 'string',
description:
'模型固定的引导词,通过调整该内容,可以引导模型聊天方向。该内容会被固定在上下文的开头。可使用变量,例如 {{language}}',
placeholder:
'模型固定的引导词,通过调整该内容,可以引导模型聊天方向。该内容会被固定在上下文的开头。可使用变量,例如 {{language}}',
value: '请直接将我的问题翻译成{{language}},不需要回答问题。',
connected: true
},
{
key: 'quoteTemplate',
type: 'hidden',
label: '引用内容模板',
valueType: 'string',
value: '',
connected: true
},
{
key: 'quotePrompt',
type: 'hidden',
label: '引用内容提示词',
valueType: 'string',
value: '',
connected: true
},
@@ -1381,8 +1399,9 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
},
{
key: 'quoteQA',
type: 'target',
type: 'custom',
label: '引用内容',
description: "对象数组格式,结构:\n [{q:'问题',a:'回答'}]",
valueType: 'kb_quote',
connected: false
},
@@ -1406,8 +1425,9 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
{
key: 'answerText',
label: '模型回复',
description: '直接响应,无需配置',
type: 'hidden',
description: '将在 stream 回复完毕后触发',
valueType: 'string',
type: 'source',
targets: []
},
{
@@ -1417,6 +1437,14 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
valueType: 'boolean',
type: 'source',
targets: []
},
{
key: 'history',
label: '新的上下文',
description: '将本次回复内容拼接上历史记录,作为新的上下文返回',
valueType: 'chat_history',
type: 'source',
targets: []
}
]
}
@@ -1427,6 +1455,7 @@ export const appTemplates: (AppItemType & { avatar: string; intro: string })[] =
avatar: '/imgs/module/cq.png',
name: '问题分类 + 知识库',
intro: '先对用户的问题进行分类,再根据不同类型问题,执行不同的操作',
type: AppTypeEnum.advanced,
modules: [
{
moduleId: '7z5g5h',

View File

@@ -8,13 +8,17 @@ export const useLoading = (props?: { defaultLoading: boolean }) => {
({
loading,
fixed = true,
text = ''
text = '',
zIndex
}: {
loading?: boolean;
fixed?: boolean;
text?: string;
zIndex?: number;
}): JSX.Element | null => {
return isLoading || loading ? <LoadingComponent fixed={fixed} text={text} /> : null;
return isLoading || loading ? (
<LoadingComponent fixed={fixed} text={text} zIndex={zIndex} />
) : null;
},
[isLoading]
);

View File

@@ -15,16 +15,6 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse)
await connectToDatabase();
await authUser({ req, authRoot: true });
console.log('add index');
await PgClient.query(
`
ALTER TABLE modeldata
ALTER COLUMN source TYPE VARCHAR(256),
ALTER COLUMN file_id TYPE VARCHAR(256);
CREATE INDEX IF NOT EXISTS modelData_fileId_index ON modeldata (file_id);
`
);
console.log('index success');
console.log('count rows');
// 去重获取 fileId
const { rows } = await PgClient.query(`SELECT DISTINCT file_id
@@ -36,8 +26,6 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse)
await init(rows.slice(i, i + limit), initFileIds);
console.log(i);
}
console.log('filter success');
console.log('start update');
for (let i = 0; i < initFileIds.length; i++) {
await PgClient.query(`UPDATE ${PgDatasetTableName}
@@ -49,9 +37,11 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse)
const { rows: emptyIds } = await PgClient.query(
`SELECT id FROM ${PgDatasetTableName} WHERE file_id IS NULL OR file_id=''`
);
console.log('filter success');
console.log(emptyIds.length);
await delay(5000);
console.log('start update');
async function start(start: number) {
for (let i = start; i < emptyIds.length; i += limit) {
@@ -65,12 +55,6 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse)
start(i);
}
// await PgClient.query(
// `UPDATE ${PgDatasetTableName}
// SET file_id = '${DatasetSpecialIdEnum.manual}'
// WHERE file_id IS NULL OR file_id = ''`
// );
console.log('update success');
jsonRes(res, {

View File

@@ -4,17 +4,17 @@ import { jsonRes } from '@/service/response';
import { connectToDatabase } from '@/service/mongo';
import { authUser } from '@/service/utils/auth';
import { App } from '@/service/models/app';
import { AppModuleItemType } from '@/types/app';
export type Props = {
name: string;
avatar?: string;
modules: AppModuleItemType[];
};
import type { CreateAppParams } from '@/types/app';
import { AppTypeEnum } from '@/constants/app';
export default async function handler(req: NextApiRequest, res: NextApiResponse<any>) {
try {
const { name, avatar, modules } = req.body as Props;
const {
name = 'APP',
avatar,
type = AppTypeEnum.advanced,
modules
} = req.body as CreateAppParams;
if (!name || !Array.isArray(modules)) {
throw new Error('缺少参数');
@@ -38,7 +38,8 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse<
avatar,
name,
userId,
modules
modules,
type
});
jsonRes(res, {

View File

@@ -9,7 +9,7 @@ import { authApp } from '@/service/utils/auth';
/* 获取我的模型 */
export default async function handler(req: NextApiRequest, res: NextApiResponse<any>) {
try {
const { name, avatar, type, chat, share, intro, modules } = req.body as AppUpdateParams;
const { name, avatar, type, share, intro, modules } = req.body as AppUpdateParams;
const { appId } = req.query as { appId: string };
if (!appId) {
@@ -37,7 +37,6 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse<
type,
avatar,
intro,
chat,
...(share && {
'share.isShare': share.isShare,
'share.isShareDetail': share.isShareDetail

View File

@@ -13,6 +13,7 @@ import { getVectorModel } from '@/service/utils/data';
import { getVector } from '@/pages/api/openapi/plugin/vector';
import { DatasetDataItemType } from '@/types/core/dataset/data';
import { countPromptTokens } from '@/utils/common/tiktoken';
import { authFileIdValid } from '@/service/dataset/auth';
export type Props = {
kbId: string;
@@ -72,6 +73,8 @@ export async function getVectorAndInsertDataset(
return Promise.reject('已经存在完全一致的数据');
}
await authFileIdValid(data.file_id);
const { vectors } = await getVector({
model: kb.vectorModel,
input: [q],

View File

@@ -1,16 +1,17 @@
/* push data to training queue */
import type { NextApiRequest, NextApiResponse } from 'next';
import { jsonRes } from '@/service/response';
import { connectToDatabase, TrainingData, KB } from '@/service/mongo';
import { authUser } from '@/service/utils/auth';
import { authKb } from '@/service/utils/auth';
import { withNextCors } from '@/service/utils/tools';
import { PgDatasetTableName, TrainingModeEnum } from '@/constants/plugin';
import { TrainingModeEnum } from '@/constants/plugin';
import { startQueue } from '@/service/utils/tools';
import { PgClient } from '@/service/pg';
import { getVectorModel } from '@/service/utils/data';
import { DatasetDataItemType } from '@/types/core/dataset/data';
import { countPromptTokens } from '@/utils/common/tiktoken';
import type { PushDataProps, PushDataResponse } from '@/api/core/dataset/data.d';
import { authFileIdValid } from '@/service/dataset/auth';
const modeMap = {
[TrainingModeEnum.index]: true,
@@ -80,69 +81,49 @@ export async function pushDataToKb({
[TrainingModeEnum.qa]: global.qaModel.maxToken * 0.8
};
// 过滤重复的 qa 内容
// filter repeat or equal content
const set = new Set();
const filterData: DatasetDataItemType[] = [];
const filterResult: Record<string, DatasetDataItemType[]> = {
success: [],
overToken: [],
fileIdInvalid: [],
error: []
};
data.forEach((item) => {
if (!item.q) return;
await Promise.all(
data.map(async (item) => {
if (!item.q) {
filterResult.error.push(item);
return;
}
const text = item.q + item.a;
const text = item.q + item.a;
// count q token
const token = countPromptTokens(item.q, 'system');
// count q token
const token = countPromptTokens(item.q, 'system');
if (token > modeMaxToken[mode]) {
return;
}
if (token > modeMaxToken[mode]) {
filterResult.overToken.push(item);
return;
}
if (!set.has(text)) {
filterData.push(item);
set.add(text);
}
});
try {
await authFileIdValid(item.file_id);
} catch (error) {
filterResult.fileIdInvalid.push(item);
return;
}
// 数据库去重
const insertData = (
await Promise.allSettled(
filterData.map(async (data) => {
let { q, a } = data;
if (mode !== TrainingModeEnum.index) {
return Promise.resolve(data);
}
if (!q) {
return Promise.reject('q为空');
}
q = q.replace(/\\n/g, '\n').trim().replace(/'/g, '"');
a = a.replace(/\\n/g, '\n').trim().replace(/'/g, '"');
// Exactly the same data, not push
try {
const { rows } = await PgClient.query(`
SELECT COUNT(*) > 0 AS exists
FROM ${PgDatasetTableName}
WHERE md5(q)=md5('${q}') AND md5(a)=md5('${a}') AND user_id='${userId}' AND kb_id='${kbId}'
`);
const exists = rows[0]?.exists || false;
if (exists) {
return Promise.reject('已经存在');
}
} catch (error) {
console.log(error);
}
return Promise.resolve(data);
})
)
)
.filter((item) => item.status === 'fulfilled')
.map<DatasetDataItemType>((item: any) => item.value);
if (!set.has(text)) {
filterResult.success.push(item);
set.add(text);
}
})
);
// 插入记录
const insertRes = await TrainingData.insertMany(
insertData.map((item) => ({
filterResult.success.map((item) => ({
...item,
userId,
kbId,
@@ -154,9 +135,11 @@ export async function pushDataToKb({
);
insertRes.length > 0 && startQueue();
delete filterResult.success;
return {
insertLen: insertRes.length
insertLen: insertRes.length,
...filterResult
};
}

View File

@@ -3,7 +3,7 @@ import { jsonRes } from '@/service/response';
import { connectToDatabase, TrainingData } from '@/service/mongo';
import { authUser } from '@/service/utils/auth';
import { GridFSStorage } from '@/service/lib/gridfs';
import { PgClient } from '@/service/pg';
import { PgClient, updateDataFileId } from '@/service/pg';
import { PgDatasetTableName } from '@/constants/plugin';
import { FileStatusEnum } from '@/constants/dataset';
import { strIsLink } from '@fastgpt/common/tools/str';
@@ -35,8 +35,8 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse<
.join(' ')}
${searchText ? `AND source ILIKE '%${searchText}%'` : ''}`;
const [{ rows }, { rowCount: total }] = await Promise.all([
PgClient.query(`SELECT file_id, COUNT(*) AS count
let [{ rows }, { rowCount: total }] = await Promise.all([
PgClient.query<{ file_id: string; count: number }>(`SELECT file_id, COUNT(*) AS count
FROM ${PgDatasetTableName}
where ${pgWhere}
GROUP BY file_id
@@ -49,6 +49,21 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse<
`)
]);
// If fileId is invalid, reset it to manual
await Promise.all(
rows.map((row) => {
if (!strIsLink(row.file_id) && row.file_id.length !== 24) {
return updateDataFileId({
oldFileId: row.file_id,
userId,
newFileId: DatasetSpecialIdEnum.manual
});
}
})
);
// just filter link or fileData
rows = rows.filter((row) => strIsLink(row.file_id) || row.file_id.length === 24);
// find files
const gridFs = new GridFSStorage('dataset', userId);
const collection = gridFs.Collection();
@@ -96,6 +111,7 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse<
const data = await Promise.all([
getSpecialData(),
...rows.map(async (row) => {
if (!row.file_id) return null;
// link data
if (strIsLink(row.file_id)) {
const { rows } = await PgClient.select(PgDatasetTableName, {

View File

@@ -47,7 +47,6 @@ export default withNextCors(async function handler(req: NextApiRequest, res: Nex
data: response?.[2]?.rows || []
});
} catch (err) {
console.log(err);
jsonRes(res, {
code: 500,
error: err

View File

@@ -2,7 +2,7 @@ import type { NextApiRequest, NextApiResponse } from 'next';
import { jsonRes } from '@/service/response';
import { authBalanceByUid, authUser } from '@/service/utils/auth';
import { withNextCors } from '@/service/utils/tools';
import { getAIChatApi, axiosConfig } from '@fastgpt/core/ai/config';
import { getAIApi } from '@fastgpt/core/ai/config';
import { pushGenerateVectorBill } from '@/service/common/bill/push';
type Props = {
@@ -54,29 +54,31 @@ export async function getVector({
}
// 获取 chatAPI
const chatAPI = getAIChatApi();
const ai = getAIApi();
// 把输入的内容转成向量
const result = await chatAPI
.createEmbedding(
const result = await ai.embeddings
.create(
{
model,
input
},
{
timeout: 60000,
...axiosConfig()
timeout: 60000
}
)
.then(async (res) => {
if (!res.data?.data?.[0]?.embedding) {
console.log(res.data);
if (!res.data) {
return Promise.reject('Embedding API 404');
}
if (!res?.data?.[0]?.embedding) {
console.log(res?.data);
// @ts-ignore
return Promise.reject(res.data?.err?.message || 'Embedding API Error');
}
return {
tokenLen: res.data.usage.total_tokens || 0,
vectors: await Promise.all(res.data.data.map((item) => unityDimensional(item.embedding)))
tokenLen: res.usage.total_tokens || 0,
vectors: await Promise.all(res.data.map((item) => unityDimensional(item.embedding)))
};
});

View File

@@ -5,7 +5,7 @@ import { User } from '@/service/models/user';
import { connectToDatabase } from '@/service/mongo';
import { authUser } from '@/service/utils/auth';
import { UserUpdateParams } from '@/types/user';
import { axiosConfig, getAIChatApi, openaiBaseUrl } from '@fastgpt/core/ai/config';
import { getAIApi, openaiBaseUrl } from '@fastgpt/core/ai/config';
/* update user info */
export default async function handler(req: NextApiRequest, res: NextApiResponse<any>) {
@@ -22,20 +22,15 @@ export default async function handler(req: NextApiRequest, res: NextApiResponse<
const baseUrl = openaiAccount?.baseUrl || openaiBaseUrl;
openaiAccount.baseUrl = baseUrl;
const chatAPI = getAIChatApi(openaiAccount);
const ai = getAIApi(openaiAccount);
const response = await chatAPI.createChatCompletion(
{
model: 'gpt-3.5-turbo',
max_tokens: 1,
messages: [{ role: 'user', content: 'hi' }]
},
{
...axiosConfig(openaiAccount)
}
);
if (response?.data?.choices?.[0]?.message?.content === undefined) {
throw new Error(JSON.stringify(response?.data));
const response = await ai.chat.completions.create({
model: 'gpt-3.5-turbo',
max_tokens: 1,
messages: [{ role: 'user', content: 'hi' }]
});
if (response?.choices?.[0]?.message?.content === undefined) {
throw new Error('Key response is empty');
}
}

View File

@@ -6,6 +6,7 @@ import dynamic from 'next/dynamic';
import { defaultApp } from '@/constants/model';
import { useToast } from '@/hooks/useToast';
import { useQuery } from '@tanstack/react-query';
import { feConfigs } from '@/store/static';
import Tabs from '@/components/Tabs';
import SideTabs from '@/components/SideTabs';
@@ -52,7 +53,9 @@ const AppDetail = ({ currentTab }: { currentTab: `${TabEnum}` }) => {
const tabList = useMemo(
() => [
{ label: '简易配置', id: TabEnum.basicEdit, icon: 'overviewLight' },
{ label: '高级编排', id: TabEnum.adEdit, icon: 'settingLight' },
...(feConfigs?.hide_app_flow
? []
: [{ label: '高级编排', id: TabEnum.adEdit, icon: 'settingLight' }]),
{ label: '外部使用', id: TabEnum.outLink, icon: 'shareLight' },
{ label: '对话日志', id: TabEnum.logs, icon: 'logsLight' },
{ label: '立即对话', id: TabEnum.startChat, icon: 'chat' }

View File

@@ -21,6 +21,7 @@ import { useRouter } from 'next/router';
import { appTemplates } from '@/constants/flow/ModuleTemplate';
import { useGlobalStore } from '@/store/global';
import { useRequest } from '@/hooks/useRequest';
import { feConfigs } from '@/store/static';
import Avatar from '@/components/Avatar';
import MyTooltip from '@/components/MyTooltip';
import MyModal from '@/components/MyModal';
@@ -74,10 +75,15 @@ const CreateModal = ({ onClose, onSuccess }: { onClose: () => void; onSuccess: (
const { mutate: onclickCreate, isLoading: creating } = useRequest({
mutationFn: async (data: FormType) => {
const template = appTemplates.find((item) => item.id === data.templateId);
if (!template) {
return Promise.reject('模板不存在');
}
return postCreateApp({
avatar: data.avatar,
name: data.name,
modules: appTemplates.find((item) => item.id === data.templateId)?.modules || []
type: template.type,
modules: template.modules || []
});
},
onSuccess(id: string) {
@@ -118,48 +124,52 @@ const CreateModal = ({ onClose, onSuccess }: { onClose: () => void; onSuccess: (
})}
/>
</Flex>
<Box mt={[4, 7]} mb={[0, 3]} color={'myGray.800'} fontWeight={'bold'}>
</Box>
<Grid
userSelect={'none'}
gridTemplateColumns={['repeat(1,1fr)', 'repeat(2,1fr)']}
gridGap={[2, 4]}
>
{appTemplates.map((item) => (
<Card
key={item.id}
border={theme.borders.base}
p={3}
borderRadius={'md'}
cursor={'pointer'}
boxShadow={'sm'}
{...(getValues('templateId') === item.id
? {
bg: 'myWhite.600'
}
: {
_hover: {
boxShadow: 'md'
}
})}
onClick={() => {
setValue('templateId', item.id);
setRefresh((state) => !state);
}}
{!feConfigs?.hide_app_flow && (
<>
<Box mt={[4, 7]} mb={[0, 3]} color={'myGray.800'} fontWeight={'bold'}>
</Box>
<Grid
userSelect={'none'}
gridTemplateColumns={['repeat(1,1fr)', 'repeat(2,1fr)']}
gridGap={[2, 4]}
>
<Flex alignItems={'center'}>
<Avatar src={item.avatar} borderRadius={'md'} w={'20px'} />
<Box ml={3} fontWeight={'bold'}>
{item.name}
</Box>
</Flex>
<Box fontSize={'sm'} mt={4}>
{item.intro}
</Box>
</Card>
))}
</Grid>
{appTemplates.map((item) => (
<Card
key={item.id}
border={theme.borders.base}
p={3}
borderRadius={'md'}
cursor={'pointer'}
boxShadow={'sm'}
{...(getValues('templateId') === item.id
? {
bg: 'myWhite.600'
}
: {
_hover: {
boxShadow: 'md'
}
})}
onClick={() => {
setValue('templateId', item.id);
setRefresh((state) => !state);
}}
>
<Flex alignItems={'center'}>
<Avatar src={item.avatar} borderRadius={'md'} w={'20px'} />
<Box ml={3} fontWeight={'bold'}>
{item.name}
</Box>
</Flex>
<Box fontSize={'sm'} mt={4}>
{item.intro}
</Box>
</Card>
))}
</Grid>
</>
)}
</ModalBody>
<ModalFooter>

View File

@@ -263,6 +263,10 @@ export function RawFileText({ fileId, filename = '', ...props }: RawFileTextProp
const { setLoading } = useGlobalStore();
const hasFile = useMemo(() => fileId && !datasetSpecialIds.includes(fileId), [fileId]);
const formatName = useMemo(
() => (filename.startsWith('kb') ? t(filename) : filename),
[filename, t]
);
return (
<MyTooltip label={hasFile ? t('file.Click to view file') || '' : ''} shouldWrapChildren={false}>
@@ -293,7 +297,7 @@ export function RawFileText({ fileId, filename = '', ...props }: RawFileTextProp
: {})}
{...props}
>
{t(filename)}
{formatName}
</Box>
</MyTooltip>
);

View File

@@ -0,0 +1,22 @@
import { isSpecialFileId } from '@fastgpt/core/dataset/utils';
import { GridFSStorage } from '../lib/gridfs';
import { Types } from 'mongoose';
export async function authFileIdValid(fileId?: string) {
if (!fileId) return true;
if (isSpecialFileId(fileId)) return true;
try {
// find file
const gridFs = new GridFSStorage('dataset', '');
const collection = gridFs.Collection();
const file = await collection.findOne(
{ _id: new Types.ObjectId(fileId) },
{ projection: { _id: 1 } }
);
if (!file) {
return Promise.reject('Invalid fileId');
}
} catch (error) {
return Promise.reject('Invalid fileId');
}
}

View File

@@ -17,19 +17,6 @@ export const TOKEN_ERROR_CODE: Record<number, string> = {
403: '登录状态无效,请重新登录'
};
export const openaiError: Record<string, string> = {
context_length_exceeded: '内容超长了,请重置对话',
Unauthorized: 'API-KEY 不合法',
rate_limit_reached: 'API被限制请稍后再试',
'Bad Request': 'Bad Request~ 可能内容太多了',
'Bad Gateway': '网关异常,请重试'
};
export const openaiAccountError: Record<string, string> = {
insufficient_quota: 'API 余额不足',
invalid_api_key: 'openai 账号异常',
account_deactivated: '账号已停用',
invalid_request_error: '无效请求'
};
export const proxyError: Record<string, boolean> = {
ECONNABORTED: true,
ECONNRESET: true

View File

@@ -4,7 +4,7 @@ import { TrainingModeEnum } from '@/constants/plugin';
import { ERROR_ENUM } from '../errorCode';
import { sendInform } from '@/pages/api/user/inform/send';
import { authBalanceByUid } from '../utils/auth';
import { axiosConfig, getAIChatApi } from '@fastgpt/core/ai/config';
import { getAIApi } from '@fastgpt/core/ai/config';
import type { ChatCompletionRequestMessage } from '@fastgpt/core/ai/type';
import { addLog } from '../utils/tools';
import { splitText2Chunks } from '@/utils/file';
@@ -58,8 +58,6 @@ export async function generateQA(): Promise<any> {
const startTime = Date.now();
const chatAPI = getAIChatApi();
// request LLM to get QA
const text = data.q;
const messages: ChatCompletionRequestMessage[] = [
@@ -73,19 +71,13 @@ export async function generateQA(): Promise<any> {
})
}
];
const { data: chatResponse } = await chatAPI.createChatCompletion(
{
model: global.qaModel.model,
temperature: 0.01,
messages,
stream: false
},
{
timeout: 480000,
...axiosConfig()
}
);
const ai = getAIApi(undefined, 480000);
const chatResponse = await ai.chat.completions.create({
model: global.qaModel.model,
temperature: 0.01,
messages,
stream: false
});
const answer = chatResponse.choices?.[0].message?.content;
const totalTokens = chatResponse.usage?.total_tokens || 0;

View File

@@ -23,7 +23,7 @@ const UserSchema = new Schema({
},
avatar: {
type: String,
default: '/icon/human.png'
default: '/icon/human.svg'
},
balance: {
type: Number,

View File

@@ -2,7 +2,7 @@ import { adaptChat2GptMessages } from '@/utils/common/adapt/message';
import { ChatContextFilter } from '@/service/common/tiktoken';
import type { ChatHistoryItemResType, ChatItemType } from '@/types/chat';
import { ChatRoleEnum, TaskResponseKeyEnum } from '@/constants/chat';
import { getAIChatApi, axiosConfig } from '@fastgpt/core/ai/config';
import { getAIApi } from '@fastgpt/core/ai/config';
import type { ClassifyQuestionAgentItemType } from '@/types/app';
import { SystemInputEnum } from '@/constants/app';
import { SpecialInputKeyEnum } from '@/constants/flow';
@@ -105,27 +105,22 @@ async function functionCall({
required: ['type']
}
};
const chatAPI = getAIChatApi(user.openaiAccount);
const ai = getAIApi(user.openaiAccount);
const response = await chatAPI.createChatCompletion(
{
model: cqModel.model,
temperature: 0,
messages: [...adaptMessages],
function_call: { name: agentFunName },
functions: [agentFunction]
},
{
...axiosConfig(user.openaiAccount)
}
);
const response = await ai.chat.completions.create({
model: cqModel.model,
temperature: 0,
messages: [...adaptMessages],
function_call: { name: agentFunName },
functions: [agentFunction]
});
try {
const arg = JSON.parse(response.data.choices?.[0]?.message?.function_call?.arguments || '');
const arg = JSON.parse(response.choices?.[0]?.message?.function_call?.arguments || '');
return {
arg,
tokens: response.data.usage?.total_tokens || 0
tokens: response.usage?.total_tokens || 0
};
} catch (error) {
console.log('Your model may not support function_call');
@@ -155,20 +150,14 @@ Human:${userChatInput}`
}
];
const chatAPI = getAIChatApi(user.openaiAccount);
const ai = getAIApi(user.openaiAccount, 480000);
const { data } = await chatAPI.createChatCompletion(
{
model: extractModel.model,
temperature: 0.01,
messages: adaptChat2GptMessages({ messages, reserveId: false }),
stream: false
},
{
timeout: 480000,
...axiosConfig(user.openaiAccount)
}
);
const data = await ai.chat.completions.create({
model: extractModel.model,
temperature: 0.01,
messages: adaptChat2GptMessages({ messages, reserveId: false }),
stream: false
});
const answer = data.choices?.[0].message?.content || '';
const totalTokens = data.usage?.total_tokens || 0;

View File

@@ -2,7 +2,7 @@ import { adaptChat2GptMessages } from '@/utils/common/adapt/message';
import { ChatContextFilter } from '@/service/common/tiktoken';
import type { ChatHistoryItemResType, ChatItemType } from '@/types/chat';
import { ChatRoleEnum, TaskResponseKeyEnum } from '@/constants/chat';
import { getAIChatApi, axiosConfig } from '@fastgpt/core/ai/config';
import { getAIApi } from '@fastgpt/core/ai/config';
import type { ContextExtractAgentItemType } from '@/types/app';
import { ContextExtractEnum } from '@/constants/flow/flowField';
import { FlowModuleTypeEnum } from '@/constants/flow';
@@ -126,30 +126,25 @@ async function functionCall({
}
};
const chatAPI = getAIChatApi(user.openaiAccount);
const ai = getAIApi(user.openaiAccount);
const response = await chatAPI.createChatCompletion(
{
model: extractModel.model,
temperature: 0,
messages: [...adaptMessages],
function_call: { name: agentFunName },
functions: [agentFunction]
},
{
...axiosConfig(user.openaiAccount)
}
);
const response = await ai.chat.completions.create({
model: extractModel.model,
temperature: 0,
messages: [...adaptMessages],
function_call: { name: agentFunName },
functions: [agentFunction]
});
const arg: Record<string, any> = (() => {
try {
return JSON.parse(response.data.choices?.[0]?.message?.function_call?.arguments || '{}');
return JSON.parse(response.choices?.[0]?.message?.function_call?.arguments || '{}');
} catch (error) {
return {};
}
})();
const tokens = response.data.usage?.total_tokens || 0;
const tokens = response.usage?.total_tokens || 0;
return {
tokens,
arg
@@ -181,20 +176,14 @@ Human: ${content}`
}
];
const chatAPI = getAIChatApi(user.openaiAccount);
const ai = getAIApi(user.openaiAccount, 480000);
const { data } = await chatAPI.createChatCompletion(
{
model: extractModel.model,
temperature: 0.01,
messages: adaptChat2GptMessages({ messages, reserveId: false }),
stream: false
},
{
timeout: 480000,
...axiosConfig(user.openaiAccount)
}
);
const data = await ai.chat.completions.create({
model: extractModel.model,
temperature: 0.01,
messages: adaptChat2GptMessages({ messages, reserveId: false }),
stream: false
});
const answer = data.choices?.[0].message?.content || '';
const totalTokens = data.usage?.total_tokens || 0;

View File

@@ -3,9 +3,9 @@ import { ChatContextFilter } from '@/service/common/tiktoken';
import type { ChatItemType, QuoteItemType } from '@/types/chat';
import type { ChatHistoryItemResType } from '@/types/chat';
import { ChatRoleEnum, sseResponseEventEnum } from '@/constants/chat';
import { SSEParseData, parseStreamChunk } from '@/utils/sse';
import { textAdaptGptResponse } from '@/utils/adapt';
import { getAIChatApi, axiosConfig } from '@fastgpt/core/ai/config';
import { getAIApi } from '@fastgpt/core/ai/config';
import type { ChatCompletion, StreamChatType } from '@fastgpt/core/ai/type';
import { TaskResponseKeyEnum } from '@/constants/chat';
import { getChatModel } from '@/service/utils/data';
import { countModelPrice } from '@/service/common/bill/push';
@@ -20,9 +20,7 @@ import type { AIChatProps } from '@/types/core/aiChat';
import { replaceVariable } from '@/utils/common/tools/text';
import { FlowModuleTypeEnum } from '@/constants/flow';
import type { ModuleDispatchProps } from '@/types/core/chat/type';
import { Readable } from 'stream';
import { responseWrite, responseWriteController } from '@/service/common/stream';
import { addLog } from '@/service/utils/tools';
export type ChatProps = ModuleDispatchProps<
AIChatProps & {
@@ -106,32 +104,25 @@ export const dispatchChatCompletion = async (props: ChatProps): Promise<ChatResp
// FastGPT temperature range: 1~10
temperature = +(modelConstantsData.maxTemperature * (temperature / 10)).toFixed(2);
temperature = Math.max(temperature, 0.01);
const chatAPI = getAIChatApi(user.openaiAccount);
const ai = getAIApi(user.openaiAccount, 480000);
const response = await chatAPI.createChatCompletion(
{
model,
temperature,
max_tokens,
messages: [
...(modelConstantsData.defaultSystem
? [
{
role: ChatCompletionRequestMessageRoleEnum.System,
content: modelConstantsData.defaultSystem
}
]
: []),
...messages
],
stream
},
{
timeout: 480000,
responseType: stream ? 'stream' : 'json',
...axiosConfig(user.openaiAccount)
}
);
const response = await ai.chat.completions.create({
model,
temperature,
max_tokens,
messages: [
...(modelConstantsData.defaultSystem
? [
{
role: ChatCompletionRequestMessageRoleEnum.System,
content: modelConstantsData.defaultSystem
}
]
: []),
...messages
],
stream
});
const { answerText, totalTokens, completeMessages } = await (async () => {
if (stream) {
@@ -139,7 +130,7 @@ export const dispatchChatCompletion = async (props: ChatProps): Promise<ChatResp
const { answer } = await streamResponse({
res,
detail,
response
stream: response
});
// count tokens
const completeMessages = filterMessages.concat({
@@ -159,8 +150,9 @@ export const dispatchChatCompletion = async (props: ChatProps): Promise<ChatResp
completeMessages
};
} else {
const answer = response.data.choices?.[0].message?.content || '';
const totalTokens = response.data.usage?.total_tokens || 0;
const unStreamResponse = response as ChatCompletion;
const answer = unStreamResponse.choices?.[0].message?.content || '';
const totalTokens = unStreamResponse.usage?.total_tokens || 0;
const completeMessages = filterMessages.concat({
obj: ChatRoleEnum.AI,
@@ -208,7 +200,7 @@ function filterQuote({
obj: ChatRoleEnum.System,
value: replaceVariable(quoteTemplate || defaultQuoteTemplate, {
...item,
index: `${index + 1}`
index: index + 1
})
}))
});
@@ -340,59 +332,40 @@ function targetResponse({
async function streamResponse({
res,
detail,
response
stream
}: {
res: NextApiResponse;
detail: boolean;
response: any;
stream: StreamChatType;
}) {
return new Promise<{ answer: string }>((resolve, reject) => {
const stream = response.data as Readable;
let answer = '';
const parseData = new SSEParseData();
const write = responseWriteController({
res,
readStream: stream
});
stream.on('data', (data) => {
if (res.closed) {
stream.destroy();
return resolve({ answer });
}
const parse = parseStreamChunk(data);
parse.forEach((item) => {
const { data } = parseData.parse(item);
if (!data || data === '[DONE]') return;
const content: string = data?.choices?.[0]?.delta?.content || '';
if (data.error) {
addLog.error(`SSE response`, data.error);
} else {
answer += content;
responseWrite({
write,
event: detail ? sseResponseEventEnum.answer : undefined,
data: textAdaptGptResponse({
text: content
})
});
}
});
});
stream.on('end', () => {
resolve({ answer });
});
stream.on('close', () => {
resolve({ answer });
});
stream.on('error', (err) => {
reject(err);
});
const write = responseWriteController({
res,
readStream: stream
});
let answer = '';
for await (const part of stream) {
if (res.closed) {
stream.controller?.abort();
break;
}
const content = part.choices[0]?.delta?.content || '';
answer += content;
responseWrite({
write,
event: detail ? sseResponseEventEnum.answer : undefined,
data: textAdaptGptResponse({
text: content
})
});
}
if (!answer) {
return Promise.reject('Chat API is error or undefined');
}
return { answer };
}
function getHistoryPreview(completeMessages: ChatItemType[]) {

View File

@@ -46,7 +46,9 @@ export async function dispatchKBSearch(props: Record<string, any>): Promise<KBSe
const res: any = await PgClient.query(
`BEGIN;
SET LOCAL ivfflat.probes = ${global.systemEnv.pgIvfflatProbe || 10};
select kb_id,id,q,a,source,file_id from ${PgDatasetTableName} where kb_id IN (${kbList
select id, kb_id, q, a, source, file_id, (vector <#> '[${
vectors[0]
}]') * -1 AS score from ${PgDatasetTableName} where kb_id IN (${kbList
.map((item) => `'${item.kbId}'`)
.join(',')}) AND vector <#> '[${vectors[0]}]' < -${similarity} order by vector <#> '[${
vectors[0]

View File

@@ -3,6 +3,7 @@ import type { QueryResultRow } from 'pg';
import { PgDatasetTableName } from '@/constants/plugin';
import { addLog } from './utils/tools';
import type { DatasetDataItemType } from '@/types/core/dataset/data';
import { DatasetSpecialIdEnum, datasetSpecialIdMap } from '@fastgpt/core/dataset/constant';
export const connectPg = async (): Promise<Pool> => {
if (global.pgClient) {
@@ -179,8 +180,13 @@ export const insertData2Dataset = ({
values: data.map((item) => [
{ key: 'user_id', value: userId },
{ key: 'kb_id', value: kbId },
{ key: 'source', value: item.source?.slice(0, 200)?.trim() || '' },
{ key: 'file_id', value: item.file_id?.slice(0, 200)?.trim() || '' },
{
key: 'source',
value:
item.source?.slice(0, 200)?.trim() ||
datasetSpecialIdMap[DatasetSpecialIdEnum.manual].sourceName
},
{ key: 'file_id', value: item.file_id?.slice(0, 200)?.trim() || DatasetSpecialIdEnum.manual },
{ key: 'q', value: item.q.replace(/'/g, '"') },
{ key: 'a', value: item.a.replace(/'/g, '"') },
{ key: 'vector', value: `[${item.vector}]` }
@@ -188,6 +194,25 @@ export const insertData2Dataset = ({
});
};
/**
* Update data file_id
*/
export const updateDataFileId = async ({
oldFileId,
userId,
newFileId = DatasetSpecialIdEnum.manual
}: {
oldFileId: string;
userId: string;
newFileId?: string;
}) => {
await PgClient.update(PgDatasetTableName, {
where: [['file_id', oldFileId], 'AND', ['user_id', userId]],
values: [{ key: 'file_id', value: newFileId }]
});
return newFileId;
};
export async function initPg() {
try {
await connectPg();
@@ -203,10 +228,6 @@ export async function initPg() {
q TEXT NOT NULL,
a TEXT
);
CREATE INDEX IF NOT EXISTS modelData_userId_index ON ${PgDatasetTableName} USING HASH (user_id);
CREATE INDEX IF NOT EXISTS modelData_kb_id_index ON ${PgDatasetTableName} (kb_id);
CREATE INDEX IF NOT EXISTS modelData_fileId_index ON ${PgDatasetTableName} (file_id);
CREATE INDEX IF NOT EXISTS idx_model_data_md5_q_a_user_id_kb_id ON ${PgDatasetTableName} (md5(q), md5(a), user_id, kb_id);
`);
console.log('init pg successful');
} catch (error) {

View File

@@ -1,12 +1,6 @@
import { sseResponseEventEnum } from '@/constants/chat';
import { NextApiResponse } from 'next';
import {
openaiError,
openaiAccountError,
proxyError,
ERROR_RESPONSE,
ERROR_ENUM
} from './errorCode';
import { proxyError, ERROR_RESPONSE, ERROR_ENUM } from './errorCode';
import { clearCookie, sseResponse, addLog } from './utils/tools';
export interface ResponseType<T = any> {
@@ -47,10 +41,8 @@ export const jsonRes = <T = any>(
msg = '网络连接异常';
} else if (error?.response?.data?.error?.message) {
msg = error?.response?.data?.error?.message;
} else if (openaiAccountError[error?.response?.data?.error?.code]) {
msg = openaiAccountError[error?.response?.data?.error?.code];
} else if (openaiError[error?.response?.statusText]) {
msg = openaiError[error.response.statusText];
} else if (error?.error?.message) {
msg = error?.error?.message;
}
addLog.error(`response error: ${msg}`, error);
@@ -88,10 +80,8 @@ export const sseErrRes = (res: NextApiResponse, error: any) => {
msg = '网络连接异常';
} else if (error?.response?.data?.error?.message) {
msg = error?.response?.data?.error?.message;
} else if (openaiAccountError[error?.response?.data?.error?.code]) {
msg = openaiAccountError[error?.response?.data?.error?.code];
} else if (openaiError[error?.response?.statusText]) {
msg = openaiError[error.response.statusText];
} else if (error?.error?.message) {
msg = error?.error?.message;
}
addLog.error(`sse error: ${msg}`, error);

View File

@@ -22,12 +22,17 @@ export type AppListItemType = {
intro: string;
};
export type CreateAppParams = {
name?: string;
avatar?: string;
type?: `${AppTypeEnum}`;
modules: AppSchema['modules'];
};
export interface AppUpdateParams {
name?: string;
type?: `${AppTypeEnum}`;
avatar?: string;
intro?: string;
chat?: AppSchema['chat'];
share?: AppSchema['share'];
modules?: AppSchema['modules'];
}

View File

@@ -45,6 +45,7 @@ export type ShareChatType = InitShareChatResponse & {
export type QuoteItemType = PgDataItemType & {
kb_id: string;
score?: number;
};
// response data

View File

@@ -3,7 +3,7 @@ import type { NextApiResponse } from 'next';
import { RunningModuleItemType } from '@/types/app';
import { UserModelSchema } from '@/types/mongoSchema';
export type MessageItemType = ChatCompletionRequestMessage & { dataId?: string };
export type MessageItemType = ChatCompletionRequestMessage & { dataId?: string; content: string };
// module dispatch props type
export type ModuleDispatchProps<T> = {

View File

@@ -29,6 +29,7 @@ export type FeConfigsType = {
show_pay?: boolean;
show_openai_account?: boolean;
show_promotion?: boolean;
hide_app_flow?: boolean;
openAPIUrl?: string;
systemTitle?: string;
authorText?: string;

View File

@@ -25,7 +25,7 @@ export const adaptBill = (bill: BillSchema): UserBillType => {
};
export const gptMessage2ChatType = (messages: MessageItemType[]): ChatItemType[] => {
const roleMap: Record<`${ChatCompletionRequestMessageRoleEnum}`, `${ChatRoleEnum}`> = {
const roleMap = {
[ChatCompletionRequestMessageRoleEnum.Assistant]: ChatRoleEnum.AI,
[ChatCompletionRequestMessageRoleEnum.User]: ChatRoleEnum.Human,
[ChatCompletionRequestMessageRoleEnum.System]: ChatRoleEnum.System,

View File

@@ -1,7 +1,7 @@
/*
replace {{variable}} to value
*/
export function replaceVariable(text: string, obj: Record<string, string>) {
export function replaceVariable(text: string, obj: Record<string, string | number>) {
for (const key in obj) {
const val = obj[key];
if (typeof val !== 'string') continue;

View File

@@ -11,7 +11,14 @@ export const splitText2Chunks = ({ text, maxLen }: { text: string; maxLen: numbe
const overlapLen = Math.floor(maxLen * 0.25); // Overlap length
try {
const splitTexts = text.split(/(?<=[。!?;.!?;\n])/g);
const tempMarker = 'SPLIT_HERE';
text = text.replace(/\n{3,}/g, '\n');
text = text.replace(/\s/g, ' ');
text = text.replace('\n\n', '');
const splitTexts = text
.replace(/([。!?;]|\.\s|!\s|\?\s|;\s|\n)/g, `$1${tempMarker}`)
.split(tempMarker)
.filter((part) => part);
const chunks: string[] = [];
let preChunk = '';

View File

@@ -9,7 +9,7 @@ export async function chunksUpload({
mode,
chunks,
prompt,
rate = 50,
rate = 150,
onUploading
}: {
kbId: string;