Training status (#4424)
* dataset data training state (#4311) * dataset data training state * fix * fix ts * fix * fix api format * fix * fix * perf: count training * format * fix: dataset training state (#4417) * fix * add test * fix * fix * fix test * fix test * perf: training count * count * loading status --------- Co-authored-by: heheer <heheer@sealos.io>
This commit is contained in:
@@ -28,9 +28,24 @@
|
||||
"custom_data_process_params_desc": "Customize data processing rules",
|
||||
"custom_split_sign_tip": "Allows you to chunk according to custom delimiters. \nUsually used for processed data, using specific separators for precise chunking. \nYou can use the | symbol to represent multiple splitters, such as: \".|.\" to represent a period in Chinese and English.\n\nTry to avoid using special symbols related to regular, such as: * () [] {}, etc.",
|
||||
"data_amount": "{{dataAmount}} Datas, {{indexAmount}} Indexes",
|
||||
"data_error_amount": "{{errorAmount}} Group training exception",
|
||||
"data_index_num": "Index {{index}}",
|
||||
"data_process_params": "Params",
|
||||
"data_process_setting": "Processing config",
|
||||
"dataset.Chunk_Number": "Block number",
|
||||
"dataset.Completed": "Finish",
|
||||
"dataset.Delete_Chunk": "delete",
|
||||
"dataset.Edit_Chunk": "edit",
|
||||
"dataset.Error_Message": "Report an error message",
|
||||
"dataset.No_Error": "No exception information yet",
|
||||
"dataset.Operation": "operate",
|
||||
"dataset.ReTrain": "Retrain",
|
||||
"dataset.Training Process": "Training status",
|
||||
"dataset.Training_Count": "{{count}} Group training",
|
||||
"dataset.Training_Errors": "Errors",
|
||||
"dataset.Training_QA": "{{count}} Group Q&A pair training",
|
||||
"dataset.Training_Status": "Training status",
|
||||
"dataset.Training_Waiting": "Need to wait for {{count}} group data",
|
||||
"dataset.Unsupported operation": "dataset.Unsupported operation",
|
||||
"dataset.no_collections": "No datasets available",
|
||||
"dataset.no_tags": "No tags available",
|
||||
@@ -82,6 +97,13 @@
|
||||
"preview_chunk_empty": "Unable to read the contents of the file",
|
||||
"preview_chunk_intro": "A total of {{total}} blocks, up to 10",
|
||||
"preview_chunk_not_selected": "Click on the file on the left to preview",
|
||||
"process.Auto_Index": "Automatic index generation",
|
||||
"process.Get QA": "Q&A extraction",
|
||||
"process.Image_Index": "Image index generation",
|
||||
"process.Is_Ready": "Ready",
|
||||
"process.Parsing": "Parsing",
|
||||
"process.Vectorizing": "Index vectorization",
|
||||
"process.Waiting": "Queue",
|
||||
"rebuild_embedding_start_tip": "Index model switching task has started",
|
||||
"rebuilding_index_count": "Number of indexes being rebuilt: {{count}}",
|
||||
"request_headers": "Request headers, will automatically append 'Bearer '",
|
||||
@@ -114,7 +136,10 @@
|
||||
"tag.total_tags": "Total {{total}} tags",
|
||||
"the_knowledge_base_has_indexes_that_are_being_trained_or_being_rebuilt": "The Dataset has indexes that are being trained or rebuilt",
|
||||
"total_num_files": "Total {{total}} files",
|
||||
"training.Error": "{{count}} Group exception",
|
||||
"training.Normal": "Normal",
|
||||
"training_mode": "Chunk mode",
|
||||
"training_ready": "{{count}} Group",
|
||||
"vector_model_max_tokens_tip": "Each chunk of data has a maximum length of 3000 tokens",
|
||||
"vllm_model": "Image understanding model",
|
||||
"website_dataset": "Website Sync",
|
||||
|
||||
Reference in New Issue
Block a user