fix(dataset): CELERY_BROKER uses amqp rabbitmq. When adding document segments in batches and uploading large files, the status will always remain stuck at "In batch processing" #22709 (#23038)

This commit is contained in:
zhaobingshuang
2025-07-28 14:24:13 +08:00
committed by GitHub
parent 3f8fb18c89
commit 5c5f61b2aa
6 changed files with 166 additions and 44 deletions

View File

@@ -17,7 +17,7 @@ import cn from '@/utils/classnames'
import Divider from '@/app/components/base/divider'
import Loading from '@/app/components/base/loading'
import { ToastContext } from '@/app/components/base/toast'
import type { ChunkingMode, ParentMode, ProcessMode } from '@/models/datasets'
import type { ChunkingMode, FileItem, ParentMode, ProcessMode } from '@/models/datasets'
import { useDatasetDetailContext } from '@/context/dataset-detail'
import FloatRightContainer from '@/app/components/base/float-right-container'
import useBreakpoints, { MediaType } from '@/hooks/use-breakpoints'
@@ -111,12 +111,10 @@ const DocumentDetail: FC<Props> = ({ datasetId, documentId }) => {
}
const { mutateAsync: segmentBatchImport } = useSegmentBatchImport()
const runBatch = async (csv: File) => {
const formData = new FormData()
formData.append('file', csv)
const runBatch = async (csv: FileItem) => {
await segmentBatchImport({
url: `/datasets/${datasetId}/documents/${documentId}/segments/batch_import`,
body: formData,
body: { upload_file_id: csv.file.id! },
}, {
onSuccess: (res) => {
setImportStatus(res.job_status)