cecilia-uu
commited on
Commit
·
8337858
1
Parent(s):
fa34f3e
API: Stop parsing (#1556)
Browse files### What problem does this PR solve?
Aims to stop the process of parsing.
### Type of change
- [x] New Feature (non-breaking change which adds functionality)
- api/apps/dataset_api.py +129 -47
- api/db/services/document_service.py +12 -1
- docs/references/ragflow_api.md +1 -1
- sdk/python/ragflow/ragflow.py +11 -0
- sdk/python/test/test_document.py +119 -0
api/apps/dataset_api.py
CHANGED
|
@@ -16,6 +16,7 @@ import os
|
|
| 16 |
import pathlib
|
| 17 |
import re
|
| 18 |
import warnings
|
|
|
|
| 19 |
from io import BytesIO
|
| 20 |
|
| 21 |
from elasticsearch_dsl import Q
|
|
@@ -26,13 +27,12 @@ from httpx import HTTPError
|
|
| 26 |
from api.contants import NAME_LENGTH_LIMIT
|
| 27 |
from api.db import FileType, ParserType, FileSource, TaskStatus
|
| 28 |
from api.db import StatusEnum
|
| 29 |
-
from api.db.db_models import File
|
| 30 |
from api.db.services import duplicate_name
|
| 31 |
from api.db.services.document_service import DocumentService
|
| 32 |
from api.db.services.file2document_service import File2DocumentService
|
| 33 |
from api.db.services.file_service import FileService
|
| 34 |
from api.db.services.knowledgebase_service import KnowledgebaseService
|
| 35 |
-
from api.db.services.task_service import TaskService
|
| 36 |
from api.db.services.user_service import TenantService
|
| 37 |
from api.settings import RetCode
|
| 38 |
from api.utils import get_uuid
|
|
@@ -233,9 +233,10 @@ def update_dataset(dataset_id):
|
|
| 233 |
if chunk_num == 0:
|
| 234 |
dataset_updating_data["embd_id"] = req["embedding_model_id"]
|
| 235 |
else:
|
| 236 |
-
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 237 |
-
|
| 238 |
-
|
|
|
|
| 239 |
# only if chunk_num is 0, the user can update the chunk_method
|
| 240 |
if "chunk_method" in req:
|
| 241 |
type_value = req["chunk_method"]
|
|
@@ -614,35 +615,39 @@ def download_document(dataset_id, document_id):
|
|
| 614 |
|
| 615 |
# ----------------------------start parsing a document-----------------------------------------------------
|
| 616 |
# helper method for parsing
|
| 617 |
-
|
| 618 |
-
|
|
|
|
|
|
|
|
|
|
| 619 |
|
| 620 |
|
| 621 |
-
def doc_parse(binary, doc_name, parser_name, tenant_id):
|
| 622 |
match parser_name:
|
| 623 |
case "book":
|
| 624 |
-
book.chunk(doc_name, binary=binary, callback=
|
| 625 |
case "laws":
|
| 626 |
-
laws.chunk(doc_name, binary=binary, callback=
|
| 627 |
case "manual":
|
| 628 |
-
manual.chunk(doc_name, binary=binary, callback=
|
| 629 |
case "naive":
|
| 630 |
# It's the mode by default, which is general in the front-end
|
| 631 |
-
naive.chunk(doc_name, binary=binary, callback=
|
| 632 |
case "one":
|
| 633 |
-
one.chunk(doc_name, binary=binary, callback=
|
| 634 |
case "paper":
|
| 635 |
-
paper.chunk(doc_name, binary=binary, callback=
|
| 636 |
case "picture":
|
| 637 |
-
picture.chunk(doc_name, binary=binary, tenant_id=tenant_id, lang="Chinese",
|
|
|
|
| 638 |
case "presentation":
|
| 639 |
-
presentation.chunk(doc_name, binary=binary, callback=
|
| 640 |
case "qa":
|
| 641 |
-
qa.chunk(doc_name, binary=binary, callback=
|
| 642 |
case "resume":
|
| 643 |
-
resume.chunk(doc_name, binary=binary, callback=
|
| 644 |
case "table":
|
| 645 |
-
table.chunk(doc_name, binary=binary, callback=
|
| 646 |
case _:
|
| 647 |
return False
|
| 648 |
|
|
@@ -658,13 +663,8 @@ def parse_document(dataset_id, document_id):
|
|
| 658 |
if not exist:
|
| 659 |
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 660 |
message=f"This dataset '{dataset_id}' cannot be found!")
|
| 661 |
-
|
| 662 |
-
|
| 663 |
-
if isinstance(res, str):
|
| 664 |
-
message += res
|
| 665 |
-
return construct_json_result(code=RetCode.SUCCESS, message=message)
|
| 666 |
-
else:
|
| 667 |
-
return res
|
| 668 |
|
| 669 |
except Exception as e:
|
| 670 |
return construct_error_response(e)
|
|
@@ -680,34 +680,31 @@ def parse_documents(dataset_id):
|
|
| 680 |
if not exist:
|
| 681 |
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 682 |
message=f"This dataset '{dataset_id}' cannot be found!")
|
| 683 |
-
|
| 684 |
-
def process(doc_ids):
|
| 685 |
-
message = ""
|
| 686 |
-
# for loop
|
| 687 |
-
for id in doc_ids:
|
| 688 |
-
res = get_message_during_parsing_document(id, message)
|
| 689 |
-
if isinstance(res, str):
|
| 690 |
-
message += res
|
| 691 |
-
else:
|
| 692 |
-
return res
|
| 693 |
-
return construct_json_result(data=True, code=RetCode.SUCCESS, message=message)
|
| 694 |
-
|
| 695 |
# two conditions
|
| 696 |
-
if doc_ids:
|
| 697 |
-
return process(doc_ids)
|
| 698 |
-
else:
|
| 699 |
# documents inside the dataset
|
| 700 |
docs, total = DocumentService.list_documents_in_dataset(dataset_id, 0, -1, "create_time",
|
| 701 |
True, "")
|
| 702 |
doc_ids = [doc["id"] for doc in docs]
|
| 703 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 704 |
|
| 705 |
except Exception as e:
|
| 706 |
return construct_error_response(e)
|
| 707 |
|
| 708 |
|
| 709 |
-
# helper method for
|
| 710 |
-
def
|
|
|
|
| 711 |
try:
|
| 712 |
# Check whether there is this document
|
| 713 |
exist, document = DocumentService.get_by_id(id)
|
|
@@ -736,7 +733,7 @@ def get_message_during_parsing_document(id, message):
|
|
| 736 |
binary = MINIO.get(bucket, doc_name)
|
| 737 |
parser_name = doc_attributes["parser_id"]
|
| 738 |
if binary:
|
| 739 |
-
res = doc_parse(binary, doc_name, parser_name, tenant_id)
|
| 740 |
if res is False:
|
| 741 |
message += f"The parser id: {parser_name} of the document {doc_id} is not supported; "
|
| 742 |
else:
|
|
@@ -744,10 +741,94 @@ def get_message_during_parsing_document(id, message):
|
|
| 744 |
# failed in parsing
|
| 745 |
if doc_attributes["status"] == TaskStatus.FAIL.value:
|
| 746 |
message += f"Failed in parsing the document: {doc_id}; "
|
| 747 |
-
return message
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 748 |
except Exception as e:
|
| 749 |
return construct_error_response(e)
|
| 750 |
-
|
| 751 |
|
| 752 |
# ----------------------------show the status of the file-----------------------------------------------------
|
| 753 |
@manager.route("/<dataset_id>/documents/<document_id>/status", methods=["GET"])
|
|
@@ -774,6 +855,7 @@ def show_parsing_status(dataset_id, document_id):
|
|
| 774 |
)
|
| 775 |
except Exception as e:
|
| 776 |
return construct_error_response(e)
|
|
|
|
| 777 |
# ----------------------------list the chunks of the file-----------------------------------------------------
|
| 778 |
|
| 779 |
# -- --------------------------delete the chunk-----------------------------------------------------
|
|
|
|
| 16 |
import pathlib
|
| 17 |
import re
|
| 18 |
import warnings
|
| 19 |
+
from functools import partial
|
| 20 |
from io import BytesIO
|
| 21 |
|
| 22 |
from elasticsearch_dsl import Q
|
|
|
|
| 27 |
from api.contants import NAME_LENGTH_LIMIT
|
| 28 |
from api.db import FileType, ParserType, FileSource, TaskStatus
|
| 29 |
from api.db import StatusEnum
|
| 30 |
+
from api.db.db_models import File
|
| 31 |
from api.db.services import duplicate_name
|
| 32 |
from api.db.services.document_service import DocumentService
|
| 33 |
from api.db.services.file2document_service import File2DocumentService
|
| 34 |
from api.db.services.file_service import FileService
|
| 35 |
from api.db.services.knowledgebase_service import KnowledgebaseService
|
|
|
|
| 36 |
from api.db.services.user_service import TenantService
|
| 37 |
from api.settings import RetCode
|
| 38 |
from api.utils import get_uuid
|
|
|
|
| 233 |
if chunk_num == 0:
|
| 234 |
dataset_updating_data["embd_id"] = req["embedding_model_id"]
|
| 235 |
else:
|
| 236 |
+
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 237 |
+
message="You have already parsed the document in this "
|
| 238 |
+
"dataset, so you cannot change the embedding "
|
| 239 |
+
"model.")
|
| 240 |
# only if chunk_num is 0, the user can update the chunk_method
|
| 241 |
if "chunk_method" in req:
|
| 242 |
type_value = req["chunk_method"]
|
|
|
|
| 615 |
|
| 616 |
# ----------------------------start parsing a document-----------------------------------------------------
|
| 617 |
# helper method for parsing
|
| 618 |
+
# callback method
|
| 619 |
+
def doc_parse_callback(doc_id, prog=None, msg=""):
|
| 620 |
+
cancel = DocumentService.do_cancel(doc_id)
|
| 621 |
+
if cancel:
|
| 622 |
+
raise Exception("The parsing process has been cancelled!")
|
| 623 |
|
| 624 |
|
| 625 |
+
def doc_parse(binary, doc_name, parser_name, tenant_id, doc_id):
|
| 626 |
match parser_name:
|
| 627 |
case "book":
|
| 628 |
+
book.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 629 |
case "laws":
|
| 630 |
+
laws.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 631 |
case "manual":
|
| 632 |
+
manual.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 633 |
case "naive":
|
| 634 |
# It's the mode by default, which is general in the front-end
|
| 635 |
+
naive.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 636 |
case "one":
|
| 637 |
+
one.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 638 |
case "paper":
|
| 639 |
+
paper.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 640 |
case "picture":
|
| 641 |
+
picture.chunk(doc_name, binary=binary, tenant_id=tenant_id, lang="Chinese",
|
| 642 |
+
callback=partial(doc_parse_callback, doc_id))
|
| 643 |
case "presentation":
|
| 644 |
+
presentation.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 645 |
case "qa":
|
| 646 |
+
qa.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 647 |
case "resume":
|
| 648 |
+
resume.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 649 |
case "table":
|
| 650 |
+
table.chunk(doc_name, binary=binary, callback=partial(doc_parse_callback, doc_id))
|
| 651 |
case _:
|
| 652 |
return False
|
| 653 |
|
|
|
|
| 663 |
if not exist:
|
| 664 |
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 665 |
message=f"This dataset '{dataset_id}' cannot be found!")
|
| 666 |
+
|
| 667 |
+
return parsing_document_internal(document_id)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 668 |
|
| 669 |
except Exception as e:
|
| 670 |
return construct_error_response(e)
|
|
|
|
| 680 |
if not exist:
|
| 681 |
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 682 |
message=f"This dataset '{dataset_id}' cannot be found!")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 683 |
# two conditions
|
| 684 |
+
if not doc_ids:
|
|
|
|
|
|
|
| 685 |
# documents inside the dataset
|
| 686 |
docs, total = DocumentService.list_documents_in_dataset(dataset_id, 0, -1, "create_time",
|
| 687 |
True, "")
|
| 688 |
doc_ids = [doc["id"] for doc in docs]
|
| 689 |
+
|
| 690 |
+
message = ""
|
| 691 |
+
# for loop
|
| 692 |
+
for id in doc_ids:
|
| 693 |
+
res = parsing_document_internal(id)
|
| 694 |
+
res_body = res.json
|
| 695 |
+
if res_body["code"] == RetCode.SUCCESS:
|
| 696 |
+
message += res_body["message"]
|
| 697 |
+
else:
|
| 698 |
+
return res
|
| 699 |
+
return construct_json_result(data=True, code=RetCode.SUCCESS, message=message)
|
| 700 |
|
| 701 |
except Exception as e:
|
| 702 |
return construct_error_response(e)
|
| 703 |
|
| 704 |
|
| 705 |
+
# helper method for parsing the document
|
| 706 |
+
def parsing_document_internal(id):
|
| 707 |
+
message = ""
|
| 708 |
try:
|
| 709 |
# Check whether there is this document
|
| 710 |
exist, document = DocumentService.get_by_id(id)
|
|
|
|
| 733 |
binary = MINIO.get(bucket, doc_name)
|
| 734 |
parser_name = doc_attributes["parser_id"]
|
| 735 |
if binary:
|
| 736 |
+
res = doc_parse(binary, doc_name, parser_name, tenant_id, doc_id)
|
| 737 |
if res is False:
|
| 738 |
message += f"The parser id: {parser_name} of the document {doc_id} is not supported; "
|
| 739 |
else:
|
|
|
|
| 741 |
# failed in parsing
|
| 742 |
if doc_attributes["status"] == TaskStatus.FAIL.value:
|
| 743 |
message += f"Failed in parsing the document: {doc_id}; "
|
| 744 |
+
return construct_json_result(code=RetCode.SUCCESS, message=message)
|
| 745 |
+
except Exception as e:
|
| 746 |
+
return construct_error_response(e)
|
| 747 |
+
|
| 748 |
+
|
| 749 |
+
# ----------------------------stop parsing a doc-----------------------------------------------------
|
| 750 |
+
@manager.route("<dataset_id>/documents/<document_id>/status", methods=["DELETE"])
|
| 751 |
+
@login_required
|
| 752 |
+
def stop_parsing_document(dataset_id, document_id):
|
| 753 |
+
try:
|
| 754 |
+
# valid dataset
|
| 755 |
+
exist, _ = KnowledgebaseService.get_by_id(dataset_id)
|
| 756 |
+
if not exist:
|
| 757 |
+
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 758 |
+
message=f"This dataset '{dataset_id}' cannot be found!")
|
| 759 |
+
|
| 760 |
+
return stop_parsing_document_internal(document_id)
|
| 761 |
+
|
| 762 |
+
except Exception as e:
|
| 763 |
+
return construct_error_response(e)
|
| 764 |
+
|
| 765 |
+
|
| 766 |
+
# ----------------------------stop parsing docs-----------------------------------------------------
|
| 767 |
+
@manager.route("<dataset_id>/documents/status", methods=["DELETE"])
|
| 768 |
+
@login_required
|
| 769 |
+
def stop_parsing_documents(dataset_id):
|
| 770 |
+
doc_ids = request.json["doc_ids"]
|
| 771 |
+
try:
|
| 772 |
+
# valid dataset?
|
| 773 |
+
exist, _ = KnowledgebaseService.get_by_id(dataset_id)
|
| 774 |
+
if not exist:
|
| 775 |
+
return construct_json_result(code=RetCode.DATA_ERROR,
|
| 776 |
+
message=f"This dataset '{dataset_id}' cannot be found!")
|
| 777 |
+
if not doc_ids:
|
| 778 |
+
# documents inside the dataset
|
| 779 |
+
docs, total = DocumentService.list_documents_in_dataset(dataset_id, 0, -1, "create_time",
|
| 780 |
+
True, "")
|
| 781 |
+
doc_ids = [doc["id"] for doc in docs]
|
| 782 |
+
|
| 783 |
+
message = ""
|
| 784 |
+
# for loop
|
| 785 |
+
for id in doc_ids:
|
| 786 |
+
res = stop_parsing_document_internal(id)
|
| 787 |
+
res_body = res.json
|
| 788 |
+
if res_body["code"] == RetCode.SUCCESS:
|
| 789 |
+
message += res_body["message"]
|
| 790 |
+
else:
|
| 791 |
+
return res
|
| 792 |
+
return construct_json_result(data=True, code=RetCode.SUCCESS, message=message)
|
| 793 |
+
|
| 794 |
+
except Exception as e:
|
| 795 |
+
return construct_error_response(e)
|
| 796 |
+
|
| 797 |
+
|
| 798 |
+
# Helper method
|
| 799 |
+
def stop_parsing_document_internal(document_id):
|
| 800 |
+
try:
|
| 801 |
+
# valid doc?
|
| 802 |
+
exist, doc = DocumentService.get_by_id(document_id)
|
| 803 |
+
if not exist:
|
| 804 |
+
return construct_json_result(message=f"This document '{document_id}' cannot be found!",
|
| 805 |
+
code=RetCode.ARGUMENT_ERROR)
|
| 806 |
+
doc_attributes = doc.to_dict()
|
| 807 |
+
|
| 808 |
+
# only when the status is parsing, we need to stop it
|
| 809 |
+
if doc_attributes["status"] == TaskStatus.RUNNING.value:
|
| 810 |
+
tenant_id = DocumentService.get_tenant_id(document_id)
|
| 811 |
+
if not tenant_id:
|
| 812 |
+
return construct_json_result(message="Tenant not found!", code=RetCode.AUTHENTICATION_ERROR)
|
| 813 |
+
|
| 814 |
+
# update successfully?
|
| 815 |
+
if not DocumentService.update_by_id(document_id, {"status": "2"}): # cancel
|
| 816 |
+
return construct_json_result(
|
| 817 |
+
code=RetCode.OPERATING_ERROR,
|
| 818 |
+
message="There was an error during the stopping parsing the document process. "
|
| 819 |
+
"Please check the status of the RAGFlow server and try the update again."
|
| 820 |
+
)
|
| 821 |
+
|
| 822 |
+
_, doc_attributes = DocumentService.get_by_id(document_id)
|
| 823 |
+
doc_attributes = doc_attributes.to_dict()
|
| 824 |
+
|
| 825 |
+
# failed in stop parsing
|
| 826 |
+
if doc_attributes["status"] == TaskStatus.RUNNING.value:
|
| 827 |
+
return construct_json_result(message=f"Failed in parsing the document: {document_id}; ", code=RetCode.SUCCESS)
|
| 828 |
+
return construct_json_result(code=RetCode.SUCCESS, message="")
|
| 829 |
except Exception as e:
|
| 830 |
return construct_error_response(e)
|
| 831 |
+
|
| 832 |
|
| 833 |
# ----------------------------show the status of the file-----------------------------------------------------
|
| 834 |
@manager.route("/<dataset_id>/documents/<document_id>/status", methods=["GET"])
|
|
|
|
| 855 |
)
|
| 856 |
except Exception as e:
|
| 857 |
return construct_error_response(e)
|
| 858 |
+
|
| 859 |
# ----------------------------list the chunks of the file-----------------------------------------------------
|
| 860 |
|
| 861 |
# -- --------------------------delete the chunk-----------------------------------------------------
|
api/db/services/document_service.py
CHANGED
|
@@ -333,6 +333,17 @@ class DocumentService(CommonService):
|
|
| 333 |
cls.model.kb_id == kb_id).dicts())
|
| 334 |
|
| 335 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 336 |
def queue_raptor_tasks(doc):
|
| 337 |
def new_task():
|
| 338 |
nonlocal doc
|
|
@@ -347,4 +358,4 @@ def queue_raptor_tasks(doc):
|
|
| 347 |
task = new_task()
|
| 348 |
bulk_insert_into_db(Task, [task], True)
|
| 349 |
task["type"] = "raptor"
|
| 350 |
-
assert REDIS_CONN.queue_product(SVR_QUEUE_NAME, message=task), "Can't access Redis. Please check the Redis' status."
|
|
|
|
| 333 |
cls.model.kb_id == kb_id).dicts())
|
| 334 |
|
| 335 |
|
| 336 |
+
@classmethod
|
| 337 |
+
@DB.connection_context()
|
| 338 |
+
def do_cancel(cls, doc_id):
|
| 339 |
+
try:
|
| 340 |
+
_, doc = DocumentService.get_by_id(doc_id)
|
| 341 |
+
return doc.run == TaskStatus.CANCEL.value or doc.progress < 0
|
| 342 |
+
except Exception as e:
|
| 343 |
+
pass
|
| 344 |
+
return False
|
| 345 |
+
|
| 346 |
+
|
| 347 |
def queue_raptor_tasks(doc):
|
| 348 |
def new_task():
|
| 349 |
nonlocal doc
|
|
|
|
| 358 |
task = new_task()
|
| 359 |
bulk_insert_into_db(Task, [task], True)
|
| 360 |
task["type"] = "raptor"
|
| 361 |
+
assert REDIS_CONN.queue_product(SVR_QUEUE_NAME, message=task), "Can't access Redis. Please check the Redis' status."
|
docs/references/ragflow_api.md
CHANGED
|
@@ -758,7 +758,7 @@ This method enables a specific document to start parsing for a specific user.
|
|
| 758 |
```json
|
| 759 |
{
|
| 760 |
"code": 102,
|
| 761 |
-
"message": "This dataset 'imagination
|
| 762 |
}
|
| 763 |
```
|
| 764 |
|
|
|
|
| 758 |
```json
|
| 759 |
{
|
| 760 |
"code": 102,
|
| 761 |
+
"message": "This dataset 'imagination' cannot be found!"
|
| 762 |
}
|
| 763 |
```
|
| 764 |
|
sdk/python/ragflow/ragflow.py
CHANGED
|
@@ -157,6 +157,17 @@ class RAGFlow:
|
|
| 157 |
return res.json()
|
| 158 |
|
| 159 |
# ----------------------------stop parsing-----------------------------------------------------
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 160 |
|
| 161 |
# ----------------------------show the status of the file-----------------------------------------------------
|
| 162 |
def show_parsing_status(self, dataset_id, document_id):
|
|
|
|
| 157 |
return res.json()
|
| 158 |
|
| 159 |
# ----------------------------stop parsing-----------------------------------------------------
|
| 160 |
+
def stop_parsing_document(self, dataset_id, document_id):
|
| 161 |
+
endpoint = f"{self.dataset_url}/{dataset_id}/documents/{document_id}/status"
|
| 162 |
+
res = requests.delete(endpoint, headers=self.authorization_header)
|
| 163 |
+
|
| 164 |
+
return res.json()
|
| 165 |
+
|
| 166 |
+
def stop_parsing_documents(self, dataset_id, doc_ids=None):
|
| 167 |
+
endpoint = f"{self.dataset_url}/{dataset_id}/documents/status"
|
| 168 |
+
res = requests.delete(endpoint, headers=self.authorization_header, json={"doc_ids": doc_ids})
|
| 169 |
+
|
| 170 |
+
return res.json()
|
| 171 |
|
| 172 |
# ----------------------------show the status of the file-----------------------------------------------------
|
| 173 |
def show_parsing_status(self, dataset_id, document_id):
|
sdk/python/test/test_document.py
CHANGED
|
@@ -949,7 +949,126 @@ class TestFile(TestSdk):
|
|
| 949 |
# parse
|
| 950 |
res = ragflow.start_parsing_documents(created_res_id, doc_ids)
|
| 951 |
assert res["code"] == RetCode.SUCCESS and res["message"] == ""
|
|
|
|
| 952 |
# ----------------------------stop parsing-----------------------------------------------------
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 953 |
|
| 954 |
# ----------------------------show the status of the file-----------------------------------------------------
|
| 955 |
def test_show_status_with_success(self):
|
|
|
|
| 949 |
# parse
|
| 950 |
res = ragflow.start_parsing_documents(created_res_id, doc_ids)
|
| 951 |
assert res["code"] == RetCode.SUCCESS and res["message"] == ""
|
| 952 |
+
|
| 953 |
# ----------------------------stop parsing-----------------------------------------------------
|
| 954 |
+
def test_stop_parsing_document_with_success(self):
|
| 955 |
+
"""
|
| 956 |
+
Test the stopping parsing of a document with success.
|
| 957 |
+
"""
|
| 958 |
+
# create a dataset
|
| 959 |
+
ragflow = RAGFlow(API_KEY, HOST_ADDRESS)
|
| 960 |
+
created_res = ragflow.create_dataset("test_start_parsing_document_with_success")
|
| 961 |
+
created_res_id = created_res["data"]["dataset_id"]
|
| 962 |
+
# upload files
|
| 963 |
+
file_paths = ["test_data/lol.txt"]
|
| 964 |
+
uploading_res = ragflow.upload_local_file(created_res_id, file_paths)
|
| 965 |
+
# get the doc_id
|
| 966 |
+
data = uploading_res["data"][0]
|
| 967 |
+
doc_id = data["id"]
|
| 968 |
+
# parse file
|
| 969 |
+
res = ragflow.start_parsing_document(created_res_id, doc_id)
|
| 970 |
+
assert res["code"] == RetCode.SUCCESS and res["message"] == ""
|
| 971 |
+
res = ragflow.stop_parsing_document(created_res_id, doc_id)
|
| 972 |
+
assert res["code"] == RetCode.SUCCESS and res["message"] == ""
|
| 973 |
+
|
| 974 |
+
def test_stop_parsing_nonexistent_document(self):
|
| 975 |
+
"""
|
| 976 |
+
Test the stopping parsing a document which does not exist.
|
| 977 |
+
"""
|
| 978 |
+
# create a dataset
|
| 979 |
+
ragflow = RAGFlow(API_KEY, HOST_ADDRESS)
|
| 980 |
+
created_res = ragflow.create_dataset("test_start_parsing_nonexistent_document")
|
| 981 |
+
created_res_id = created_res["data"]["dataset_id"]
|
| 982 |
+
res = ragflow.stop_parsing_document(created_res_id, "imagination.txt")
|
| 983 |
+
assert res["code"] == RetCode.ARGUMENT_ERROR and res["message"] == "This document 'imagination.txt' cannot be found!"
|
| 984 |
+
|
| 985 |
+
def test_stop_parsing_document_in_nonexistent_dataset(self):
|
| 986 |
+
"""
|
| 987 |
+
Test the stopping parsing a document whose dataset is nonexistent.
|
| 988 |
+
"""
|
| 989 |
+
# create a dataset
|
| 990 |
+
ragflow = RAGFlow(API_KEY, HOST_ADDRESS)
|
| 991 |
+
created_res = ragflow.create_dataset("test_download_nonexistent_document")
|
| 992 |
+
created_res_id = created_res["data"]["dataset_id"]
|
| 993 |
+
# upload files
|
| 994 |
+
file_paths = ["test_data/test.txt"]
|
| 995 |
+
uploading_res = ragflow.upload_local_file(created_res_id, file_paths)
|
| 996 |
+
# get the doc_id
|
| 997 |
+
data = uploading_res["data"][0]
|
| 998 |
+
doc_id = data["id"]
|
| 999 |
+
# parse
|
| 1000 |
+
res = ragflow.stop_parsing_document("imagination", doc_id)
|
| 1001 |
+
assert res["code"] == RetCode.DATA_ERROR and res["message"] == "This dataset 'imagination' cannot be found!"
|
| 1002 |
+
|
| 1003 |
+
# ------------------------stop parsing multiple documents----------------------------
|
| 1004 |
+
def test_stop_parsing_documents_in_nonexistent_dataset(self):
|
| 1005 |
+
"""
|
| 1006 |
+
Test the stopping parsing documents whose dataset is nonexistent.
|
| 1007 |
+
"""
|
| 1008 |
+
# create a dataset
|
| 1009 |
+
ragflow = RAGFlow(API_KEY, HOST_ADDRESS)
|
| 1010 |
+
created_res = ragflow.create_dataset("test_download_nonexistent_document")
|
| 1011 |
+
created_res_id = created_res["data"]["dataset_id"]
|
| 1012 |
+
# upload files
|
| 1013 |
+
file_paths = ["test_data/test.txt"]
|
| 1014 |
+
uploading_res = ragflow.upload_local_file(created_res_id, file_paths)
|
| 1015 |
+
# parse
|
| 1016 |
+
res = ragflow.stop_parsing_documents("imagination")
|
| 1017 |
+
assert res["code"] == RetCode.DATA_ERROR and res["message"] == "This dataset 'imagination' cannot be found!"
|
| 1018 |
+
|
| 1019 |
+
def test_stop_parsing_multiple_documents(self):
|
| 1020 |
+
"""
|
| 1021 |
+
Test the stopping parsing documents with a success.
|
| 1022 |
+
"""
|
| 1023 |
+
# create a dataset
|
| 1024 |
+
ragflow = RAGFlow(API_KEY, HOST_ADDRESS)
|
| 1025 |
+
created_res = ragflow.create_dataset("test_start_parsing_multiple_documents")
|
| 1026 |
+
created_res_id = created_res["data"]["dataset_id"]
|
| 1027 |
+
# upload files
|
| 1028 |
+
file_paths = ["test_data/test.txt", "test_data/test1.txt"]
|
| 1029 |
+
ragflow.upload_local_file(created_res_id, file_paths)
|
| 1030 |
+
res = ragflow.start_parsing_documents(created_res_id)
|
| 1031 |
+
assert res["code"] == RetCode.SUCCESS and res["data"] is True and res["message"] == ""
|
| 1032 |
+
|
| 1033 |
+
res = ragflow.stop_parsing_documents(created_res_id)
|
| 1034 |
+
assert res["code"] == RetCode.SUCCESS and res["data"] is True and res["message"] == ""
|
| 1035 |
+
|
| 1036 |
+
def test_stop_parsing_multiple_documents_with_one_empty_file(self):
|
| 1037 |
+
"""
|
| 1038 |
+
Test the stopping parsing documents, one of which is empty.
|
| 1039 |
+
"""
|
| 1040 |
+
# create a dataset
|
| 1041 |
+
ragflow = RAGFlow(API_KEY, HOST_ADDRESS)
|
| 1042 |
+
created_res = ragflow.create_dataset(" test_start_parsing_multiple_documents")
|
| 1043 |
+
created_res_id = created_res["data"]["dataset_id"]
|
| 1044 |
+
# upload files
|
| 1045 |
+
file_paths = ["test_data/test.txt", "test_data/test1.txt", "test_data/empty.txt"]
|
| 1046 |
+
ragflow.upload_local_file(created_res_id, file_paths)
|
| 1047 |
+
res = ragflow.start_parsing_documents(created_res_id)
|
| 1048 |
+
assert res["code"] == RetCode.SUCCESS and res["message"] == "Empty data in the document: empty.txt; "
|
| 1049 |
+
res = ragflow.stop_parsing_documents(created_res_id)
|
| 1050 |
+
assert res["code"] == RetCode.SUCCESS and res["data"] is True and res["message"] == ""
|
| 1051 |
+
|
| 1052 |
+
def test_stop_parsing_multiple_specific_documents(self):
|
| 1053 |
+
"""
|
| 1054 |
+
Test the stopping parsing documents whose document ids are specified.
|
| 1055 |
+
"""
|
| 1056 |
+
# create a dataset
|
| 1057 |
+
ragflow = RAGFlow(API_KEY, HOST_ADDRESS)
|
| 1058 |
+
created_res = ragflow.create_dataset(" test_start_parsing_multiple_documents")
|
| 1059 |
+
created_res_id = created_res["data"]["dataset_id"]
|
| 1060 |
+
# upload files
|
| 1061 |
+
file_paths = ["test_data/test.txt", "test_data/test1.txt"]
|
| 1062 |
+
uploading_res = ragflow.upload_local_file(created_res_id, file_paths)
|
| 1063 |
+
# get the doc_id
|
| 1064 |
+
data = uploading_res["data"]
|
| 1065 |
+
doc_ids = []
|
| 1066 |
+
for d in data:
|
| 1067 |
+
doc_ids.append(d["id"])
|
| 1068 |
+
res = ragflow.start_parsing_documents(created_res_id, doc_ids)
|
| 1069 |
+
assert res["code"] == RetCode.SUCCESS and res["message"] == ""
|
| 1070 |
+
res = ragflow.stop_parsing_documents(created_res_id, doc_ids)
|
| 1071 |
+
assert res["code"] == RetCode.SUCCESS and res["data"] is True and res["message"] == ""
|
| 1072 |
|
| 1073 |
# ----------------------------show the status of the file-----------------------------------------------------
|
| 1074 |
def test_show_status_with_success(self):
|