|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import os |
|
import random |
|
|
|
from api.db.db_utils import bulk_insert_into_db |
|
from deepdoc.parser import PdfParser |
|
from peewee import JOIN |
|
from api.db.db_models import DB, File2Document, File |
|
from api.db import StatusEnum, FileType, TaskStatus |
|
from api.db.db_models import Task, Document, Knowledgebase, Tenant |
|
from api.db.services.common_service import CommonService |
|
from api.db.services.document_service import DocumentService |
|
from api.utils import current_timestamp, get_uuid |
|
from deepdoc.parser.excel_parser import RAGFlowExcelParser |
|
from rag.settings import SVR_QUEUE_NAME |
|
from rag.utils.storage_factory import STORAGE_IMPL |
|
from rag.utils.redis_conn import REDIS_CONN |
|
|
|
|
|
class TaskService(CommonService): |
|
model = Task |
|
|
|
@classmethod |
|
@DB.connection_context() |
|
def get_task(cls, task_id): |
|
fields = [ |
|
cls.model.id, |
|
cls.model.doc_id, |
|
cls.model.from_page, |
|
cls.model.to_page, |
|
cls.model.retry_count, |
|
Document.kb_id, |
|
Document.parser_id, |
|
Document.parser_config, |
|
Document.name, |
|
Document.type, |
|
Document.location, |
|
Document.size, |
|
Knowledgebase.tenant_id, |
|
Knowledgebase.language, |
|
Knowledgebase.embd_id, |
|
Knowledgebase.pagerank, |
|
Tenant.img2txt_id, |
|
Tenant.asr_id, |
|
Tenant.llm_id, |
|
cls.model.update_time] |
|
docs = cls.model.select(*fields) \ |
|
.join(Document, on=(cls.model.doc_id == Document.id)) \ |
|
.join(Knowledgebase, on=(Document.kb_id == Knowledgebase.id)) \ |
|
.join(Tenant, on=(Knowledgebase.tenant_id == Tenant.id)) \ |
|
.where(cls.model.id == task_id) |
|
docs = list(docs.dicts()) |
|
if not docs: return None |
|
|
|
msg = "\nTask has been received." |
|
prog = random.random() / 10. |
|
if docs[0]["retry_count"] >= 3: |
|
msg = "\nERROR: Task is abandoned after 3 times attempts." |
|
prog = -1 |
|
|
|
cls.model.update(progress_msg=cls.model.progress_msg + msg, |
|
progress=prog, |
|
retry_count=docs[0]["retry_count"]+1 |
|
).where( |
|
cls.model.id == docs[0]["id"]).execute() |
|
|
|
if docs[0]["retry_count"] >= 3: return None |
|
|
|
return docs[0] |
|
|
|
@classmethod |
|
@DB.connection_context() |
|
def get_ongoing_doc_name(cls): |
|
with DB.lock("get_task", -1): |
|
docs = cls.model.select(*[Document.id, Document.kb_id, Document.location, File.parent_id]) \ |
|
.join(Document, on=(cls.model.doc_id == Document.id)) \ |
|
.join(File2Document, on=(File2Document.document_id == Document.id), join_type=JOIN.LEFT_OUTER) \ |
|
.join(File, on=(File2Document.file_id == File.id), join_type=JOIN.LEFT_OUTER) \ |
|
.where( |
|
Document.status == StatusEnum.VALID.value, |
|
Document.run == TaskStatus.RUNNING.value, |
|
~(Document.type == FileType.VIRTUAL.value), |
|
cls.model.progress < 1, |
|
cls.model.create_time >= current_timestamp() - 1000 * 600 |
|
) |
|
docs = list(docs.dicts()) |
|
if not docs: return [] |
|
|
|
return list(set([(d["parent_id"] if d["parent_id"] else d["kb_id"], d["location"]) for d in docs])) |
|
|
|
@classmethod |
|
@DB.connection_context() |
|
def do_cancel(cls, id): |
|
try: |
|
task = cls.model.get_by_id(id) |
|
_, doc = DocumentService.get_by_id(task.doc_id) |
|
return doc.run == TaskStatus.CANCEL.value or doc.progress < 0 |
|
except Exception: |
|
pass |
|
return False |
|
|
|
@classmethod |
|
@DB.connection_context() |
|
def update_progress(cls, id, info): |
|
if os.environ.get("MACOS"): |
|
if info["progress_msg"]: |
|
cls.model.update(progress_msg=cls.model.progress_msg + "\n" + info["progress_msg"]).where( |
|
cls.model.id == id).execute() |
|
if "progress" in info: |
|
cls.model.update(progress=info["progress"]).where( |
|
cls.model.id == id).execute() |
|
return |
|
|
|
with DB.lock("update_progress", -1): |
|
if info["progress_msg"]: |
|
cls.model.update(progress_msg=cls.model.progress_msg + "\n" + info["progress_msg"]).where( |
|
cls.model.id == id).execute() |
|
if "progress" in info: |
|
cls.model.update(progress=info["progress"]).where( |
|
cls.model.id == id).execute() |
|
|
|
|
|
def queue_tasks(doc: dict, bucket: str, name: str): |
|
def new_task(): |
|
return { |
|
"id": get_uuid(), |
|
"doc_id": doc["id"] |
|
} |
|
tsks = [] |
|
|
|
if doc["type"] == FileType.PDF.value: |
|
file_bin = STORAGE_IMPL.get(bucket, name) |
|
do_layout = doc["parser_config"].get("layout_recognize", True) |
|
pages = PdfParser.total_page_number(doc["name"], file_bin) |
|
page_size = doc["parser_config"].get("task_page_size", 12) |
|
if doc["parser_id"] == "paper": |
|
page_size = doc["parser_config"].get("task_page_size", 22) |
|
if doc["parser_id"] in ["one", "knowledge_graph"] or not do_layout: |
|
page_size = 10 ** 9 |
|
page_ranges = doc["parser_config"].get("pages") or [(1, 10 ** 5)] |
|
for s, e in page_ranges: |
|
s -= 1 |
|
s = max(0, s) |
|
e = min(e - 1, pages) |
|
for p in range(s, e, page_size): |
|
task = new_task() |
|
task["from_page"] = p |
|
task["to_page"] = min(p + page_size, e) |
|
tsks.append(task) |
|
|
|
elif doc["parser_id"] == "table": |
|
file_bin = STORAGE_IMPL.get(bucket, name) |
|
rn = RAGFlowExcelParser.row_number(doc["name"], file_bin) |
|
for i in range(0, rn, 3000): |
|
task = new_task() |
|
task["from_page"] = i |
|
task["to_page"] = min(i + 3000, rn) |
|
tsks.append(task) |
|
else: |
|
tsks.append(new_task()) |
|
|
|
bulk_insert_into_db(Task, tsks, True) |
|
DocumentService.begin2parse(doc["id"]) |
|
|
|
for t in tsks: |
|
assert REDIS_CONN.queue_product(SVR_QUEUE_NAME, message=t), "Can't access Redis. Please check the Redis' status." |
|
|