ragflow / docs /guides /accelerate_question_answering.mdx
writinwaters
Added document: Accelerate document indexing and retrieval (#4600)
0d43cf3
raw
history blame
2.77 kB
---
sidebar_position: 2
slug: /accelerate_doc_indexing_and_question_answering
---
# Accelerate document indexing and question answering
import APITable from '@site/src/components/APITable';
A checklist to speed up document parsing and question answering.
---
Please note that some of your settings may consume a significant amount of time. If you often find that document parsing and question answering are time-consuming, here is a checklist to consider:
## 1. Accelerate document indexing
- Use GPU to reduce embedding time.
- On the configuration page of your knowledge base, toggle off **Use RAPTOR to enhance retrieval**.
- The **Knowledge Graph** chunk method (GraphRAG) is time-consuming.
- Disable **Auto-keyword** and **Auto-question** on the configuration page of yor knowledge base, as both depend on the LLM.
## 2. Accelerate question answering
- In the **Prompt Engine** tab of your **Chat Configuration** dialogue, disabling **Multi-turn optimization** will reduce the time required to get an answer from the LLM.
- In the **Prompt Engine** tab of your **Chat Configuration** dialogue, leaving the **Rerank model** field empty will significantly decrease retrieval time.
- In the **Assistant Setting** tab of your **Chat Configuration** dialogue, disabling **Keyword analysis** will reduce the time to get get an answer from the LLM.
- When chatting with your chat assistant, click the light bulb icon above the *current* dialogue and scroll down the popup window to view the time taken for each task:
![enlighten](https://github.com/user-attachments/assets/fedfa2ee-21a7-451b-be66-20125619923c)
```mdx-code-block
<APITable>
```
| Item name | Description |
| ----------------- | ------------------------------------------------------------ |
| Total | Total time spent on this conversation round, including chunk retrieval and answer generation. |
| Check LLM | Time to validate the specified LLM. |
| Create retriever | Time to create a chunk retriever. |
| Bind embedding | Time to initialize an embedding model instance. |
| Bind LLM | Time to intialize an LLM instance. |
| Tune question | Time to optimize the user query using the context of the mult-turn conversation. |
| Bind reranker | Time to initialize an reranker model instance for chunk retrieval. |
| Generate keywords | Time to extract keywords from the user query. |
| Retrieval | Time to retrieve the chunks. |
| Generate answer | Time to generate the answer. |
```mdx-code-block
</APITable>
```