writinwaters commited on
Commit
ccccbdd
·
1 Parent(s): 5d3a620

Added supported LLMs (#1517)

Browse files

### What problem does this PR solve?

_Briefly describe what this PR aims to solve. Include background context
that will help reviewers understand the purpose of the PR._

### Type of change

- [x] Documentation Update

docs/guides/configure_knowledge_base.md CHANGED
@@ -107,6 +107,10 @@ RAGFlow features visibility and explainability, allowing you to view the chunkin
107
 
108
  ![update chunk](https://github.com/infiniflow/ragflow/assets/93570324/1d84b408-4e9f-46fd-9413-8c1059bf9c76)
109
 
 
 
 
 
110
  4. In Retrieval testing, ask a quick question in **Test text** to double check if your configurations work:
111
 
112
  _As you can tell from the following, RAGFlow responds with truthful citations._
 
107
 
108
  ![update chunk](https://github.com/infiniflow/ragflow/assets/93570324/1d84b408-4e9f-46fd-9413-8c1059bf9c76)
109
 
110
+ :::caution NOTE
111
+ You can add keywords to a file chunk to increase its relevance. This action increases its keyword weight and can improve its position in search list.
112
+ :::
113
+
114
  4. In Retrieval testing, ask a quick question in **Test text** to double check if your configurations work:
115
 
116
  _As you can tell from the following, RAGFlow responds with truthful citations._
docs/guides/llm_api_key_setup.md CHANGED
@@ -12,8 +12,14 @@ An API key is required for RAGFlow to interact with an online AI model. This gui
12
  For now, RAGFlow supports the following online LLMs. Click the corresponding link to apply for your API key. Most LLM providers grant newly-created accounts trial credit, which will expire in a couple of months, or a promotional amount of free quota.
13
 
14
  - [OpenAI](https://platform.openai.com/login?launch),
 
 
 
 
 
15
  - [Tongyi-Qianwen](https://dashscope.console.aliyun.com/model),
16
  - [ZHIPU-AI](https://open.bigmodel.cn/),
 
17
  - [Moonshot](https://platform.moonshot.cn/docs),
18
  - [DeepSeek](https://platform.deepseek.com/api-docs/),
19
  - [Baichuan](https://www.baichuan-ai.com/home),
 
12
  For now, RAGFlow supports the following online LLMs. Click the corresponding link to apply for your API key. Most LLM providers grant newly-created accounts trial credit, which will expire in a couple of months, or a promotional amount of free quota.
13
 
14
  - [OpenAI](https://platform.openai.com/login?launch),
15
+ - Azure-OpenAI,
16
+ - Gemini,
17
+ - Groq,
18
+ - Mistral,
19
+ - Bedrock,
20
  - [Tongyi-Qianwen](https://dashscope.console.aliyun.com/model),
21
  - [ZHIPU-AI](https://open.bigmodel.cn/),
22
+ - MiniMax
23
  - [Moonshot](https://platform.moonshot.cn/docs),
24
  - [DeepSeek](https://platform.deepseek.com/api-docs/),
25
  - [Baichuan](https://www.baichuan-ai.com/home),
docs/quickstart.mdx CHANGED
@@ -177,14 +177,22 @@ With the default settings, you only need to enter `http://IP_OF_YOUR_MACHINE` (*
177
  RAGFlow is a RAG engine, and it needs to work with an LLM to offer grounded, hallucination-free question-answering capabilities. For now, RAGFlow supports the following LLMs, and the list is expanding:
178
 
179
  - OpenAI
 
 
 
 
 
180
  - Tongyi-Qianwen
181
  - ZHIPU-AI
 
182
  - Moonshot
183
  - DeepSeek-V2
184
  - Baichuan
185
  - VolcEngine
186
 
187
- > RAGFlow also supports deploying LLMs locally using Ollama or Xinference, but this part is not covered in this quick start guide.
 
 
188
 
189
  To add and configure an LLM:
190
 
@@ -192,7 +200,7 @@ To add and configure an LLM:
192
 
193
  ![add llm](https://github.com/infiniflow/ragflow/assets/93570324/10635088-028b-4b3d-add9-5c5a6e626814)
194
 
195
- > Each RAGFlow account is able to use **text-embedding-v2** for free, a embedding model of Tongyi-Qianwen. This is why you can see Tongyi-Qianwen in the **Added models** list. And you may need to update your Tongyi-Qianwen API key at a later point.
196
 
197
  2. Click on the desired LLM and update the API key accordingly (DeepSeek-V2 in this case):
198
 
@@ -228,7 +236,9 @@ To create your first knowledge base:
228
 
229
  3. RAGFlow offers multiple chunk templates that cater to different document layouts and file formats. Select the embedding model and chunk method (template) for your knowledge base.
230
 
231
- > IMPORTANT: Once you have selected an embedding model and used it to parse a file, you are no longer allowed to change it. The obvious reason is that we must ensure that all files in a specific knowledge base are parsed using the *same* embedding model (ensure that they are being compared in the same embedding space).
 
 
232
 
233
  _You are taken to the **Dataset** page of your knowledge base._
234
 
@@ -240,6 +250,11 @@ To create your first knowledge base:
240
 
241
  _When the file parsing completes, its parsing status changes to **SUCCESS**._
242
 
 
 
 
 
 
243
  ## Intervene with file parsing
244
 
245
  RAGFlow features visibility and explainability, allowing you to view the chunking results and intervene where necessary. To do so:
@@ -256,6 +271,10 @@ RAGFlow features visibility and explainability, allowing you to view the chunkin
256
 
257
  ![update chunk](https://github.com/infiniflow/ragflow/assets/93570324/1d84b408-4e9f-46fd-9413-8c1059bf9c76)
258
 
 
 
 
 
259
  4. In Retrieval testing, ask a quick question in **Test text** to double check if your configurations work:
260
 
261
  _As you can tell from the following, RAGFlow responds with truthful citations._
 
177
  RAGFlow is a RAG engine, and it needs to work with an LLM to offer grounded, hallucination-free question-answering capabilities. For now, RAGFlow supports the following LLMs, and the list is expanding:
178
 
179
  - OpenAI
180
+ - Azure-OpenAI
181
+ - Gemini
182
+ - Groq
183
+ - Mistral
184
+ - Bedrock
185
  - Tongyi-Qianwen
186
  - ZHIPU-AI
187
+ - MiniMax
188
  - Moonshot
189
  - DeepSeek-V2
190
  - Baichuan
191
  - VolcEngine
192
 
193
+ :::note
194
+ RAGFlow also supports deploying LLMs locally using Ollama or Xinference, but this part is not covered in this quick start guide.
195
+ :::
196
 
197
  To add and configure an LLM:
198
 
 
200
 
201
  ![add llm](https://github.com/infiniflow/ragflow/assets/93570324/10635088-028b-4b3d-add9-5c5a6e626814)
202
 
203
+ > Each RAGFlow account is able to use **text-embedding-v2** for free, an embedding model of Tongyi-Qianwen. This is why you can see Tongyi-Qianwen in the **Added models** list. And you may need to update your Tongyi-Qianwen API key at a later point.
204
 
205
  2. Click on the desired LLM and update the API key accordingly (DeepSeek-V2 in this case):
206
 
 
236
 
237
  3. RAGFlow offers multiple chunk templates that cater to different document layouts and file formats. Select the embedding model and chunk method (template) for your knowledge base.
238
 
239
+ :::danger IMPORTANT
240
+ Once you have selected an embedding model and used it to parse a file, you are no longer allowed to change it. The obvious reason is that we must ensure that all files in a specific knowledge base are parsed using the *same* embedding model (ensure that they are being compared in the same embedding space).
241
+ :::
242
 
243
  _You are taken to the **Dataset** page of your knowledge base._
244
 
 
250
 
251
  _When the file parsing completes, its parsing status changes to **SUCCESS**._
252
 
253
+ :::alert NOTE
254
+ - If your file parsing gets stuck at below 1%, see [FAQ 4.3](https://ragflow.io/docs/dev/faq#43-why-does-my-document-parsing-stall-at-under-one-percent).
255
+ - If your file parsing gets stuck at near completion, see [FAQ 4.4](https://ragflow.io/docs/dev/faq#44-why-does-my-pdf-parsing-stall-near-completion-while-the-log-does-not-show-any-error)
256
+ :::
257
+
258
  ## Intervene with file parsing
259
 
260
  RAGFlow features visibility and explainability, allowing you to view the chunking results and intervene where necessary. To do so:
 
271
 
272
  ![update chunk](https://github.com/infiniflow/ragflow/assets/93570324/1d84b408-4e9f-46fd-9413-8c1059bf9c76)
273
 
274
+ :::caution NOTE
275
+ You can add keywords to a file chunk to increase its relevance. This action increases its keyword weight and can improve its position in search list.
276
+ :::
277
+
278
  4. In Retrieval testing, ask a quick question in **Test text** to double check if your configurations work:
279
 
280
  _As you can tell from the following, RAGFlow responds with truthful citations._
docs/references/faq.md CHANGED
@@ -198,7 +198,7 @@ Ignore this warning and continue. All system warnings can be ignored.
198
 
199
  ![stall](https://github.com/infiniflow/ragflow/assets/93570324/3589cc25-c733-47d5-bbfc-fedb74a3da50)
200
 
201
- If your RAGFlow is deployed *locally*, try the following:
202
 
203
  1. Check the log of your RAGFlow server to see if it is running properly:
204
  ```bash
@@ -209,15 +209,17 @@ docker logs -f ragflow-server
209
 
210
  #### 4.4 Why does my pdf parsing stall near completion, while the log does not show any error?
211
 
212
- If your RAGFlow is deployed *locally*, the parsing process is likely killed due to insufficient RAM. Try increasing your memory allocation by increasing the `MEM_LIMIT` value in **docker/.env**.
213
 
214
- > Ensure that you restart up your RAGFlow server for your changes to take effect!
215
- > ```bash
216
- > docker compose stop
217
- > ```
218
- > ```bash
219
- > docker compose up -d
220
- > ```
 
 
221
 
222
  ![nearcompletion](https://github.com/infiniflow/ragflow/assets/93570324/563974c3-f8bb-4ec8-b241-adcda8929cbb)
223
 
 
198
 
199
  ![stall](https://github.com/infiniflow/ragflow/assets/93570324/3589cc25-c733-47d5-bbfc-fedb74a3da50)
200
 
201
+ Click the red cross beside the 'parsing status' bar, then restart the parsing process to see if the issue remains. If the issue persists and your RAGFlow is deployed locally, try the following:
202
 
203
  1. Check the log of your RAGFlow server to see if it is running properly:
204
  ```bash
 
209
 
210
  #### 4.4 Why does my pdf parsing stall near completion, while the log does not show any error?
211
 
212
+ Click the red cross beside the 'parsing status' bar, then restart the parsing process to see if the issue remains. If the issue persists and your RAGFlow is deployed locally, the parsing process is likely killed due to insufficient RAM. Try increasing your memory allocation by increasing the `MEM_LIMIT` value in **docker/.env**.
213
 
214
+ :::note
215
+ Ensure that you restart up your RAGFlow server for your changes to take effect!
216
+ ```bash
217
+ docker compose stop
218
+ ```
219
+ ```bash
220
+ docker compose up -d
221
+ ```
222
+ :::
223
 
224
  ![nearcompletion](https://github.com/infiniflow/ragflow/assets/93570324/563974c3-f8bb-4ec8-b241-adcda8929cbb)
225