writinwaters
commited on
Commit
·
dc34855
1
Parent(s):
5281167
Fixed a broken link (#2190)
Browse filesTo fix a broken link
### Type of change
- [x] Documentation Update
- docs/references/faq.md +2 -2
docs/references/faq.md
CHANGED
@@ -357,7 +357,7 @@ This exception occurs when starting up the RAGFlow server. Try the following:
|
|
357 |
|
358 |
1. Right click the desired dialog to display the **Chat Configuration** window.
|
359 |
2. Switch to the **Model Setting** tab and adjust the **Max Tokens** slider to get the desired length.
|
360 |
-
3. Click **OK** to confirm your change.
|
361 |
|
362 |
|
363 |
### 2. What does Empty response mean? How to set it?
|
@@ -370,7 +370,7 @@ You limit what the system responds to what you specify in **Empty response** if
|
|
370 |
|
371 |
### 4. How to run RAGFlow with a locally deployed LLM?
|
372 |
|
373 |
-
You can use Ollama to deploy local LLM. See [here](
|
374 |
|
375 |
### 5. How to link up ragflow and ollama servers?
|
376 |
|
|
|
357 |
|
358 |
1. Right click the desired dialog to display the **Chat Configuration** window.
|
359 |
2. Switch to the **Model Setting** tab and adjust the **Max Tokens** slider to get the desired length.
|
360 |
+
3. Click **OK** to confirm your change.
|
361 |
|
362 |
|
363 |
### 2. What does Empty response mean? How to set it?
|
|
|
370 |
|
371 |
### 4. How to run RAGFlow with a locally deployed LLM?
|
372 |
|
373 |
+
You can use Ollama to deploy local LLM. See [here](../guides/deploy_local_llm.mdx) for more information.
|
374 |
|
375 |
### 5. How to link up ragflow and ollama servers?
|
376 |
|