Skip to content

Commit 2a6756d

Browse files
committed
Update docs
Signed-off-by: Jin Hai <haijin.chn@gmail.com>
1 parent 6567ecf commit 2a6756d

File tree

5 files changed

+3
-18
lines changed

5 files changed

+3
-18
lines changed

docs/faq.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -151,7 +151,7 @@ See [Build a RAGFlow Docker image](./develop/build_docker_image.mdx).
151151

152152
### Cannot access https://huggingface.co
153153

154-
A locally deployed RAGflow downloads OCR and embedding modules from [Huggingface website](https://huggingface.co) by default. If your machine is unable to access this site, the following error occurs and PDF parsing fails:
154+
A locally deployed RAGflow downloads OCR models from [Huggingface website](https://huggingface.co) by default. If your machine is unable to access this site, the following error occurs and PDF parsing fails:
155155

156156
```
157157
FileNotFoundError: [Errno 2] No such file or directory: '/root/.cache/huggingface/hub/models--InfiniFlow--deepdoc/snapshots/be0c1e50eef6047b412d1800aa89aba4d275f997/ocr.res'

docs/guides/agent/agent_component_reference/retrieval.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ By default, you can use `sys.query`, which is the user query and the default out
3838

3939
### 3. Select dataset(s) to query
4040

41-
You can specify one or multiple datasets to retrieve data from. If selecting mutiple, ensure they use the same embedding model.
41+
You can specify one or multiple datasets to retrieve data from. If selecting multiple, ensure they use the same embedding model.
4242

4343
### 4. Expand **Advanced Settings** to configure the retrieval method
4444

docs/guides/dataset/best_practices/accelerate_doc_indexing.mdx

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,6 @@ A checklist to speed up document parsing and indexing.
1212

1313
Please note that some of your settings may consume a significant amount of time. If you often find that document parsing is time-consuming, here is a checklist to consider:
1414

15-
- Use GPU to reduce embedding time.
1615
- On the configuration page of your dataset, switch off **Use RAPTOR to enhance retrieval**.
1716
- Extracting knowledge graph (GraphRAG) is time-consuming.
1817
- Disable **Auto-keyword** and **Auto-question** on the configuration page of your dataset, as both depend on the LLM.

docs/guides/models/deploy_local_llm.mdx

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -107,7 +107,6 @@ Max retries exceeded with url: /api/chat (Caused by NewConnectionError('<urllib3
107107
Click on your logo **>** **Model providers** **>** **System Model Settings** to update your model:
108108

109109
- *You should now be able to find **llama3.2** from the dropdown list under **Chat model**, and **bge-m3** from the dropdown list under **Embedding model**.*
110-
- _If your local model is an embedding model, you should find it under **Embedding model**._
111110

112111
### 6. Update Chat Configuration
113112

@@ -158,14 +157,10 @@ Click on your logo **>** **Model providers** **>** **System Model Settings** to
158157

159158
*You should now be able to find **mistral** from the dropdown list under **Chat model**.*
160159

161-
> If your local model is an embedding model, you should find your local model under **Embedding model**.
162-
163160
### 7. Update Chat Configuration
164161

165162
Update your chat model accordingly in **Chat Configuration**:
166163

167-
> If your local model is an embedding model, update it on the configuration page of your dataset.
168-
169164
## Deploy a local model using IPEX-LLM
170165

171166
[IPEX-LLM](https://github.com/intel-analytics/ipex-llm) is a PyTorch library for running LLMs on local Intel CPUs or GPUs (including iGPU or discrete GPUs like Arc, Flex, and Max) with low latency. It supports Ollama on Linux and Windows systems.

docs/quickstart.mdx

Lines changed: 1 addition & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -190,7 +190,7 @@ This section provides instructions on setting up the RAGFlow server on Linux. If
190190
3. Use the pre-built Docker images and start up the server:
191191

192192
```bash
193-
# Use CPU for embedding and DeepDoc tasks:
193+
# Use CPU for DeepDoc tasks:
194194
$ docker compose -f docker-compose.yml up -d
195195
```
196196

@@ -207,15 +207,6 @@ This section provides instructions on setting up the RAGFlow server on Linux. If
207207
</APITable>
208208
```
209209

210-
:::danger IMPORTANT
211-
The embedding models included in `v0.21.1` and `nightly` are:
212-
213-
- BAAI/bge-large-zh-v1.5
214-
- maidalun1020/bce-embedding-base_v1
215-
216-
These two embedding models are optimized specifically for English and Chinese, so performance will be compromised if you use them to embed documents in other languages.
217-
:::
218-
219210
:::tip NOTE
220211
The image size shown refers to the size of the *downloaded* Docker image, which is compressed. When Docker runs the image, it unpacks it, resulting in significantly greater disk usage. A Docker image will expand to around 7 GB once unpacked.
221212
:::

0 commit comments

Comments
 (0)