Skip to content

Commit bdac7ba

Browse files
committed
update readme
1 parent bf1bcce commit bdac7ba

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ You can deploy Flare AI RAG using Docker or set up the backend and frontend manu
3333
1. **Prepare the Environment File:**
3434
Rename `.env.example` to `.env` and update the variables accordingly. (e.g. your [Gemini API key](https://aistudio.google.com/app/apikey))
3535

36-
### Build using Docker (Recommended) -- [WIP]
36+
### Build using Docker (Recommended)
3737

3838
1. **Build the Docker Image:**
3939

@@ -257,10 +257,10 @@ Design and implement a knowledge ingestion pipeline, with a demonstration interf
257257

258258
_N.B._ Other vector databases can be used, provided they run within the same Docker container as the RAG system, since the deployment will occur in a TEE.
259259

260-
- **Enhanced Data Ingestion & Indexing**: Explore more sophisticated data structures for improved indexing and retrieval, and expand beyond a CSV format to include additional data sources (_e.g._, Flares GitHub, blogs, documentation). BigQuery integration would be desirable.
260+
- **Enhanced Data Ingestion & Indexing**: Explore more sophisticated data structures for improved indexing and retrieval, and expand beyond a CSV format to include additional data sources (_e.g._, Flare's GitHub, blogs, documentation). BigQuery integration would be desirable.
261261
- **Intelligent Query & Data Processing**: Use recommended AI models to refine the data processing pipeline, including pre-processing steps that optimize and clean incoming data, ensuring higher-quality context retrieval. (_e.g._ Use an LLM to reformulate or expand user queries before passing them to the retriever, improving the precision and recall of the semantic search.)
262262
- **Advanced Context Management**: Develop an intelligent context management system that:
263263
- Implements Dynamic Relevance Scoring to rank documents by their contextual importance.
264264
- Optimizes the Context Window to balance the amount of information sent to LLMs.
265265
- Includes Source Verification Mechanisms to assess and validate the reliability of the data sources.
266-
- **Improved Retrieval & Response Pipelines**: Integrate hybrid search techniques (combining semantic and keyword-based methods) for better retrieval, and implement completion checks to verify that the responders output is complete and accurate (potentially allow an iterative feedback loop for refining the final answer).
266+
- **Improved Retrieval & Response Pipelines**: Integrate hybrid search techniques (combining semantic and keyword-based methods) for better retrieval, and implement completion checks to verify that the responder's output is complete and accurate (potentially allow an iterative feedback loop for refining the final answer).

0 commit comments

Comments
 (0)