You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[rerank]: Refine documentation for rerank comps (opea-project#758)
* [rerank]: Refine documentation for rerank comps
Restructure and refine README documentation for rerank components.
Signed-off-by: Yeoh, Hoong Tee <[email protected]>
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
* rerank-doc: Break up README lines for viewing experience
Signed-off-by: Yeoh, Hoong Tee <[email protected]>
---------
Signed-off-by: Yeoh, Hoong Tee <[email protected]>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: chen, suyue <[email protected]>
Copy file name to clipboardExpand all lines: comps/reranks/fastrag/README.md
+29-16Lines changed: 29 additions & 16 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,13 @@
1
-
# Reranking Microservice
1
+
# Reranking Microservice with fastRAG
2
2
3
-
The Reranking Microservice, fueled by reranking models, stands as a straightforward yet immensely potent tool for semantic search. When provided with a query and a collection of documents, reranking swiftly indexes the documents based on their semantic relevance to the query, arranging them from most to least pertinent. This microservice significantly enhances overall accuracy. In a text retrieval system, either a dense embedding model or a sparse lexical search index is often employed to retrieve relevant text documents based on the input. However, a reranking model can further refine this process by rearranging potential candidates into a final, optimized order.
3
+
`fastRAG` is a research framework for efficient and optimized retrieval augmented generative pipelines, incorporating state-of-the-art LLMs and Information Retrieval.
4
+
5
+
Please refer to [Official fastRAG repo](https://github.com/IntelLabs/fastRAG/tree/main)
6
+
for more information.
7
+
8
+
This README provides set-up instructions and comprehensive details regarding the reranking microservice via fastRAG.
Copy file name to clipboardExpand all lines: comps/reranks/tei/README.md
+37-24Lines changed: 37 additions & 24 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,11 @@
1
-
# Reranking Microservice
1
+
# Reranking Microservice via TEI
2
2
3
-
The Reranking Microservice, fueled by reranking models, stands as a straightforward yet immensely potent tool for semantic search. When provided with a query and a collection of documents, reranking swiftly indexes the documents based on their semantic relevance to the query, arranging them from most to least pertinent. This microservice significantly enhances overall accuracy. In a text retrieval system, either a dense embedding model or a sparse lexical search index is often employed to retrieve relevant text documents based on the input. However, a reranking model can further refine this process by rearranging potential candidates into a final, optimized order.
3
+
`Text Embeddings Inference (TEI)` is a comprehensive toolkit designed for efficient deployment and serving of open source text embeddings models.
4
+
It enable us to host our own reranker endpoint seamlessly.
5
+
6
+
This README provides set-up instructions and comprehensive details regarding the reranking microservice via TEI.
Copy file name to clipboardExpand all lines: comps/reranks/videoqna/README.md
+23-11Lines changed: 23 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,8 +1,14 @@
1
-
# Rerank Microservice
1
+
# Rerank Microservice with VideoQnA
2
2
3
-
This is a Docker-based microservice that do result rerank for VideoQnA use case. Local rerank is used rather than rerank model.
3
+
This README provides set-up instructions and comprehensive details regarding the reranking microservice with VideoQnA.
4
+
This microservice is designed that do result rerank for VideoQnA use case. Local rerank is used rather than rerank model.
4
5
5
-
For the `VideoQnA` usecase, during the data preparation phase, frames are extracted from videos and stored in a vector database. To identify the most relevant video, we count the occurrences of each video source among the retrieved data with rerank function `get_top_doc`. This sorts the video as a descending list of names, ranked by their degree of match with the query. Then we could send the `top_n` videos to the downstream LVM.
6
+
For the `VideoQnA` usecase, during the data preparation phase, frames are extracted from videos and stored in a vector database.
7
+
To identify the most relevant video, we count the occurrences of each video source among the retrieved data with rerank function `get_top_doc`.
8
+
This sorts the video as a descending list of names, ranked by their degree of match with the query.
9
+
Then we could send the `top_n` videos to the downstream LVM.
{"id":"random number","video_url":"http://0.0.0.0:6005/top_video_name","chunk_start":20.0,"chunk_duration":10.0,"prompt":"this is the query","max_new_tokens":512}
62
+
# Expected output result:
63
+
#{"id":"random number","video_url":"http://0.0.0.0:6005/top_video_name","chunk_start":20.0,"chunk_duration":10.0,"prompt":"this is the query","max_new_tokens":512}
0 commit comments