Skip to content

Commit ab2d503

Browse files
authored
Kheiss/qa review2 (#1583)
1 parent 15e5406 commit ab2d503

File tree

5 files changed

+18
-6
lines changed

5 files changed

+18
-6
lines changed

docs/docs/extraction/audio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ to transcribe speech to text, which is then embedded by using the Nemotron embed
2727

2828
!!! important
2929

30-
Due to limitations in available VRAM controls in the current release, the RIVA ASR NIM microservice must run on a [dedicated additional GPU](support-matrix.md). For the full list of requirements, refer to [Support Matrix](https://docs.nvidia.com/deeplearning/riva/user-guide/docs/support-matrix.html).
30+
Due to limitations in available VRAM controls in the current release, the RIVA ASR NIM microservice must run on a [dedicated additional GPU](support-matrix.md). For the full list of requirements, refer to [Support Matrix](https://docs.nvidia.com/deeplearning/riva/user-guide/docs/support-matrix/support-matrix.html).
3131

3232
This pipeline enables users to retrieve speech files at the segment level.
3333

docs/docs/extraction/prerequisites.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@ Before you begin using [NeMo Retriever Library](overview.md), ensure the followi
1111
## Software Requirements
1212

1313
- Linux operating systems (Ubuntu 22.04 or later recommended)
14+
- **Python 3.12 or later** (required for NeMo Retriever Library packages; see note below)
1415
- [Docker](https://docs.docker.com/engine/install/)
1516
- [Docker Compose](https://docs.docker.com/compose/install/)
1617
- [Docker Buildx](https://docs.docker.com/build/concepts/overview/#buildx) `>= 0.17` (Compose 2.40+ enforces this)
@@ -21,7 +22,7 @@ Before you begin using [NeMo Retriever Library](overview.md), ensure the followi
2122

2223
!!! note
2324

24-
You install Python later.
25+
Install **Python 3.12 or later** before creating your environment. Using Python 3.10 or 3.11 will cause dependency resolution failures when installing NeMo Retriever Library packages.
2526

2627

2728

docs/docs/extraction/quickstart-guide.md

Lines changed: 8 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -82,14 +82,20 @@ h. Run the command `docker ps`. You should see output similar to the following.
8282

8383
```
8484
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
85+
...
86+
```
87+
88+
To run the NeMo Retriever Library Python client from your host machine, **Python 3.12 or later is required**. Create a virtual environment and install the client packages:
89+
90+
```shell
8591
uv venv --python 3.12 nv-ingest-dev
8692
source nv-ingest-dev/bin/activate
8793
uv pip install nv-ingest==26.1.2 nv-ingest-api==26.1.2 nv-ingest-client==26.1.2
8894
```
8995

9096
!!! tip
9197

92-
To confirm that you have activated your Conda environment, run `which pip` and `which python`, and confirm that you see `nemo_retriever` in the result. You can do this before any pip or python command that you run.
98+
To confirm that you have activated your virtual environment, run `which pip` and `which python`, and confirm that you see `nemo_retriever` or your venv path in the result. You can do this before any pip or python command that you run.
9399

94100

95101
!!! note
@@ -459,7 +465,7 @@ docker compose \
459465

460466
## Specify MIG slices for NIM models
461467

462-
When you deploy NeMo Retriever Library with NIM models on MIG‑enabled GPUs, MIG device slices are requested and scheduled through the `values.yaml` file for the corresponding NIM microservice. For IBM Content-Aware Storage (CAS) deployments, this allows NeMo Retriever Library NIM pods to land only on nodes that expose the desired MIG profiles [raw.githubusercontent](https://raw.githubusercontent.com/NVIDIA/NeMo-Retriever/main/helm/README.md%E2%80%8B).​
468+
When you deploy NeMo Retriever Library with NIM models on MIG‑enabled GPUs, MIG device slices are requested and scheduled through the `values.yaml` file for the corresponding NIM microservice. For IBM Content-Aware Storage (CAS) deployments, this allows NeMo Retriever Library NIM pods to land only on nodes that expose the desired MIG profiles [raw.githubusercontent](https://raw.githubusercontent.com/NVIDIA/NeMo-Retriever/main/helm/README.md).​
463469

464470
To target a specific MIG profile—for example, a 3g.20gb slice on an A100, which is a hardware-partitioned virtual GPU instance that gives your workload a fixed mid-sized share of the A100’s compute plus 20 GB of dedicated GPU memory and behaves like a smaller independent GPU—for a given NIM, configure the `resources` and `nodeSelector` under that NIM’s values path in `values.yaml`.
465471

docs/docs/extraction/support-matrix.md

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,12 +7,17 @@ Before you begin using [NeMo Retriever Library](overview.md), ensure that you ha
77
NVIDIA Ingest (nv-ingest) has been renamed to the NeMo Retriever Library.
88

99

10+
## Software Requirements
11+
12+
- **Python**: 3.12 or later. The NeMo Retriever Library core and harness require Python 3.12+; the client supports Python 3.11+. Using Python 3.10 or earlier will cause dependency resolution failures. For details, see [Prerequisites](prerequisites.md).
13+
14+
1015
## Core and Advanced Pipeline Features
1116

1217
The NeMo Retriever Library core pipeline features run on a single A10G or better GPU.
1318
The core pipeline features include the following:
1419

15-
- llama3.2-nv-embedqa-1b-v2 — Embedding model for converting text chunks into vectors.
20+
- llama-nemotron-embed-1b-v2 — Embedding model for converting text chunks into vectors.
1621
- nemotron-page-elements-v3 — Detects and classifies images on a page as a table, chart or infographic.
1722
- nemotron-table-structure-v1 — Detects rows, columns, and cells within a table to preserve table structure and convert to Markdown format.
1823
- nemotron-graphic-elements-v1 — Detects graphic elements within chart images such as titles, legends, axes, and numerical values.

docs/docs/extraction/vlm-embed.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Use Multimodal Embedding with NeMo Retriever Library
22

3-
This guide explains how to use the [NeMo Retriever Library](https://www.perplexity.ai/search/overview.md) with the multimodal embedding model [Llama Nemotron Embed VL 1B v2](https://build.nvidia.com/nvidia/llama-nemotron-embed-vl-1b-v2).
3+
This guide explains how to use the [NeMo Retriever Library](overview.md) with the multimodal embedding model [Llama Nemotron Embed VL 1B v2](https://build.nvidia.com/nvidia/llama-nemotron-embed-vl-1b-v2).
44

55
The `Llama Nemotron Embed VL 1B v2` model is optimized for multimodal question-answering and retrieval tasks.
66
It can embed documents as text, images, or paired text-image combinations.

0 commit comments

Comments
 (0)