The inference container for the clip module
Documentation for this module can be found here.
LOCAL_REPO="multi2vec-clip" \
TEXT_MODEL_NAME="sentence-transformers/clip-ViT-B-32-multilingual-v1" \
CLIP_MODEL_NAME="clip-ViT-B-32" \
./cicd/build.sh
LOCAL_REPO="multi2vec-clip" ./cicd/test.sh
In order to run CLIP models on NVIDIA Jetson device one needs to have NVIDIA JetPack installed and configured.
This module only supports Jetson devices with JetPack 6 configured (JetPack 7 support coming soon).
In order to run a CLIP embedding model using Jetson GPU's one needs to install requirements-nvidia-jetpack6.txt:
uv venv --python 3.10
source .venv/bin/activate
uv pip install -r requirements-nvidia-jetpack6.txtUse ./cicd/download_model.sh script to download a model:
./cicd/download_model.shStart CLIP inference server:
ENABLE_CUDA=1 uvicorn app:app --host 0.0.0.0 --port 8000Run smoke tests:
uv pip install -r requirements-test.txt
python3 smoke_test.py