File tree Expand file tree Collapse file tree 6 files changed +25
-0
lines changed
Expand file tree Collapse file tree 6 files changed +25
-0
lines changed Original file line number Diff line number Diff line change 3131 runs-on : ubuntu-latest
3232 env :
3333 INFERENCE_MODEL : meta-llama/Llama-3.2-1B-Instruct
34+ EMBEDDING_MODEL : granite-embedding-125m
3435 VLLM_URL : http://localhost:8000/v1
3536 strategy :
3637 matrix :
Original file line number Diff line number Diff line change @@ -41,6 +41,8 @@ RUN pip install \
4141 uvicorn
4242RUN pip install \
4343 llama_stack_provider_lmeval==0.2.4
44+ RUN pip install \
45+ llama_stack_provider_ragas[remote]==0.3.0
4446RUN pip install \
4547 --extra-index-url https://test.pypi.org/simple/ llama_stack_provider_trustyai_fms==0.2.3
4648RUN pip install --extra-index-url https://download.pytorch.org/whl/cpu torch 'torchao>=0.12.0' torchvision
Original file line number Diff line number Diff line change @@ -13,7 +13,9 @@ You can see an overview of the APIs and Providers the image ships with in the ta
1313| agents | inline::meta-reference | Yes | N/A |
1414| datasetio | inline::localfs | Yes | N/A |
1515| datasetio | remote::huggingface | Yes | N/A |
16+ | eval | inline::trustyai_ragas | No | Set the ` EMBEDDING_MODEL ` environment variable |
1617| eval | remote::trustyai_lmeval | Yes | N/A |
18+ | eval | remote::trustyai_ragas | No | Set the ` KUBEFLOW_LLAMA_STACK_URL ` environment variable |
1719| files | inline::localfs | Yes | N/A |
1820| inference | inline::sentence-transformers | Yes | N/A |
1921| inference | remote::azure | No | Set the ` AZURE_API_KEY ` environment variable |
Original file line number Diff line number Diff line change @@ -21,6 +21,8 @@ distribution_spec:
2121 eval :
2222 - provider_type : remote::trustyai_lmeval
2323 module : llama_stack_provider_lmeval==0.2.4
24+ - provider_type : inline::trustyai_ragas
25+ module : llama_stack_provider_ragas[remote]==0.3.0
2426 datasetio :
2527 - provider_type : remote::huggingface
2628 - provider_type : inline::localfs
Original file line number Diff line number Diff line change @@ -108,6 +108,23 @@ providers:
108108 config :
109109 use_k8s : ${env.TRUSTYAI_LMEVAL_USE_K8S:=true}
110110 base_url : ${env.VLLM_URL:=}
111+ - provider_id : ${env.EMBEDDING_MODEL:+trustyai_ragas_inline}
112+ provider_type : inline::trustyai_ragas
113+ module : llama_stack_provider_ragas.inline
114+ config :
115+ embedding_model : ${env.EMBEDDING_MODEL:=}
116+ - provider_id : ${env.KUBEFLOW_LLAMA_STACK_URL:+trustyai_ragas_remote}
117+ provider_type : remote::trustyai_ragas
118+ module : llama_stack_provider_ragas.remote
119+ config :
120+ embedding_model : ${env.EMBEDDING_MODEL:=}
121+ kubeflow_config :
122+ results_s3_prefix : ${env.KUBEFLOW_RESULTS_S3_PREFIX:=}
123+ s3_credentials_secret_name : ${env.KUBEFLOW_S3_CREDENTIALS_SECRET_NAME:=}
124+ pipelines_endpoint : ${env.KUBEFLOW_PIPELINES_ENDPOINT:=}
125+ namespace : ${env.KUBEFLOW_NAMESPACE:=}
126+ llama_stack_url : ${env.KUBEFLOW_LLAMA_STACK_URL:=}
127+ base_image : ${env.KUBEFLOW_BASE_IMAGE:=}
111128 datasetio :
112129 - provider_id : huggingface
113130 provider_type : remote::huggingface
Original file line number Diff line number Diff line change @@ -10,6 +10,7 @@ function start_and_wait_for_llama_stack_container {
1010 --net=host \
1111 -p 8321:8321 \
1212 --env INFERENCE_MODEL=" $INFERENCE_MODEL " \
13+ --env EMBEDDING_MODEL=" $EMBEDDING_MODEL " \
1314 --env VLLM_URL=" $VLLM_URL " \
1415 --env TRUSTYAI_LMEVAL_USE_K8S=False \
1516 --name llama-stack \
You can’t perform that action at this time.
0 commit comments