feat: add watsonx inference provider and bump lmeval#28
feat: add watsonx inference provider and bump lmeval#28leseb merged 2 commits intoopendatahub-io:mainfrom
Conversation
New provider, env variable WATSONX_API_KEY and WATSONX_PROJECT_ID are expected to be populated to enable the provider. Relates to: RHAIENG-1026 Signed-off-by: Sébastien Han <seb@redhat.com>
|
@coderabbitai ignore |
|
Note Reviews pausedUse the following commands to manage reviews:
Caution Failed to replace (edit) comment. This is likely due to insufficient permissions or the comment being deleted. Error detailsPre-merge checks (3 passed)✅ Passed checks (3 passed)
✨ Finishing touches
🧪 Generate unit tests
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
There was a problem hiding this comment.
Actionable comments posted: 1
🧹 Nitpick comments (2)
distribution/Containerfile (1)
18-18: Use canonical PyPI name and (ideally) pin the version for reproducible builds.pip normalizes names, but the published package is “ibm-watsonx-ai”. Consider also pinning to a known-good version to avoid surprise breakages.
Apply:
- ibm_watsonx_ai \ + ibm-watsonx-ai \Optionally:
- ibm-watsonx-ai \ + ibm-watsonx-ai==X.Y.Z \If you want me to propose a tested version matrix for UBI9 + Py3.12, say the word.
distribution/run.yaml (1)
39-44: Wire a model to the new provider so it’s usable out-of-the-box.No model currently targets
provider_id: watsonx. Add a model entry controlled by an env var.Add under
models:(outside this hunk):- metadata: {} model_id: ${env.WATSONX_MODEL_ID:=watsonx-llm} provider_id: watsonx provider_model_id: ${env.WATSONX_PROVIDER_MODEL_ID:=} model_type: llmI can align names with your existing conventions if you share the desired default model(s).
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (3)
distribution/Containerfile(1 hunks)distribution/build.yaml(1 hunks)distribution/run.yaml(1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (1)
- GitHub Check: build-test-push (linux/amd64)
🔇 Additional comments (1)
distribution/build.yaml (1)
9-9: Verifyremote::watsonxexists in the bundled llama-stack (0.2.21) or add a module mapping.File: distribution/build.yaml (line 9) — provider_type:
remote::watsonx.Automated check failed in the verification environment (ModuleNotFoundError: No module named 'llama_stack'). Run this locally to confirm the provider registry:
python - <<'PY' import importlib, pkgutil m = importlib.import_module("llama_stack.providers") print([name for _, name, _ in pkgutil.walk_packages(m.__path__, m.__name__ + ".") if name.endswith(".watsonx")]) PYIf
remote::watsonxis absent in the pinned 0.2.21, either bumpllama-stackto a version that includes it or add amodule: <watsonx provider package>==<version>entry alongside this provider_type in distribution/build.yaml.
✅ Actions performedReviews paused. |
Fix auth token propagation in the safety provider. Relates to: RHOAIENG-34050 Signed-off-by: Sébastien Han <seb@redhat.com>
|
draft since downstream hasn't merged yet |
|
Ready! |
8b82642 feat: add watsonx inference provider
ffce08d chore: bump lmeval
commit 8b82642
Author: Sébastien Han seb@redhat.com
Date: Fri Sep 12 17:10:05 2025 +0200
commit ffce08d
Author: Sébastien Han seb@redhat.com
Date: Fri Sep 12 17:22:21 2025 +0200
@coderabbitai ignore
Summary by CodeRabbit
New Features
Chores