fix: Fiw qwen model name in llama-stack core tests#587
fix: Fiw qwen model name in llama-stack core tests#587jgarciao merged 1 commit intoopendatahub-io:mainfrom
Conversation
Signed-off-by: Jorge Garcia Oncins <jgarciao@redhat.com>
📝 WalkthroughWalkthroughUpdates include renaming the Qwen model constant, using it in tests, adding a served model name flag to the vLLM CPU runtime container args, and adjusting a core test to expect a namespaced model identifier and flexible model count. Changes
Estimated code review effort🎯 2 (Simple) | ⏱️ ~10 minutes ✨ Finishing Touches
🧪 Generate unit tests
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. 🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
SupportNeed help? Create a ticket on our support page for assistance with any issues or questions. CodeRabbit Commands (Invoked using PR/Issue comments)Type Other keywords and placeholders
Status, Documentation and Community
|
|
The following are automatically added/executed:
Available user actions:
Supported labels{'/hold', '/cherry-pick', '/lgtm', '/wip', '/build-push-pr-image', '/verified'} |
There was a problem hiding this comment.
Actionable comments posted: 3
🧹 Nitpick comments (1)
tests/fixtures/inference.py (1)
35-35: Good: pass served model name to vLLM.
The adapter will expose the namespaced id as expected once registered. Consider lifting 8032 into a shared constant to avoid drift.- "args": ["--port=8032", "--model=/mnt/models", "--served-model-name={{.Name}}"], + "args": ["--port=8032", "--model=/mnt/models", "--served-model-name={{.Name}}"], # TODO: use a shared PORT constant
📜 Review details
Configuration used: Path: .coderabbit.yaml
Review profile: CHILL
Plan: Pro
💡 Knowledge Base configuration:
- MCP integration is disabled by default for public repositories
- Jira integration is disabled by default for public repositories
- Linear integration is disabled by default for public repositories
You can enable these sources in your CodeRabbit configuration.
📒 Files selected for processing (3)
tests/fixtures/inference.py(3 hunks)tests/llama_stack/core/test_llamastack_core.py(1 hunks)utilities/constants.py(1 hunks)
🧰 Additional context used
🧬 Code graph analysis (2)
tests/llama_stack/core/test_llamastack_core.py (1)
tests/llama_stack/constants.py (2)
LlamaStackProviders(4-14)Inference(7-8)
tests/fixtures/inference.py (1)
utilities/constants.py (2)
RuntimeTemplates(65-78)KServeDeploymentType(6-9)
🔇 Additional comments (1)
tests/fixtures/inference.py (1)
12-12: Importing QWEN_MODEL_NAME into fixtures is the right move.
Keeps the ISVC name and test expectations in sync.
|
Status of building tag latest: success. |
Signed-off-by: Jorge Garcia Oncins <jgarciao@redhat.com>
Summary by CodeRabbit