Skip to content

fix: Fiw qwen model name in llama-stack core tests#587

Merged
jgarciao merged 1 commit intoopendatahub-io:mainfrom
jgarciao:fix-qwen-modelname
Sep 5, 2025
Merged

fix: Fiw qwen model name in llama-stack core tests#587
jgarciao merged 1 commit intoopendatahub-io:mainfrom
jgarciao:fix-qwen-modelname

Conversation

@jgarciao
Copy link
Copy Markdown
Contributor

@jgarciao jgarciao commented Sep 5, 2025

  • Fixes how qwen model name is set in llama-stack core tests
  • Fixes test test_model_list, which was failing because there are 2 embedding models included in RHOAI 2.24

Summary by CodeRabbit

  • Chores
    • Updated the default Qwen model identifier to the new naming convention.
    • Model list output now shows namespaced identifiers aligned with the inference provider.
    • Serving configuration now includes the explicit served model name for clearer deployments.
  • Tests
    • Updated tests to reflect the new model naming and listing behavior.

Signed-off-by: Jorge Garcia Oncins <jgarciao@redhat.com>
@jgarciao jgarciao requested a review from a team as a code owner September 5, 2025 10:23
@coderabbitai
Copy link
Copy Markdown
Contributor

coderabbitai bot commented Sep 5, 2025

📝 Walkthrough

Walkthrough

Updates include renaming the Qwen model constant, using it in tests, adding a served model name flag to the vLLM CPU runtime container args, and adjusting a core test to expect a namespaced model identifier and flexible model count.

Changes

Cohort / File(s) Summary
Constants update
utilities/constants.py
Changed QWEN_MODEL_NAME value from "qwen2.5-0.5b-instruct" to "qwen25-05b-instruct".
Test fixtures: inference
tests/fixtures/inference.py
Imported QWEN_MODEL_NAME; vllm_cpu_runtime kserve-container args now include --served-model-name={{.Name}}; qwen_isvc now names resource via QWEN_MODEL_NAME instead of a literal.
Core tests adjustment
tests/llama_stack/core/test_llamastack_core.py
Updated test_model_list to not assert exact count and to expect first model id as VLLM-prefixed QWEN_MODEL_NAME; updated comment on embedding models.

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~10 minutes

✨ Finishing Touches
  • 📝 Generate Docstrings
🧪 Generate unit tests
  • Create PR with unit tests
  • Post copyable unit tests in a comment

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share
🪧 Tips

Chat

There are 3 ways to chat with CodeRabbit:

  • Review comments: Directly reply to a review comment made by CodeRabbit. Example:
    • I pushed a fix in commit <commit_id>, please review it.
    • Open a follow-up GitHub issue for this discussion.
  • Files and specific lines of code (under the "Files changed" tab): Tag @coderabbitai in a new review comment at the desired location with your query.
  • PR comments: Tag @coderabbitai in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples:
    • @coderabbitai gather interesting stats about this repository and render them as a table. Additionally, render a pie chart showing the language distribution in the codebase.
    • @coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.

Support

Need help? Create a ticket on our support page for assistance with any issues or questions.

CodeRabbit Commands (Invoked using PR/Issue comments)

Type @coderabbitai help to get the list of available commands.

Other keywords and placeholders

  • Add @coderabbitai ignore or @coderabbit ignore anywhere in the PR description to prevent this PR from being reviewed.
  • Add @coderabbitai summary to generate the high-level summary at a specific location in the PR description.
  • Add @coderabbitai anywhere in the PR title to generate the title automatically.

Status, Documentation and Community

  • Visit our Status Page to check the current availability of CodeRabbit.
  • Visit our Documentation for detailed information on how to use CodeRabbit.
  • Join our Discord Community to get help, request features, and share feedback.
  • Follow us on X/Twitter for updates and announcements.

@github-actions
Copy link
Copy Markdown

github-actions bot commented Sep 5, 2025

The following are automatically added/executed:

  • PR size label.
  • Run pre-commit
  • Run tox
  • Add PR author as the PR assignee
  • Build image based on the PR

Available user actions:

  • To mark a PR as WIP, add /wip in a comment. To remove it from the PR comment /wip cancel to the PR.
  • To block merging of a PR, add /hold in a comment. To un-block merging of PR comment /hold cancel.
  • To mark a PR as approved, add /lgtm in a comment. To remove, add /lgtm cancel.
    lgtm label removed on each new commit push.
  • To mark PR as verified comment /verified to the PR, to un-verify comment /verified cancel to the PR.
    verified label removed on each new commit push.
  • To Cherry-pick a merged PR /cherry-pick <target_branch_name> to the PR. If <target_branch_name> is valid,
    and the current PR is merged, a cherry-picked PR would be created and linked to the current PR.
  • To build and push image to quay, add /build-push-pr-image in a comment. This would create an image with tag
    pr-<pr_number> to quay repository. This image tag, however would be deleted on PR merge or close action.
Supported labels

{'/hold', '/cherry-pick', '/lgtm', '/wip', '/build-push-pr-image', '/verified'}

Copy link
Copy Markdown
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 3

🧹 Nitpick comments (1)
tests/fixtures/inference.py (1)

35-35: Good: pass served model name to vLLM.
The adapter will expose the namespaced id as expected once registered. Consider lifting 8032 into a shared constant to avoid drift.

-                "args": ["--port=8032", "--model=/mnt/models", "--served-model-name={{.Name}}"],
+                "args": ["--port=8032", "--model=/mnt/models", "--served-model-name={{.Name}}"],  # TODO: use a shared PORT constant
📜 Review details

Configuration used: Path: .coderabbit.yaml

Review profile: CHILL

Plan: Pro

💡 Knowledge Base configuration:

  • MCP integration is disabled by default for public repositories
  • Jira integration is disabled by default for public repositories
  • Linear integration is disabled by default for public repositories

You can enable these sources in your CodeRabbit configuration.

📥 Commits

Reviewing files that changed from the base of the PR and between f074dcf and f45dba3.

📒 Files selected for processing (3)
  • tests/fixtures/inference.py (3 hunks)
  • tests/llama_stack/core/test_llamastack_core.py (1 hunks)
  • utilities/constants.py (1 hunks)
🧰 Additional context used
🧬 Code graph analysis (2)
tests/llama_stack/core/test_llamastack_core.py (1)
tests/llama_stack/constants.py (2)
  • LlamaStackProviders (4-14)
  • Inference (7-8)
tests/fixtures/inference.py (1)
utilities/constants.py (2)
  • RuntimeTemplates (65-78)
  • KServeDeploymentType (6-9)
🔇 Additional comments (1)
tests/fixtures/inference.py (1)

12-12: Importing QWEN_MODEL_NAME into fixtures is the right move.
Keeps the ISVC name and test expectations in sync.

@jgarciao jgarciao merged commit c92cdee into opendatahub-io:main Sep 5, 2025
10 checks passed
@github-actions
Copy link
Copy Markdown

github-actions bot commented Sep 5, 2025

Status of building tag latest: success.
Status of pushing tag latest to image registry: success.

mwaykole pushed a commit to mwaykole/opendatahub-tests that referenced this pull request Jan 23, 2026
Signed-off-by: Jorge Garcia Oncins <jgarciao@redhat.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants