Skip to content

bump llm-d-kv-cache-manager version#359

Merged
github-actions[bot] merged 1 commit intollm-d:mainfrom
vMaroon:kvc-v0.3.1
Sep 25, 2025
Merged

bump llm-d-kv-cache-manager version#359
github-actions[bot] merged 1 commit intollm-d:mainfrom
vMaroon:kvc-v0.3.1

Conversation

@vMaroon
Copy link
Member

@vMaroon vMaroon commented Sep 25, 2025

Summary

Upgrade the llm-d-kv-cache-manager dependency to v0.3.1, which adds support for the updated vLLM v0.11.0 KVEvents format while maintaining backwards compatibility. This was tested and benchmarked with vLLM v0.10.2 and v0.11.0, built into the image: quay.io/vmaroon/llm-d-inference-scheduler:v0.3.1.

This is required for the llm-d v0.3.0 release.

Context: since the llm-d release took (much) longer than planned, and the vLLM native CPU offloading work landed, this llm-d v0.4.0 change is required now.

Signed-off-by: Maroon Ayoub <maroon.ayoub@ibm.com>
@kfswain
Copy link
Collaborator

kfswain commented Sep 25, 2025

/lgtm
/approve

@github-actions github-actions bot added the lgtm "Looks good to me", indicates that a PR is ready to be merged. label Sep 25, 2025
@kfswain kfswain mentioned this pull request Sep 25, 2025
@github-actions github-actions bot merged commit 4829137 into llm-d:main Sep 25, 2025
6 checks passed
kfswain pushed a commit that referenced this pull request Sep 26, 2025
Signed-off-by: Maroon Ayoub <maroon.ayoub@ibm.com>
usize pushed a commit to usize/llm-d-inference-scheduler that referenced this pull request Oct 6, 2025
Signed-off-by: Maroon Ayoub <maroon.ayoub@ibm.com>
Signed-off-by: usize <mofoster@redhat.com>
usize pushed a commit to usize/llm-d-inference-scheduler that referenced this pull request Oct 8, 2025
Signed-off-by: Maroon Ayoub <maroon.ayoub@ibm.com>
Signed-off-by: usize <mofoster@redhat.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

lgtm "Looks good to me", indicates that a PR is ready to be merged.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants