-
Notifications
You must be signed in to change notification settings - Fork 527
change vllm version to v0.10.2 #5264
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from all commits
4b82ca4
e6fcc65
a5cbefe
d684e6e
22c22e5
cd9b0db
ed0ba83
36613dc
40d3959
4988ca1
d33ef27
68e985e
bb17420
1a894ae
7408479
762152d
1554684
46abfdf
bdacfc7
9400690
928efe7
aefc773
df03922
b8fa81e
d470b5c
2d9f45c
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
|
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Same as above, errors out about tmp folder. |
|
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. benchmark_serving is deprecated so we need to use the bench serve command. |
|
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I want to call this out, we are using latest pytorch, latest cuda and latest vllm version which is a win. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I removed this check because vllm installs deepgemm scripts to tmp folder.