Skip to content

Conversation

@dtrifiro
Copy link

  • using uv instead of pip
  • adding --no-binary=vllm to force a vllm rebuild

@dtrifiro
Copy link
Author

Based on #57

uv pip install \
--force-reinstall \
--no-cache-dir \
--no-binary "vllm" \
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I guess this works well in CI scenarios, but we need a way to make sure that when users install the plugin from pypi, they have

VLLM_TARGET_DEVICE=cpu
UV_EXTRA_INDEX_URL=https://download.pytorch.org/whl/cpu
UV_INDEX_STRATEGY="unsafe-best-match"`

set and use --no-binary=vllm.

Perhaps we can provide a vllm cpu wheel that you can install via an extra index url that could simplify this process. More thought needed.

@dtrifiro dtrifiro force-pushed the do-not-use-cuda-torch-improvements branch from 1f513bc to 4396b22 Compare August 29, 2025 09:09
Copy link
Collaborator

@rebel-jonghewk rebel-jonghewk left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants