Skip to content

Local inference with Blackwell (RTX 50 series) GPUs defaults to using CPU #64

@theendlesswatch

Description

@theendlesswatch

This is more of a heads up than an issue, though maybe a little note in the README for other folks would be nice too.
Anyone who has an RTX 50 series GPU should activate the Python venv then run:

pip install --upgrade torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128

This will upgrade pytorch to the latest stable version (2.8.0 as of this writing) compiled with CUDA 12.8 to enable GPU acceleration for Blackwell GPUs. I've personally tested this and everything works as expected. Though as previously noted by the author of the repo, it's currently not possible to fully utilize the GPU, save for batch inference. However, it'll still be much faster than CPU inference.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions