Multi-backend refactor: Alpha release ( INTEL ONLY ) #1338
Replies: 8 comments 19 replies
-
I'd be interested in trying it out. How can I do that? |
Beta Was this translation helpful? Give feedback.
-
I'm interested as well! |
Beta Was this translation helpful? Give feedback.
-
Hello, I'm wondering is the intel backend ready to use? My attempts are all failed with
for either or
despite the model and inputs are all on the cpu. on |
Beta Was this translation helpful? Give feedback.
-
The following code crashes with import torch
from transformers import AutoTokenizer, BitsAndBytesConfig, SiglipTextModel
quantization_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_quant_type="nf4",
)
model = SiglipTextModel.from_pretrained(
"google/siglip-so400m-patch14-384",
quantization_config=quantization_config,
)
tokenizer = AutoTokenizer.from_pretrained(
"google/siglip-so400m-patch14-384"
)
with torch.no_grad():
text = tokenizer(inputs, padding="max_length", return_tensors="pt")
text_features = model(**text).pooler_output.tolist()
return text_features Full stack trace:
|
Beta Was this translation helpful? Give feedback.
-
What model are you trying to run? Also, does it happen if you try "fp4"? |
Beta Was this translation helpful? Give feedback.
-
Hi I've been testing it on Intel GPU python version: Model: DeepSeek-R1-Distill-Qwen-1.5B
Observation 1: Observation 2: with
is this error expected with load_in_4bit? |
Beta Was this translation helpful? Give feedback.
-
I will check this issue as soon as I get back to work. We indeed need an alpha release for INTEL-ONLY. |
Beta Was this translation helpful? Give feedback.
-
Hi, I am running Openvla quantization on Intel Arc 770. While runing on
The error report Running figure: ReproduceEnvironment
Reproduce quick scripts
Any comments on Bus Error? |
Beta Was this translation helpful? Give feedback.
-
This space is intended to receive feedback from users that are willing to help us by alpha testing the current implementation of the Intel backends (Intel CPU and XPU).
Issues to discuss could be
Thanks everyone for you kind support and please remember to have a constructive tone 🤗
Beta Was this translation helpful? Give feedback.
All reactions