-
Notifications
You must be signed in to change notification settings - Fork 165
[OpenVINO] Transformers 4.56/4.57 support 2 (awaiting npu testing) #1541
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
|
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
|
@IlyasMoutawwakil what are your expectations for merging this PR? We're considering using transformers 4.57 for one of our testing pipelines, so having the corresponding support on optimum-intel's side would be a very good thing. |
|
@CuriousPanCake my expectation is that it is good to go 😁 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
there is an issue with gemma models from new transformers reported from GenAI:
Traceback (most recent call last):
File "/home/jenkins/agent/workspace/DL-Benchmark/prod/WW46-2025.4.0-20398-RC2-Transformers4.57/LO_CPU_acc_wwb_wwb_ref_nat_vs_genai_CPU_ICX/venv/bin/wwb", line 9, in <module>
sys.exit(main())
File "/home/jenkins/agent/workspace/DL-Benchmark/prod/WW46-2025.4.0-20398-RC2-Transformers4.57/LO_CPU_acc_wwb_wwb_ref_nat_vs_genai_CPU_ICX/genai/tools/who_what_benchmark/whowhatbench/wwb.py", line 764, in main
all_metrics_per_question, all_metrics = evaluator.score(
File "/home/jenkins/agent/workspace/DL-Benchmark/prod/WW46-2025.4.0-20398-RC2-Transformers4.57/LO_CPU_acc_wwb_wwb_ref_nat_vs_genai_CPU_ICX/genai/tools/who_what_benchmark/whowhatbench/visualtext_evaluator.py", line 84, in score
predictions = self._generate_data(model_or_data, gen_answer_fn, self.generation_config)
File "/home/jenkins/agent/workspace/DL-Benchmark/prod/WW46-2025.4.0-20398-RC2-Transformers4.57/LO_CPU_acc_wwb_wwb_ref_nat_vs_genai_CPU_ICX/genai/tools/who_what_benchmark/whowhatbench/visualtext_evaluator.py", line 177, in _generate_data
gen_answer_fn(
File "/home/jenkins/agent/workspace/DL-Benchmark/prod/WW46-2025.4.0-20398-RC2-Transformers4.57/LO_CPU_acc_wwb_wwb_ref_nat_vs_genai_CPU_ICX/genai/tools/who_what_benchmark/whowhatbench/wwb.py", line 469, in genai_gen_visual_text
out = model.generate(
RuntimeError: Exception from src/inference/src/cpp/infer_request.cpp:75:
Check '::getPort(port, name, {_impl->get_inputs(), _impl->get_outputs()})' failed at src/inference/src/cpp/infer_request.cpp:77:
Port for tensor name position_ids was not found.
@IlyasMoutawwakil, please double check and clarify why this issue appeared.
Thanks,
Roman
|
@rkazants no idea, there's not enough context in the traceback you shared to know what went wrong exactly |
What does this PR do?
Fixes # (issue)
Before submitting