Skip to content

Commit b3e43ec

Browse files
authored
Merge branch 'master' into master
2 parents a77ef02 + 3de2766 commit b3e43ec

File tree

44 files changed

+3015
-306
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

44 files changed

+3015
-306
lines changed

.github/copilot-instructions.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -99,6 +99,7 @@ Before posting any comment, apply this gate:
9999
- Do not review vendored/third-party sources under `thirdparty/` unless the PR explicitly modifies integration or patch logic.
100100
- Do not enforce component-specific runtime behavior rules on docs-only PRs.
101101
- For generated or auto-updated files (for example stubs/version bumps), comment only if there is clear breakage risk.
102+
- Python stub files (`.pyi`) under `src/bindings/python/` are auto-generated by a scheduled GitHub Actions workflow (`update_pyapi_stubs.yml`) using `pybind11-stubgen`. Do not request manual fixes to stub files.
102103

103104
## Component-Specific Expectations
104105

.github/workflows/linux_conditional_compilation.yml

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -192,12 +192,14 @@ jobs:
192192
-DENABLE_DEBUG_CAPS=ON \
193193
-DCMAKE_C_COMPILER_LAUNCHER=${{ env.CMAKE_C_COMPILER_LAUNCHER }} \
194194
-DCMAKE_CXX_COMPILER_LAUNCHER=${{ env.CMAKE_CXX_COMPILER_LAUNCHER }} \
195+
-DCMAKE_JOB_POOLS:STRING=link_jobs=8 \
196+
-DCMAKE_JOB_POOL_LINK:STRING=link_jobs \
195197
-S ${OPENVINO_REPO} \
196198
-B ${BUILD_DIR}
197199
198200
- name: Cmake build - CC COLLECT
199201
run: |
200-
cmake --build ${BUILD_DIR} --parallel 8 --config ${{ env.CMAKE_BUILD_TYPE }}
202+
cmake --build ${BUILD_DIR} --parallel $(nproc) --config ${{ env.CMAKE_BUILD_TYPE }}
201203
cmake --build ${BUILD_DIR} --parallel $(nproc) --config ${{ env.CMAKE_BUILD_TYPE }} --target sea_itt_lib
202204
203205
- name: Show sccache stats
@@ -289,7 +291,7 @@ jobs:
289291
defaults:
290292
run:
291293
shell: bash
292-
runs-on: aks-linux-16-cores-32gb
294+
runs-on: aks-linux-8-cores-16gb
293295
container:
294296
image: ${{ fromJSON(needs.docker.outputs.images).ov_build.ubuntu_22_04_x64_cc }}
295297
volumes:

docs/articles_en/about-openvino/performance-benchmarks.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@ For a listing of all platforms and configurations used for testing, refer to the
158158
**Disclaimers**
159159

160160
* System configurations used for Intel® Distribution of OpenVINO™ toolkit performance results
161-
are based on release 2025.4, as of December 1st, 2025.
161+
are based on release 2026.0, as of February 25, 2026.
162162

163163
* OpenVINO Model Server performance results are based on release 2025.3, as of September 3rd, 2025.
164164

docs/articles_en/about-openvino/performance-benchmarks/generative-ai-performance.rst

Lines changed: 22 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -3,20 +3,31 @@ Most Efficient Large Language Models for AI PC
33

44
This page is regularly updated to help you identify the best-performing LLMs on the
55
Intel® Core™ Ultra processor family and AI PCs.
6-
The current data is as of OpenVINO 2025.3, 3 September 2025.
6+
The current data is as of OpenVINO 2026.0, 26 February 2026.
77

88
The tables below list the key performance indicators for inference on built-in GPUs.
99

1010

1111

1212
.. tab-set::
1313

14-
.. tab-item:: 7-255H
14+
.. tab-item:: 5-338H
1515

1616
.. data-table::
1717
:class: modeldata stripe
1818
:name: supportedModelsTable_V1
1919
:header-rows: 1
20+
:file: ../../_static/benchmarks_files/llm_models_5-338H.csv
21+
:data-column-hidden: [3,4,6]
22+
:data-order: [[ 0, "asc" ]]
23+
:data-page-length: 10
24+
25+
.. tab-item:: 7-255H
26+
27+
.. data-table::
28+
:class: modeldata stripe
29+
:name: supportedModelsTable_V2
30+
:header-rows: 1
2031
:file: ../../_static/benchmarks_files/llm_models_7-255H.csv
2132
:data-column-hidden: [3,4,6]
2233
:data-order: [[ 0, "asc" ]]
@@ -26,7 +37,7 @@ The tables below list the key performance indicators for inference on built-in G
2637

2738
.. data-table::
2839
:class: modeldata stripe
29-
:name: supportedModelsTable_V2
40+
:name: supportedModelsTable_V3
3041
:header-rows: 1
3142
:file: ../../_static/benchmarks_files/llm_models_7-258V.csv
3243
:data-column-hidden: [3,4,6]
@@ -36,7 +47,7 @@ The tables below list the key performance indicators for inference on built-in G
3647

3748
.. data-table::
3849
:class: modeldata stripe
39-
:name: supportedModelsTable_V3
50+
:name: supportedModelsTable_V4
4051
:header-rows: 1
4152
:file: ../../_static/benchmarks_files/llm_models_7-165H.csv
4253
:data-column-hidden: [3,4,6]
@@ -64,6 +75,13 @@ The tables below list the key performance indicators for inference on built-in G
6475

6576
:material-regular:`download;1.5em` Get system descriptions [PDF]
6677

78+
.. button-link:: ../../_static/benchmarks_files/llm_models_5-338H.csv
79+
:color: primary
80+
:outline:
81+
:expand:
82+
83+
:material-regular:`download;1.5em` Get the 5-338H table [PDF]
84+
6785
.. button-link:: ../../_static/benchmarks_files/llm_models_7-258V.csv
6886
:color: primary
6987
:outline:

docs/articles_en/about-openvino/performance-benchmarks/model-accuracy-int8-fp32.rst

Lines changed: 69 additions & 55 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ the table for more information.
1111
* A - Intel® Core™ Ultra 9-185H (AVX2), INT8 and FP32
1212
* B - Intel® Xeon® 6338, (VNNI), INT8 and FP32
1313
* C - Intel® Xeon 6972P (VNNI, AMX), INT8, BF16, FP32
14-
* D - Intel® Arc-B580, INT8 and FP16
14+
* D - Intel® Arc-B60, INT8 and FP16
1515

1616

1717
.. list-table:: Model Accuracy for INT8
@@ -27,10 +27,10 @@ the table for more information.
2727
* - bert-base-cased
2828
- SST-2_bert_cased_padded
2929
- spearman@cosine
30-
- 2.57%
31-
- 2.65%
32-
- 2.95%
33-
- 2.63%
30+
- 2.60%
31+
- 2.70%
32+
- 3.00%
33+
- 2.60%
3434
* - Detectron-V2
3535
- COCO2017_detection_91cl_bkgr
3636
- coco_orig_precision
@@ -43,22 +43,22 @@ the table for more information.
4343
- accuracy @ top1
4444
- -0.91%
4545
- -0.93%
46-
- -0.91%
47-
- -1.03%
46+
- -1.01%
47+
- -1.01%
4848
* - resnet-50
4949
- ImageNet2012
5050
- accuracy @ top1
51-
- -0.12%
52-
- -0.12%
53-
- -0.15%
54-
- -0.15%
51+
- 0.73%
52+
- 0.73%
53+
- 0.73%
54+
- 0.73%
5555
* - ssd-resnet34-1200
5656
- COCO2017_detection_80cl_bkgr
5757
- map
58-
- 0.00%
59-
- 0.00%
60-
- 0.03%
61-
- 0.07%
58+
- 0.02%
59+
- 0.02%
60+
- 0.02%
61+
- 0.02%
6262
* - yolo_v11
6363
- COCO2017_detection_80cl
6464
- AP@0.5:0.05:0.95
@@ -93,31 +93,31 @@ the table for more information.
9393
* - mobilenet-v2
9494
- ImageNet2012
9595
- accuracy @ top1
96-
- 0.00%
97-
- 0.00%
98-
- 0.02%
99-
- 0.02%
96+
- -0.01%
97+
- -0.01%
98+
- -0.01%
99+
- -0.01%
100100
* - resnet-50
101101
- ImageNet2012
102102
- accuracy @ top1
103103
- 0.00%
104104
- 0.00%
105105
- 0.00%
106-
- -0.01%
106+
- 0.00%
107107
* - ssd-resnet34-1200
108108
- COCO2017_detection_80cl_bkgr
109109
- map
110110
- 0.02%
111111
- 0.02%
112112
- 0.02%
113-
- -0.23%
113+
- 0.02%
114114
* - yolo_v11
115115
- COCO2017_detection_80cl
116116
- AP@0.5:0.05:0.95
117-
- 0.03%
118-
- -2.21%
119-
- -2.21%
120-
- -2.21%
117+
- 0.00%
118+
- -2.18%
119+
- -2.18%
120+
- -2.18%
121121
.. list-table:: Model Accuracy for AMX-FP16, AMX-INT4, Arc-FP16 and Arc-INT4 (Arc™ B-series)
122122
:header-rows: 1
123123

@@ -132,79 +132,93 @@ the table for more information.
132132
- Data Default WWB
133133
- Similarity
134134
- 98.1%
135-
- 94.4%
136-
- 99.5%
135+
- 94.1%
136+
- 99.6%
137137
- 94.0%
138138
* - DeepSeek-R1-Distill-Qwen-1.5B
139139
- Data Default WWB
140140
- Similarity
141141
- 96.5%
142142
- 92.4%
143143
- 99.7%
144-
- 92.3%
144+
- 92.7%
145145
* - Gemma-3-4B-it
146146
- Data Default WWB
147147
- Similarity
148-
- 92.0%
148+
- 92.2%
149149
- 83.9%
150150
-
151-
- 84.9%
151+
- 92.9%
152+
* - GPT-OSS-20B
153+
- Data Default WWB
154+
- Similarity
155+
- 94.9%
156+
- 92.2%
157+
-
158+
- 92.9%
152159
* - Llama-2-7B-chat
153160
- Data Default WWB
154161
- Similarity
155162
- 99.3%
156-
- 93.4%
157-
- 99.8%
158-
- 93.4%
163+
- 93.3%
164+
- 99.6%
165+
- 93.3%
159166
* - Llama-3-8B
160167
- Data Default WWB
161168
- Similarity
162169
- 98.8%
163-
- 94.3%
164-
- 99.7%
165-
- 94.5%
170+
- 94.7%
171+
- 99.9%
172+
- 94.4%
166173
* - Llama-3.2-3b-instruct
167174
- Data Default WWB
168175
- Similarity
169-
- 97.9%
170-
- 94.2%
171-
- 99.7%
172-
- 94.1%
173-
* - Phi4-mini-instruct
176+
- 98.3%
177+
- 94.8%
178+
- 99.9%
179+
- 94.3%
180+
* - MiniCPM-V-2.6
174181
- Data Default WWB
175182
- Similarity
183+
- 90.6%
184+
- 90.1%
185+
- 88.1%
176186
- 89.1%
187+
* - Phi4-mini-instruct
188+
- Data Default WWB
189+
- Similarity
190+
- 95.1%
191+
- 92.5%
192+
- 99.1%
177193
- 92.1%
178-
- 99.5%
179-
- 92.4%
180-
* - Qwen2-VL-7B
194+
* - Qwen2.5-VL-7B
181195
- Data Default WWB
182196
- Similarity
183-
- 97.5%
184-
- 88.1%
197+
- 93.7%
198+
- 90.7%
185199
- 99.8%
186-
- 91.4%
200+
- 89.9%
187201
* - Qwen3-8B
188202
- Data Default WWB
189203
- Similarity
190-
- 97.8%
191-
- 92.3%
192-
-
193-
- 93.0%
204+
- 97.9%
205+
- 93.6%
206+
- 99.8%
207+
- 92.8%
194208
* - Flux.1-schnell
195209
- Data Default WWB
196210
- Similarity
197211
- 95.4%
198212
- 96.1%
199213
-
200-
- 92.1%
214+
- 95.1%
201215
* - Stable-Diffusion-V1-5
202216
- Data Default WWB
203217
- Similarity
204-
- 96.3%
205-
- 93.3%
218+
- 96.7%
219+
- 95.5%
206220
- 99.5%
207-
- 93.7%
221+
- 92.1%
208222

209223
Notes: For all accuracy metrics a "-", (minus sign), indicates an accuracy drop.
210224
The Similarity metric is the distance from "perfect" and as such always positive.

docs/articles_en/about-openvino/performance-benchmarks/performance-benchmarks-faq.rst

Lines changed: 15 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -59,6 +59,10 @@ Performance Information F.A.Q.
5959
- Hugginface
6060
- Text-To-Text Decoder-only
6161
- 128K
62+
* - `GPT-OSS-20B <https://huggingface.co/openai/gpt-oss-20b>`__
63+
- Hugginface
64+
- Reasoning, agentic and developer use cases.
65+
- 128K
6266
* - `Llama-2-7b-chat <https://www.llama.com/>`__
6367
- Meta AI
6468
- Auto regressive language
@@ -71,13 +75,21 @@ Performance Information F.A.Q.
7175
- Meta AI
7276
- Auto regressive language
7377
- 128K
78+
* - `MiniCPM-v-2_6 <https://huggingface.co/openbmb/MiniCPM-V-2_6>`__
79+
- OpenBMB
80+
- Multimodal, image-text to text
81+
- 32K
82+
* - `Mistral-7B-Instruct-V0.3 <https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3>`__
83+
- Mistral AI
84+
- Auto regressive language
85+
- 32K
7486
* - `Phi4-mini-Instruct <https://huggingface.co/microsoft/Phi-4-mini-instruct>`__
75-
- Huggingface
87+
- Hugginface
7688
- Auto regressive language
7789
- 4096
78-
* - `Qwen-2-VL-7B-instruct <https://huggingface.co/Qwen/Qwen2-VL-7B-instruct>`__
90+
* - `Qwen-2.5-VL-7B-instruct <https://huggingface.co/Qwen/Qwen2.5-VL-7B-instruct>`__
7991
- Huggingface
80-
- Auto regressive language
92+
- Auto regressive language, Agentic
8193
- 128K
8294
* - `Qwen-3-8B <https://huggingface.co/Qwen/Qwen3-8B>`__
8395
- Huggingface

0 commit comments

Comments
 (0)