You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/source/benchmark.md
+4-2
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,7 @@
1
-
Benchmarking
2
-
============
1
+
Benchmarking (Deprecated)
2
+
===============
3
+
> Since version >= 3.3, this functionality has been replaced by [incbench](https://github.com/intel/neural-compressor/blob/master/docs/source/3x/benchmark.md), which is easier to use.
4
+
----------------------------
3
5
1.[Introduction](#Introduction)
4
6
2.[Benchmark Support Matrix](#Benchmark-Support-Matrix)
5
7
3.[Get Started with Benchmark](#Get-Started-with-Benchmark)
- Please follow this README.md exactly:[QAT](../../examples/pytorch/image_recognition/torchvision_models/quantization/qat/eager/distributed)
170
170
171
-
### TensorFlow Examples:
171
+
### TensorFlow Examples: (Deprecated)
172
172
- TensorFlow example-1: 'ResNet50 V1.0' PTQ (Post Training Quantization) with distributed inference
173
173
- Step-1: Please cd (change directory) to the [TensorFlow Image Recognition Example](../../examples/tensorflow/image_recognition) and follow the readme to run PTQ, ensure that PTQ of 'ResNet50 V1.0' can be successfully executed.
174
174
- Step-2: We only need to modify the [resnet50_v1.yaml](../../examples/tensorflow/image_recognition/tensorflow_models/quantization/ptq/resnet50_v1.yaml), add a line 'distributed: True' in the 'evaluation' field.
Copy file name to clipboardExpand all lines: docs/source/quantization_mixed_precision.md
+1-1
Original file line number
Diff line number
Diff line change
@@ -15,7 +15,7 @@ q_model = quantization.fit(
15
15
)
16
16
```
17
17
18
-
### Tensorflow
18
+
### Tensorflow (Deprecated)
19
19
20
20
Intel has worked with the TensorFlow development team to enhance TensorFlow to include bfloat16 data support for CPUs. For more information about BF16 in TensorFlow, please read [Accelerating AI performance on 3rd Gen Intel® Xeon® Scalable processors with TensorFlow and Bfloat16](https://blog.tensorflow.org/2020/06/accelerating-ai-performance-on-3rd-gen-processors-with-tensorflow-bfloat16.html).
Copy file name to clipboardExpand all lines: examples/README.md
+17-14
Original file line number
Diff line number
Diff line change
@@ -2,25 +2,28 @@ Examples
2
2
==========
3
3
Intel® Neural Compressor validated examples with multiple compression techniques, including quantization, pruning, knowledge distillation and orchestration. Part of the validated cases can be found in the example tables, and the release data is available [here](../docs/source/validated_model_list.md).
4
4
5
+
> Note: `3.x_api` here is the target example folder for version >= 3.0, which contains pytorch and tensorflow related examples. TensorFlow and ONNX related examples here are no longer maintained.
6
+
7
+
5
8
# Quick Get Started Notebook Examples
6
-
*[Quick Get Started Notebook of Intel® Neural Compressor for ONNXRuntime](/examples/notebook/onnxruntime/Quick_Started_Notebook_of_INC_for_ONNXRuntime.ipynb)
9
+
*[Quick Get Started Notebook of Intel® Neural Compressor for ONNXRuntime (Deprecated)](/examples/notebook/onnxruntime/Quick_Started_Notebook_of_INC_for_ONNXRuntime.ipynb)
7
10
8
-
*[Quick Get Started Notebook of Intel® Neural Compressor for Tensorflow](/examples/notebook/tensorflow/resnet/resnet_quantization.ipynb)
11
+
*[Quick Get Started Notebook of Intel® Neural Compressor for Tensorflow (Deprecated)](/examples/notebook/tensorflow/resnet/resnet_quantization.ipynb)
9
12
10
13
*[Quick Get Started Notebook of Intel® Neural Compressor for Pytorch](/examples/notebook/pytorch/Quick_Started_Notebook_of_INC_for_Pytorch.ipynb)
11
14
12
15
# Helloworld Examples
13
16
*[torch_llm](/examples/helloworld/torch_llm): apply the weight-only quantization to LLMs.
14
17
*[torch_non_llm](/examples/helloworld/torch_non_llm): apply the static quantization to non-LLMs.
15
-
*[tf_example1](/examples/helloworld/tf_example1): quantize with built-in dataloader and metric.
16
-
*[tf_example2](/examples/helloworld/tf_example2): quantize keras model with customized metric and dataloader.
17
-
*[tf_example3](/examples/helloworld/tf_example3): convert model with mix precision.
18
-
*[tf_example4](/examples/helloworld/tf_example4): quantize checkpoint with dummy dataloader.
19
-
*[tf_example5](/examples/helloworld/tf_example5): config performance and accuracy measurement.
20
-
*[tf_example6](/examples/helloworld/tf_example6): use default user-facing APIs to quantize a pb model.
21
-
*[tf_example7](/examples/helloworld/tf_example7): quantize and benchmark with pure python API.
18
+
*[tf_example1 (Deprecated)](/examples/helloworld/tf_example1): quantize with built-in dataloader and metric.
19
+
*[tf_example2 (Deprecated)](/examples/helloworld/tf_example2): quantize keras model with customized metric and dataloader.
20
+
*[tf_example3 (Deprecated)](/examples/helloworld/tf_example3): convert model with mix precision.
21
+
*[tf_example4 (Deprecated)](/examples/helloworld/tf_example4): quantize checkpoint with dummy dataloader.
22
+
*[tf_example5 (Deprecated)](/examples/helloworld/tf_example5): config performance and accuracy measurement.
23
+
*[tf_example6 (Deprecated)](/examples/helloworld/tf_example6): use default user-facing APIs to quantize a pb model.
24
+
*[tf_example7 (Deprecated)](/examples/helloworld/tf_example7): quantize and benchmark with pure python API.
*[Performance of FP32 Vs. INT8 ResNet50 Model](/examples/notebook/perf_fp32_int8_tf): compare existed FP32 & INT8 ResNet50 model directly.
1476
-
*[Intel® Neural Compressor Sample for TensorFlow*](/examples/notebook/tensorflow/alexnet_mnist): an End-To-End pipeline to build up a CNN model by TensorFlow to recognize handwriting number and speed up AI model by Intel® Neural Compressor.
1477
-
*[Accelerate VGG19 Inference on Intel® Gen4 Xeon® Sapphire Rapids](/examples/notebook/tensorflow/vgg19_ibean): an End-To-End pipeline to train VGG19 model by transfer learning based on pre-trained model from [TensorFlow Hub](https://tfhub.dev); quantize it by Intel® Neural Compressor on Intel® Gen4 Xeon® Sapphire Rapids.
1478
+
*[Performance of FP32 Vs. INT8 ResNet50 Model (Deprecated)](/examples/notebook/perf_fp32_int8_tf): compare existed FP32 & INT8 ResNet50 model directly.
1479
+
*[Intel® Neural Compressor Sample for TensorFlow* (Deprecated)](/examples/notebook/tensorflow/alexnet_mnist): an End-To-End pipeline to build up a CNN model by TensorFlow to recognize handwriting number and speed up AI model by Intel® Neural Compressor.
1480
+
*[Accelerate VGG19 Inference on Intel® Gen4 Xeon® Sapphire Rapids (Deprecated)](/examples/notebook/tensorflow/vgg19_ibean): an End-To-End pipeline to train VGG19 model by transfer learning based on pre-trained model from [TensorFlow Hub](https://tfhub.dev); quantize it by Intel® Neural Compressor on Intel® Gen4 Xeon® Sapphire Rapids.
0 commit comments