You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: AI-and-Analytics/Getting-Started-Samples/INC-Sample-for-Tensorflow/README.md
+11-11Lines changed: 11 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,15 +1,15 @@
1
-
# `Intel® Neural Compressor (Intel® INC)` Sample for TensorFlow*
1
+
# `Intel® Neural Compressor` Sample for TensorFlow*
2
2
3
3
## Background
4
4
Low-precision inference can speed up inference obviously, by converting the fp32 model to int8 or bf16 model. Intel provides Intel® Deep Learning Boost technology in the Second Generation Intel® Xeon® Scalable Processors and newer Xeon®, which supports to speed up int8 and bf16 model by hardware.
5
5
6
-
Intel®Low Precision Optimization Tool (Intel INC) helps the user to simplify the processing to convert the fp32 model to int8/bf16.
6
+
Intel®Neural Compressor helps the user to simplify the processing to convert the fp32 model to int8/bf16.
7
7
8
-
At the same time, Intel INC will tune the quanization method to reduce the accuracy loss, which is a big blocker for low-precision inference.
8
+
At the same time, Intel® Neural Compressor will tune the quanization method to reduce the accuracy loss, which is a big blocker for low-precision inference.
9
9
10
-
Intel INC is released in Intel® AI Analytics Toolkit and works with Intel® Optimization of TensorFlow*.
10
+
Intel® Neural Compressor is released in Intel® AI Analytics Toolkit and works with Intel® Optimization of TensorFlow*.
11
11
12
-
Please refer to the official website for detailed info and news: [https://github.com/intel/lp-opt-tool](https://github.com/intel/lp-opt-tool)
12
+
Please refer to the official website for detailed info and news: [https://github.com/intel/neural-compressor](https://github.com/intel/neural-compressor)
13
13
14
14
## License
15
15
@@ -19,18 +19,18 @@ Code samples are licensed under the MIT license. See
19
19
Third party program Licenses can be found here: [third-party-programs.txt](https://github.com/oneapi-src/oneAPI-samples/blob/master/third-party-programs.txt)
20
20
21
21
## Purpose
22
-
This sample will show a whole process to build up a CNN model to recognize handwriting number and speed up it by Intel INC.
22
+
This sample will show a whole process to build up a CNN model to recognize handwriting number and speed up it by Intel® Neural Compressor.
23
23
24
-
We will learn how to train a CNN model based on Keras with TensorFlow, use Intel INC to quantize the model and compare the performance to understand the benefit of Intel INC.
24
+
We will learn how to train a CNN model based on Keras with TensorFlow, use Intel® Neural Compressor to quantize the model and compare the performance to understand the benefit of Intel® Neural Compressor.
25
25
26
26
## Key Implementation Details
27
27
28
28
- Use Keras on TensorFlow to build and train the CNN model.
29
29
30
30
31
-
- Define function and class for Intel INC to quantize the CNN model.
31
+
- Define function and class for Intel® Neural Compressor to quantize the CNN model.
32
32
33
-
The Intel INC can run on any Intel® CPU to quantize the AI model.
33
+
The Intel® Neural Compressor can run on any Intel® CPU to quantize the AI model.
34
34
35
35
The quantized AI model has better inference performance than the FP32 model on Intel CPU.
36
36
@@ -47,7 +47,7 @@ We will learn how to train a CNN model based on Keras with TensorFlow, use Intel
47
47
| OS | Linux* Ubuntu* 18.04
48
48
| Hardware | The Second Generation Intel® Xeon® Scalable processor family or newer
49
49
| Software | Intel® oneAPI AI Analytics Toolkit 2021.1 or newer
50
-
| What you will learn | How to use Intel INC tool to quantize the AI model based on TensorFlow and speed up the inference on Intel® Xeon® CPU
50
+
| What you will learn | How to use Intel® Neural Compressor tool to quantize the AI model based on TensorFlow and speed up the inference on Intel® Xeon® CPU
51
51
| Time to complete | 10 minutes
52
52
53
53
## Running Environment
@@ -127,7 +127,7 @@ conda activate tensorflow
127
127
128
128
```
129
129
130
-
### Install Intel INC by Local Channel
130
+
### Install Intel® Neural Compressor by Local Channel
| daal4py | [IntelPython_daal4py_GettingStarted](IntelPython_daal4py_GettingStarted) | Batch linear regression using the python API package daal4py from oneAPI Data Analytics Library (oneDAL) .
20
-
| INC | [INC-Sample-for-Tensorflow](INC-Sample-for-Tensorflow) |Quantize a fp32 model into int8 by Intel® Neural Compressor (INC), and compare the performance between fp32 and int8 .
20
+
| Intel® Neural Compressor | [INC-Sample-for-Tensorflow](INC-Sample-for-Tensorflow) |Quantize a fp32 model into int8 by Intel® Neural Compressor, and compare the performance between fp32 and int8 .
21
21
| Modin | [IntelModin_GettingStarted](IntelModin_GettingStarted) | Run Modin-accelerated Pandas functions and note the performance gain .
22
22
| PyTorch | [IntelPyTorch_GettingStarted](IntelPyTorch_GettingStarted) | A simple training example for PyTorch.
23
23
| TensorFlow | [IntelTensorFlow_GettingStarted](IntelTensorFlow_GettingStarted) | A simple training example for TensorFlow.
0 commit comments