Skip to content

Add Docstring for TF 3x API and Torch 3x Mixed Precision #1944

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 9 commits into from
Jul 25, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 3 additions & 1 deletion .azure-pipelines/scripts/codeScan/pydocstyle/scan_path.txt
Original file line number Diff line number Diff line change
Expand Up @@ -22,4 +22,6 @@
/neural_compressor/torch/export
/neural_compressor/common
/neural_compressor/torch/algorithms/weight_only
/neural_compressor/torch/algorithms/layer_wise
/neural_compressor/torch/algorithms/layer_wise
/neural_compressor/torch/algorithms/mixed_precision
/neural_compressor/tensorflow
33 changes: 24 additions & 9 deletions docs/source/3x/TensorFlow.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,12 +2,16 @@ TensorFlow
===============


1. [Introduction](#introduction)
2. [API for TensorFlow](#api-for-tensorflow)
3. [Support Matrix](#support-matrix)
3.1 [Quantization Scheme](#quantization-scheme)
3.2 [Quantization Approaches](#quantization-approaches)
3.3 [Backend and Device](#backend-and-device)
- [TensorFlow](#tensorflow)
- [Introduction](#introduction)
- [API for TensorFlow](#api-for-tensorflow)
- [Support Matrix](#support-matrix)
- [Quantization Scheme](#quantization-scheme)
- [Quantization Approaches](#quantization-approaches)
- [Post Training Static Quantization](#post-training-static-quantization)
- [Smooth Quantization](#smooth-quantization)
- [Mixed Precision](#mixed-precison)
- [Backend and Device](#backend-and-device)

## Introduction

Expand Down Expand Up @@ -152,9 +156,16 @@ The supported Quantization methods for TensorFlow and Keras are listed below:
<td align="center"><a href="https://github.com/tensorflow/tensorflow">TensorFlow</a>/<a href="https://github.com/Intel-tensorflow/tensorflow">Intel TensorFlow</a></td>
</tr>
<tr>
<td rowspan="2" align="center">Smooth Quantization(SQ)</td>
<td rowspan="2" align="center">weights</td>
<td rowspan="2" align="center">calibration</td>
<td rowspan="1" align="center">Smooth Quantization(SQ)</td>
<td rowspan="1" align="center">weights</td>
<td rowspan="1" align="center">calibration</td>
<td align="center">Tensorflow</td>
<td align="center"><a href="https://github.com/tensorflow/tensorflow">TensorFlow</a>/<a href="https://github.com/Intel-tensorflow/tensorflow">Intel TensorFlow</a></td>
</tr>
<tr>
<td rowspan="1" align="center">Mixed Precision(MP)</td>
<td rowspan="1" align="center">weights and activations</td>
<td rowspan="1" align="center">NA</td>
<td align="center">Tensorflow</td>
<td align="center"><a href="https://github.com/tensorflow/tensorflow">TensorFlow</a>/<a href="https://github.com/Intel-tensorflow/tensorflow">Intel TensorFlow</a></td>
</tr>
Expand All @@ -175,6 +186,10 @@ Smooth Quantization (SQ) is an advanced quantization technique designed to optim

Refer to the [SQ Guide](./TF_SQ.md) for detailed information.

##### Mixed Precision
The Mixed Precision (MP) is enabled with Post Training Static Quantization. Once `BF16` is supported on machine, the matched operators will be automatically converted.


#### Backend and Device
Intel(R) Neural Compressor supports TF GPU with [ITEX-XPU](https://github.com/intel/intel-extension-for-tensorflow). We will automatically run model on GPU by checking if it has been installed.

Expand Down
2 changes: 1 addition & 1 deletion neural_compressor/tensorflow/algorithms/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

"""Initialize all algorithms."""

from neural_compressor.tensorflow.algorithms.smoother import SmoothQuant
from neural_compressor.tensorflow.algorithms.static_quant import KerasAdaptor, TensorFlowAdaptor, Tensorflow_ITEXAdaptor
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize classes of smooth quant algorithm."""

from neural_compressor.tensorflow.algorithms.smoother.core import SmoothQuant
from neural_compressor.tensorflow.algorithms.smoother.scaler import (
Expand Down
1 change: 1 addition & 0 deletions neural_compressor/tensorflow/algorithms/smoother/core.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""The core components of sq."""

from typing import Callable, Dict

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize classes of static quant algorithm."""

from neural_compressor.tensorflow.algorithms.static_quant.keras import KerasAdaptor
from neural_compressor.tensorflow.algorithms.static_quant.tensorflow import TensorFlowAdaptor, Tensorflow_ITEXAdaptor
7 changes: 4 additions & 3 deletions neural_compressor/tensorflow/algorithms/static_quant/keras.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""The quantization classes for Keras."""

import copy
import json
Expand Down Expand Up @@ -373,9 +374,9 @@ def _calibrate(self, model, dataloader=None, calib_interation=None):
"""Apply calibration.

Args:
model (tf.keras.Model): The model inserted with FakeQuant layers for calibration.
model(tf.keras.Model): The model inserted with FakeQuant layers for calibration.
dataloader(object): The calibration dataloader used to load quantization dataset.
iteration(int): The iteration of calibration.
calib_interation(int): The iteration of calibration.
"""
# run eagerly to fetch the numpy min/max
results = {}
Expand Down Expand Up @@ -580,9 +581,9 @@ def _one_shot_query(self):

def _get_specified_version_cfg(self, data):
"""Get the configuration for the current runtime.

If there's no matched configuration in the input yaml, we'll
use the `default` field of yaml.

Args:
data (Yaml content): input yaml file.

Expand Down
1 change: 1 addition & 0 deletions neural_compressor/tensorflow/keras/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize classes for Keras quantization."""

from neural_compressor.tensorflow.keras.quantization import (
StaticQuantConfig,
Expand Down
1 change: 1 addition & 0 deletions neural_compressor/tensorflow/keras/layers/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize custom layers for Keras quantization."""

from neural_compressor.tensorflow.keras.layers.conv2d import QConv2D
from neural_compressor.tensorflow.keras.layers.dense import QDense
Expand Down
14 changes: 14 additions & 0 deletions neural_compressor/tensorflow/keras/layers/conv2d.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize custom conv2d layers for Keras quantization."""

import json

Expand All @@ -33,6 +34,8 @@
if version1_gte_version2(tf.__version__, "2.16.1"): # pragma: no cover

class QConv2D(BaseConv):
"""The custom quantized Conv2D layer."""

def __init__(
self,
name,
Expand Down Expand Up @@ -65,6 +68,7 @@ def __init__(
quant_axis=None,
**kwargs
):
"""Initialize custom quantized Conv2D layer."""
super(QConv2D, self).__init__(
name=name,
rank=2,
Expand Down Expand Up @@ -100,6 +104,7 @@ def __init__(
self.quant_axis = quant_axis

def call(self, inputs):
"""The __call__ function of custom quantized Conv2D layer."""
if self.quant_status == "calib" and not isinstance(inputs, tf.keras.KerasTensor):
if self.granularity == "per_tensor":
self.act_min_value = tf.math.reduce_min(inputs)
Expand Down Expand Up @@ -168,9 +173,11 @@ def call(self, inputs):

@classmethod
def from_config(cls, config):
"""Deserialize this class from a config dict."""
return cls(**config)

def get_config(self):
"""Serialize this class to a config dict."""
config = super(QConv2D, self).get_config()
config.update(
{
Expand All @@ -193,6 +200,8 @@ def get_config(self):
else:

class QConv2D(Conv):
"""The custom quantized Conv2D layer."""

def __init__(
self,
name,
Expand Down Expand Up @@ -225,6 +234,7 @@ def __init__(
quant_axis=None,
**kwargs
):
"""Initialize custom quantized Conv2D layer."""
super(QConv2D, self).__init__(
name=name,
rank=2,
Expand Down Expand Up @@ -260,6 +270,7 @@ def __init__(
self.quant_axis = quant_axis

def call(self, inputs):
"""The __call__ function of custom quantized Conv2D layer."""
if self.quant_status == "calib":
if self.granularity == "per_tensor":
self.act_min_value = tf.math.reduce_min(inputs)
Expand Down Expand Up @@ -328,9 +339,11 @@ def call(self, inputs):

@classmethod
def from_config(cls, config):
"""Deserialize this class from a config dict."""
return cls(**config)

def get_config(self):
"""Serialize this class to a config dict."""
config = super(QConv2D, self).get_config()
config.update(
{
Expand All @@ -352,6 +365,7 @@ def get_config(self):


def initialize_int8_conv2d(fp32_layer, q_config):
"""Initialize int8 conv2d."""
kwargs = fp32_layer.get_config()

param_list = [
Expand Down
8 changes: 8 additions & 0 deletions neural_compressor/tensorflow/keras/layers/dense.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize custom dense layers for Keras quantization."""

import json

Expand All @@ -25,6 +26,8 @@


class QDense(Dense):
"""The custom quantized Dense layer."""

def __init__(
self,
name,
Expand All @@ -51,6 +54,7 @@ def __init__(
quant_axis=None,
**kwargs
):
"""Initialize custom quantized Dense layer."""
super(QDense, self).__init__(
name=name,
units=units,
Expand Down Expand Up @@ -79,6 +83,7 @@ def __init__(
self.quant_axis = quant_axis

def call(self, inputs):
"""The __call__ function of custom quantized Dense layer."""
if self.quant_status == "calib" and not (
version1_gte_version2(tf.__version__, "2.16.1") and isinstance(inputs, tf.keras.KerasTensor)
):
Expand Down Expand Up @@ -144,9 +149,11 @@ def call(self, inputs):

@classmethod
def from_config(cls, config):
"""Deserialize this class from a config dict."""
return cls(**config)

def get_config(self):
"""Serialize this class to a config dict."""
config = super(QDense, self).get_config()
config.update(
{
Expand All @@ -168,6 +175,7 @@ def get_config(self):


def initialize_int8_dense(fp32_layer, q_config):
"""Initialize int8 dense."""
kwargs = fp32_layer.get_config()

param_list = [
Expand Down
15 changes: 15 additions & 0 deletions neural_compressor/tensorflow/keras/layers/depthwise_conv2d.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize custom depthwise conv2d layers for Keras quantization."""

import json

Expand All @@ -35,6 +36,8 @@
if version1_gte_version2(tf.__version__, "2.16.1"):

class QDepthwiseConv2D(BaseDepthwiseConv): # pragma: no cover
"""The custom quantized DepthwiseConv2D layer."""

def __init__(
self,
kernel_size,
Expand Down Expand Up @@ -65,6 +68,7 @@ def __init__(
quant_axis=None,
**kwargs
):
"""Initialize custom quantized DepthwiseConv2D layer."""
super().__init__(
2,
kernel_size=kernel_size,
Expand Down Expand Up @@ -98,6 +102,7 @@ def __init__(
self.quant_axis = quant_axis

def call(self, inputs):
"""The __call__ function of custom quantized DepthwiseConv2D layer."""
if self.quant_status == "calib" and not isinstance(inputs, tf.keras.KerasTensor):
if self.granularity == "per_tensor":
self.act_min_value = tf.math.reduce_min(inputs)
Expand Down Expand Up @@ -165,9 +170,11 @@ def call(self, inputs):

@classmethod
def from_config(cls, config):
"""Deserialize this class from a config dict."""
return cls(**config)

def get_config(self):
"""Serialize this class to a config dict."""
config = super(QDepthwiseConv2D, self).get_config()
config.update(
{
Expand All @@ -190,6 +197,8 @@ def get_config(self):
else:

class QDepthwiseConv2D(DepthwiseConv):
"""The custom quantized DepthwiseConv2D layer."""

def __init__(
self,
kernel_size,
Expand Down Expand Up @@ -220,6 +229,7 @@ def __init__(
quant_axis=None,
**kwargs
):
"""Initialize custom quantized DepthwiseConv2D layer."""
super().__init__(
2,
kernel_size=kernel_size,
Expand Down Expand Up @@ -253,6 +263,7 @@ def __init__(
self.quant_axis = quant_axis

def call(self, inputs):
"""The __call__ function of custom quantized DepthwiseConv2D layer."""
if self.quant_status == "calib":
if self.granularity == "per_tensor":
self.act_min_value = tf.math.reduce_min(inputs)
Expand Down Expand Up @@ -315,9 +326,11 @@ def call(self, inputs):

@classmethod
def from_config(cls, config):
"""Deserialize this class from a config dict."""
return cls(**config)

def get_config(self):
"""Serialize this class to a config dict."""
config = super(QDepthwiseConv2D, self).get_config()
config.update(
{
Expand All @@ -339,6 +352,7 @@ def get_config(self):

@tf_utils.shape_type_conversion
def compute_output_shape(self, input_shape):
"""Compute_output_shape."""
if self.data_format == "channels_first":
rows = input_shape[2]
cols = input_shape[3]
Expand Down Expand Up @@ -369,6 +383,7 @@ def compute_output_shape(self, input_shape):


def initialize_int8_depthwise_conv2d(fp32_layer, q_config):
"""Initialize int8 depthwise conv2d."""
kwargs = fp32_layer.get_config()
q_name = fp32_layer.name

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Initialize layer initializer functions."""

from neural_compressor.tensorflow.keras.layers.conv2d import initialize_int8_conv2d
from neural_compressor.tensorflow.keras.layers.dense import initialize_int8_dense
Expand Down
Loading
Loading