Supported Layers for Quantization - MATLAB & Simulink (original) (raw)

Main Content

The deep neural network layers supported for the quantization workflow depend on the choice of execution environment for the quantization workflow. The target hardware associated with each execution environment determines the supported networks and layers for that environment.

For more information about the quantization workflow and choosing an execution environment, see Quantization of Deep Neural Networks.

Two types of support for layers are:

To determine which layers in your network can be quantized, analyze your network for compression in the Deep Network Designer app.

Networks and Layers Supported in the Quantization Workflow

These layers can exist in your network for the quantization workflow. For additional details about support for the layers, see Considerations for Supported Layers for Quantization.

Layers That Can Be Quantized

This table lists the layers that can be quantized. For details about additional considerations for support for particular layers and architectures, see Considerations for Supported Layers for Quantization.

For a list of layers that can be quantized for the CPU execution environment, seeGenerate int8 Code for Deep Learning Networks (MATLAB Coder).

Convolutional and Fully Connected Layers

Input Layer Normalization

Normalization, Dropout, and Cropping Layers

See Also

dlquantizer | Deep Network Quantizer

Topics