PACT: Parameterized Clipping Activation for Quantized Neural Networks (original) (raw)
Related papers
Accurate and Efficient 2-bit Quantized Neural Networks
2019
Low-bit Quantization of Neural Networks for Efficient Inference
2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), 2019
Ultra-Low Precision 4-bit Training of Deep Neural Networks
2020
Bit Efficient Quantization for Deep Neural Networks
2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing - NeurIPS Edition (EMC2-NIPS), 2019
A White Paper on Neural Network Quantization
2021
ACIQ: Analytical Clipping for Integer Quantization of neural networks
ArXiv, 2018
Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights
ArXiv, 2017
Dataflow-based Joint Quantization of Weights and Activations for Deep Neural Networks
ArXiv, 2019
A Novel Low-Bit Quantization Strategy for Compressing Deep Neural Networks
Computational Intelligence and Neuroscience, 2020
BMPQ: Bit-Gradient Sensitivity-Driven Mixed-Precision Quantization of DNNs from Scratch
2022 Design, Automation & Test in Europe Conference & Exhibition (DATE)
Adaptive Precision Training (AdaPT): A dynamic quantized training approach for DNNs
Society for Industrial and Applied Mathematics eBooks, 2023
Post training 4-bit quantization of convolutional networks for rapid-deployment
2019
Improved Techniques for Quantizing Deep Networks with Adaptive Bit-Widths
2021
A Comprehensive Survey on Model Quantization for Deep Neural Networks
arXiv (Cornell University), 2022
A Fixed-Point Quantization Technique for Convolutional Neural Networks Based on Weight Scaling
2019 IEEE International Conference on Image Processing (ICIP)
Scalable Methods for 8-bit Training of Neural Networks
2018
AdaQAT: Adaptive Bit-Width Quantization-Aware Training
2024
SYQ: Learning Symmetric Quantization for Efficient Deep Neural Networks
2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018
Adaptive Precision Training (AdaPT): A dynamic fixed point quantized training approach for DNNs
2021
Learned Step Size Quantization
ArXiv, 2020
BitPruning: Learning Bitlengths for Aggressive and Accurate Quantization
Cornell University - arXiv, 2020
Blended coarse gradient descent for full quantization of deep neural networks
Research in the Mathematical Sciences, 2019
ReLeQ : A Reinforcement Learning Approach for Automatic Deep Quantization of Neural Networks
IEEE Micro
QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
2022
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework
2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA), 2021
IEEE Access, 2022
Power-of-Two Quantization for Low Bitwidth and Hardware Compliant Neural Networks
ArXiv, 2022
Differentiable Quantization of Deep Neural Networks
ArXiv, 2019
Post-training Piecewise Linear Quantization for Deep Neural Networks
Computer Vision – ECCV 2020, 2020
DoubleQExt: Hardware and Memory Efficient CNN Through Two Levels of Quantization
IEEE Access, 2021
2021
Quantizing Convolutional Neural Networks for Low-Power High-Throughput Inference Engines
ArXiv, 2018
BinaryRelax: A Relaxation Approach for Training Deep Neural Networks with Quantized Weights
SIAM Journal on Imaging Sciences, 2018
Dither NN: Hardware/Algorithm Co-Design for Accurate Quantized Neural Networks
IEICE Transactions on Information and Systems, 2019
HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision
In Proceedings of the IEEE International Conference on Computer Vision, 2019