Author of the publication

ZeroQuant-HERO: Hardware-Enhanced Robust Optimized Post-Training Quantization Framework for W8A8 Transformers.

, , , , , and . CoRR, (2023)

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed. You can also use the button next to the name to display some publications already assigned to the person.

 

Other publications of authors with the same name

Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT., , , , , , , and . AAAI, page 8815-8821. AAAI Press, (2020)ADAHESSIAN: An Adaptive Second Order Optimizer for Machine Learning., , , , , and . AAAI, page 10665-10673. AAAI Press, (2021)ZeroQuant-HERO: Hardware-Enhanced Robust Optimized Post-Training Quantization Framework for W8A8 Transformers., , , , , and . CoRR, (2023)Selective Guidance: Are All the Denoising Steps of Guided Diffusion Important?, , and . CoRR, (2023)Residual Networks as Nonlinear Systems: Stability Analysis using Linearization., , , and . CoRR, (2019)PyHessian: Neural Networks Through the Lens of the Hessian., , , and . CoRR, (2019)HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks., , , , , and . NeurIPS, (2020)HAWQ-V3: Dyadic Neural Network Quantization., , , , , , , , , and 1 other author(s). ICML, volume 139 of Proceedings of Machine Learning Research, page 11875-11886. PMLR, (2021)PowerNorm: Rethinking Batch Normalization in Transformers., , , , and . ICML, volume 119 of Proceedings of Machine Learning Research, page 8741-8751. PMLR, (2020)ActNN: Reducing Training Memory Footprint via 2-Bit Activation Compressed Training., , , , , , and . ICML, volume 139 of Proceedings of Machine Learning Research, page 1803-1813. PMLR, (2021)