arXiv Analytics

Sign in

arXiv:2102.09642 [cs.LG]AbstractReferencesReviewsResources

Control Variate Approximation for DNN Accelerators

Georgios Zervakis, Ourania Spantidi, Iraklis Anagnostopoulos, Hussam Amrouch, Jörg Henkel

Published 2021-02-18Version 1

In this work, we introduce a control variate approximation technique for low error approximate Deep Neural Network (DNN) accelerators. The control variate technique is used in Monte Carlo methods to achieve variance reduction. Our approach significantly decreases the induced error due to approximate multiplications in DNN inference, without requiring time-exhaustive retraining compared to state-of-the-art. Leveraging our control variate method, we use highly approximated multipliers to generate power-optimized DNN accelerators. Our experimental evaluation on six DNNs, for Cifar-10 and Cifar-100 datasets, demonstrates that, compared to the accurate design, our control variate approximation achieves same performance and 24% power reduction for a merely 0.16% accuracy loss.

Comments: Accepted for publication at the 58th Design Automation Conference (DAC'21), December 5-9, 2021, San Francisco, USA
Categories: cs.LG, cs.AR
Related articles: Most relevant | Search more
arXiv:2306.04645 [cs.LG] (Published 2023-05-31)
Special Session: Approximation and Fault Resiliency of DNN Accelerators
arXiv:2010.14778 [cs.LG] (Published 2020-10-28)
DNA: Differentiable Network-Accelerator Co-Search
Yongan Zhang et al.
arXiv:2201.11218 [cs.LG] (Published 2022-01-26)
DNNFuser: Generative Pre-Trained Transformer as a Generalized Mapper for Layer Fusion in DNN Accelerators