Analytical Guarantees on Numerical Precision of Deep Neural Networks

Abstract

The acclaimed successes of neural networks often overshadow their tremendous complexity. We focus on numerical precision – a key parameter defining the complexity of neural networks. First, we present theoretical bounds on the accuracy in presence of limited precision. Interestingly, these bounds can be computed via the back-propagation algorithm. Hence, by combining our theoretical analysis and the back-propagation algorithm, we are able to readily determine the minimum precision needed to preserve accuracy without having to resort to time-consuming fixed-point simulations. We provide numerical evidence showing how our approach allows us to maintain high accuracy but with lower complexity than state-of-the-art binary networks.

Cite

Text

Sakr et al. "Analytical Guarantees on Numerical Precision of Deep Neural Networks." International Conference on Machine Learning, 2017.

Markdown

[Sakr et al. "Analytical Guarantees on Numerical Precision of Deep Neural Networks." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/sakr2017icml-analytical/)

BibTeX

@inproceedings{sakr2017icml-analytical,
  title     = {{Analytical Guarantees on Numerical Precision of Deep Neural Networks}},
  author    = {Sakr, Charbel and Kim, Yongjune and Shanbhag, Naresh},
  booktitle = {International Conference on Machine Learning},
  year      = {2017},
  pages     = {3007-3016},
  volume    = {70},
  url       = {https://mlanthology.org/icml/2017/sakr2017icml-analytical/}
}