Flexpoint: Predictive Numerics for Deep Learning
Deep learning has been undergoing rapid growth in recent years thanks to its state-of-the-art performance across a wide range of real-world applications. Traditionally neural networks were trained in IEEE-754 binary64 or binary32 format, a common practice in general scientific computing. However, th...
Saved in:
| Published in | Proceedings - Symposium on Computer Arithmetic pp. 1 - 4 |
|---|---|
| Main Authors | , , , , |
| Format | Conference Proceeding |
| Language | English |
| Published |
IEEE
01.06.2018
|
| Subjects | |
| Online Access | Get full text |
| ISSN | 2576-2265 |
| DOI | 10.1109/ARITH.2018.8464801 |
Cover
| Summary: | Deep learning has been undergoing rapid growth in recent years thanks to its state-of-the-art performance across a wide range of real-world applications. Traditionally neural networks were trained in IEEE-754 binary64 or binary32 format, a common practice in general scientific computing. However, the unique computational requirements of deep neural network training workloads allow for much more efficient and inexpensive alternatives, unleashing a new wave of numerical innovations powering specialized computing hardware. We previously presented Flexpoint, a blocked fixed-point data type combined with a novel predictive exponent management algorithm designed to support training of deep networks without modifications, aiming at a seamless replacement of the binary32 widely in practice today. We showed that Flexpoint with 16-bit mantissa and 5-bit shared exponent (flex16+S) achieved numerical parity to binary32 in training a number of convolutional neural networks. In the current paper we review the continuing trend of predictive numerics enhancing deep neural network training in specialized computing devices such as the Intel ® N ervana ™ Neural Network Processor. |
|---|---|
| ISSN: | 2576-2265 |
| DOI: | 10.1109/ARITH.2018.8464801 |