%P 629-632 %T Dynamic point stochastic rounding algorithm for limited precision arithmetic in Deep Belief Network training %I IEEE Computer Society %A M. Essam %A T.B. Tang %A E.T.W. Ho %A H. Chen %O cited By 7; Conference of 8th International IEEE EMBS Conference on Neural Engineering, NER 2017 ; Conference Date: 25 May 2017 Through 28 May 2017; Conference Code:129986 %J International IEEE/EMBS Conference on Neural Engineering, NER %L scholars8456 %D 2017 %R 10.1109/NER.2017.8008430 %K Digital arithmetic, Deep belief network (DBN); Deep belief networks; Fixed points; Floating points; Mnist database; Precision arithmetic; Rounding algorithm; Scaling factors, Stochastic systems %X This paper reports how to train a Deep Belief Network (DBN) using only 8-bit fixed-point parameters. We propose a dynamic-point stochastic rounding algorithm that provides enhanced results compared to the existing stochastic rounding. We show that by using a variable scaling factor, the fixed-point parameter updates are enhanced. To be more hardware amenable, the use of common scaling factor at each layer of DBN is further proposed. Using publicly available MNIST database, we show that the proposed algorithm can train a 3-layer DBN with an average accuracy of 98.49, with a drop of 0.08 from the double floating-point average accuracy. © 2017 IEEE.