TY - GEN
T1 - Hardware acceleration of Bayesian neural networks using RAM based linear feedback Gaussian random number generators
AU - Cai, Ruizhe
AU - Ren, Ao
AU - Wangy, Luhao
AU - Pedramy, Massoud
AU - Wang, Yanzhi
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017/11/22
Y1 - 2017/11/22
N2 - Bayesian neural networks (BNNs) have been proposed to address the problem of model uncertainty in training. By introducing weights associated with conditioned probability distributions, BNN is capable to resolve overfitting issues commonly seen in conventional neural networks. Frequent usage of Gaussian random variables requires a properly optimized Gaussian Random Number Generator (GRNG). The high hardware cost of conventional GRNG makes the hardware realization of BNN challenging. In this paper, a new hardware acceleration architecture for variational inference in BNNs is proposed to facilitate the applicability of BNN in larger-scale applications. In addition, the proposed implementation introduced the RAM based Linear Feedback based GRNG (RLF-GRNG) for effective weight sampling in BNNs. The RAM based Linear Feedback method can effectively utilize RAM resources for parallel Gaussian random number generation while requiring limited and sharable control logic. Implementation on an Altera Cyclone V FPGA suggests that the RLF-GRNG utilizes much less RAM resources compared to other GRNG methods. Experiments results show that the proposed hardware implementation of a BNN can still attain similar accuracy compared to software implementation.
AB - Bayesian neural networks (BNNs) have been proposed to address the problem of model uncertainty in training. By introducing weights associated with conditioned probability distributions, BNN is capable to resolve overfitting issues commonly seen in conventional neural networks. Frequent usage of Gaussian random variables requires a properly optimized Gaussian Random Number Generator (GRNG). The high hardware cost of conventional GRNG makes the hardware realization of BNN challenging. In this paper, a new hardware acceleration architecture for variational inference in BNNs is proposed to facilitate the applicability of BNN in larger-scale applications. In addition, the proposed implementation introduced the RAM based Linear Feedback based GRNG (RLF-GRNG) for effective weight sampling in BNNs. The RAM based Linear Feedback method can effectively utilize RAM resources for parallel Gaussian random number generation while requiring limited and sharable control logic. Implementation on an Altera Cyclone V FPGA suggests that the RLF-GRNG utilizes much less RAM resources compared to other GRNG methods. Experiments results show that the proposed hardware implementation of a BNN can still attain similar accuracy compared to software implementation.
UR - http://www.scopus.com/inward/record.url?scp=85041656548&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85041656548&partnerID=8YFLogxK
U2 - 10.1109/ICCD.2017.51
DO - 10.1109/ICCD.2017.51
M3 - Conference contribution
AN - SCOPUS:85041656548
T3 - Proceedings - 35th IEEE International Conference on Computer Design, ICCD 2017
SP - 289
EP - 296
BT - Proceedings - 35th IEEE International Conference on Computer Design, ICCD 2017
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 35th IEEE International Conference on Computer Design, ICCD 2017
Y2 - 5 November 2017 through 8 November 2017
ER -