HSBNN: A High-Scalable Bayesian Neural Networks Accelerator Based on Field Programmable Gate Arrays (FPGA)

Traditional artificial neural networks have inherent overfitting problems and tend to produce overly confident predictions due to their reliance on point estimation methods. In contrast, Bayesian theory offers a probabilistic framework that replaces point estimation with probability distributions, e...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Cognitive computation Jg. 17; H. 3; S. 100
Hauptverfasser: Liu, Yinghao, Zhang, Hao, Sun, Zhe, Duan, Feng, Ma, Yuan, Lu, Wenyi, Caiafa, Cesar F., Solé-Casals, Jordi
Format: Journal Article
Sprache:Englisch
Veröffentlicht: New York Springer US 01.06.2025
Springer Nature B.V
Schlagworte:
ISSN:1866-9956, 1866-9964
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Traditional artificial neural networks have inherent overfitting problems and tend to produce overly confident predictions due to their reliance on point estimation methods. In contrast, Bayesian theory offers a probabilistic framework that replaces point estimation with probability distributions, effectively addressing issues of overconfidence. The brain is also believed working under the Bayesian rules, the neural networks of which evaluate the precision of prior knowledge and incoming evidence, achieving the balance of weight updating to the most reliable information sources [ 1 ]. By integrating Bayesian principles with artificial neural networks, the bio-inspired Bayesian Neural Networks (BNNs) can generate predictions accompanied by confidence evaluations, enhancing their practical applicability. To further improve the computational efficiency of BNNs and enable scalable deployment on edge devices, we propose a High-Scalable Bayesian Neural Network (HSBNN) accelerator based on field-programmable gate arrays (FPGAs) with multiple optimizations. A resource-saving Gaussian random number generator (RS-GRNG) optimized for FPGAs shows high efficiency, which seamlessly extends to support parallel sampling of weight distributions, enabling reliable confidence probability evaluations. Furthermore, the parameterization of BNN architectures with configuration files and employment of a layer-by-layer computing mode ensure that different BNNs can be accelerated without reprogramming the FPGA, offering excellent scalability. The entire system, implemented with the OpenCL heterogeneous computing library, leverages parallel processing units and pipeline channels to achieve high acceleration performance and efficient data transfer. The experiment results demonstrate that the system achieves a data processing throughput of 1.002 milliseconds per image, exceeding CPU performance by 1000-fold and GPU performance by nearly 500-fold.
Bibliographie:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:1866-9956
1866-9964
DOI:10.1007/s12559-025-10455-9