Unknown

Dataset Information

0

Hardware implementation of backpropagation using progressive gradient descent for in situ training of multilayer neural networks.


ABSTRACT: Neural network training can be slow and energy-expensive due to the frequent transfer of weight data between digital memory and processing units. Neuromorphic systems can accelerate neural networks by performing multiply-accumulate operations in parallel using nonvolatile analog memory. However, executing the widely used backpropagation training algorithm in multilayer neural networks requires information-and therefore storage-of the partial derivatives of the weight values preventing suitable and scalable implementation in hardware. Here, we propose a hardware implementation of the backpropagation algorithm that progressively updates each layer using in situ stochastic gradient descent, avoiding this storage requirement. We experimentally demonstrate the in situ error calculation and the proposed progressive backpropagation method in a multilayer hardware-implemented neural network. We confirm identical learning characteristics and classification performance compared to conventional backpropagation in software. We show that our approach can be scaled to large and deep neural networks, enabling highly efficient training of advanced artificial intelligence computing systems.

SUBMITTER: van Doremaele ERW 

PROVIDER: S-EPMC11244533 | biostudies-literature | 2024 Jul

REPOSITORIES: biostudies-literature

altmetric image

Publications

Hardware implementation of backpropagation using progressive gradient descent for in situ training of multilayer neural networks.

van Doremaele Eveline R W ERW   Stevens Tim T   Stevens Tim T   Ringeling Stijn S   Spolaor Simone S   Fattori Marco M   van de Burgt Yoeri Y  

Science advances 20240712 28


Neural network training can be slow and energy-expensive due to the frequent transfer of weight data between digital memory and processing units. Neuromorphic systems can accelerate neural networks by performing multiply-accumulate operations in parallel using nonvolatile analog memory. However, executing the widely used backpropagation training algorithm in multilayer neural networks requires information-and therefore storage-of the partial derivatives of the weight values preventing suitable a  ...[more]

Similar Datasets

| S-EPMC11628607 | biostudies-literature
| S-EPMC11623183 | biostudies-literature
| S-EPMC7610561 | biostudies-literature
| S-EPMC7039878 | biostudies-literature
| S-EPMC11549378 | biostudies-literature
| S-EPMC10912231 | biostudies-literature
| S-EPMC4382991 | biostudies-literature
| S-EPMC7358558 | biostudies-literature
| S-EPMC9792515 | biostudies-literature
| S-EPMC8627686 | biostudies-literature