Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell.

Sec. Machine Learning and Artificial Intelligence

Volume 8 - 2025 | doi: 10.3389/frai.2025.1605706

Effective Methods and Framework for Energy-Based Local Learning of Deep Neural Networks

Provisionally accepted
Haibo  ChenHaibo Chen1Bangcheng  YangBangcheng Yang1Fucun  HeFucun He1Fei  ZhouFei Zhou1,2Shuai  ChenShuai Chen1,2Chunpeng  WuChunpeng Wu1,2Fan  LiFan Li1,2Yam Song (Yansong)  ChuaYam Song (Yansong) Chua1*
  • 1China Nanhu Academy of Electronics and Information Technology, Jiaxing, China
  • 2State Grid Shanghai Municipal Electric Power Company, Shanghai, China

The final, formatted version of the article will be published soon.

From a neuroscience perspective, artificial neural networks are regarded as abstract models of biological neurons, yet they rely on biologically implausible backpropagation for training. Energy-based models represent a class of brain-inspired learning frameworks that adjust system states by minimizing an energy function.Predictive coding (PC), a theoretical model within energy-based models, constructs its energy function from forward prediction errors, with optimization achieved by minimizing local layered errors. Owing to its local plasticity, PC emerges as the most promising alternative to backpropagation. However, PC face gradient explosion and vanishing challenges in deep networks with multiple layers. Gradient explosion occurs when layer-wise prediction errors are excessively large, while gradient vanishing arises when they are excessively small. To address these challenges, we propose bidirectional energy to stabilize prediction errors and mitigate gradient explosion, while using skip connections to resolve gradient vanishing problems. We also introduce a layer-adaptive learning rate (LALR) to enhance training efficiency. Our model achieves accuracies of 99.22% on MNIST, and 73.35% on Tiny ImageNet, comparable to the performance of identically structed networks trained with backprop. Finally, we developed a Jax-based framework for efficient training of energy-based models, reducing training time by half compared to PyTorch.

Keywords: artificial neural network, biologically plausible learning rule, Local learning, Energy-based model, predictive coding

Received: 03 Apr 2025; Accepted: 23 Jul 2025.

Copyright: © 2025 Chen, Yang, He, Zhou, Chen, Wu, Li and Chua. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Yam Song (Yansong) Chua, China Nanhu Academy of Electronics and Information Technology, Jiaxing, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.