You're viewing our updated article page. If you need more time to adjust, you can return to the old layout.

EDITORIAL article

Front. Neurosci., 15 July 2024

Sec. Decision Neuroscience

Volume 18 - 2024 | https://doi.org/10.3389/fnins.2024.1448517

Editorial: Information theory meets deep neural networks: theory and applications

  • 1. Institute of Microelectronics, University of Macau, Taipa, China

  • 2. School of Creative and Digital Industries, Buckinghamshire New University, Bradford, United Kingdom

  • 3. School of Automation, Chongqing University, Chongqing, China

Article metrics

View details

1,6k

Views

659

Downloads

We are delighted to introduce this Research Topic, titled “Information Theory Meets Deep Neural Networks: Theory and Applications”. Deep neural networks (DNNs) have become a focal point in machine learning research, achieving impressive results across various tasks. However, understanding their workings and mechanisms remains challenging (Samek et al., 2021; Gawlikowski et al., 2023). Information theory, a mathematical framework for representing and analyzing information, has been widely applied to study the fundamental characteristics of data, such as structure and distribution. In the context of DNNs, information theory has been instrumental in explaining and optimizing their performance (Zhang and Li, 2019; Zhang et al., 2022, 2023). For instance, the information bottleneck theory has shed light on the abstract representations of neural networks, while entropy and mutual information have been used to evaluate model complexity and generalization performance (Wu et al., 2023). This Research Topic aims to explore the intersection of information theory and DNNs, two fields that have profoundly impacted the understanding and advancement of neural networks and their applications. The synergy between these disciplines offers promising avenues for developing more efficient, robust, and interpretable AI systems. In this Research Topic, we present four papers that illustrate the breadth and depth of research at this intersection, highlighting innovative methodologies and their applications in various domains.

You and Wang proposed a novel approach to genealogy layout recognition. Recognizing the significance of genealogies in cultural heritage, the authors introduced a sublinear information bottleneck (SIB) for feature extraction and a two-stage deep learning model combining SIB-ResNet and SIB-YOLOv5. This method surpassed existing state-of-the-art techniques, offering promising results in identifying and localizing components in genealogy images. This advancement not only aids in genealogy research but also in preserving cultural heritage through improved recognition technologies.

Li and Peng addressed the challenges of synthetic aperture radar (SAR) automatic target recognition (ATR). The study introduced a data augmentation technique that mitigates SAR image noise and a weighted ResNet with residual strain control. This approach not only enhances computational efficiency but also improves recognition accuracy, significantly reducing training time and data requirements. The experimental results demonstrated the superior performance of this method, paving the way for more efficient SAR ATR systems.

Alazeb et al. focused on shifting to the realm of robotic environments and scene classification. The paper presented a robust framework for multi-object detection and scene understanding, leveraging advanced visual sensor technologies and deep learning models. By integrating preprocessing, semantic segmentation, feature extraction, and object recognition, the proposed system achieved remarkable accuracy on standard datasets such as PASCALVOC-12, Cityscapes, and Caltech 101. This work represented a significant step forward in enhancing the capabilities of vision-based systems in various applications, from autonomous driving to augmented reality.

Finally, Chen et al. delved into the theoretical aspects of neural network training. The authors propose a novel method for adaptive learning rate estimation in restricted Boltzmann machines (RBMs) using rectified linear units (ReLUs). By providing mathematical expressions for adaptive learning step calculation, this approach optimized the learning rate dynamically, improving the generalization ability and reducing the loss function more effectively than traditional methods. This theoretical contribution offers valuable insights into the optimization of unsupervised learning algorithms.

In conclusion, this Research Topic showcased the innovative research at the crossroads of information theory and deep neural networks. The contributions presented here not only advance theoretical understanding but also demonstrate practical applications that hold the potential to transform various fields. We extended our gratitude to the authors for their exceptional work and to the reviewers for their rigorous evaluation. We hope this Research Topic inspires further research and collaboration in this exciting domain.

Statements

Author contributions

AZ: Resources, Writing – original draft. QZ: Writing – review & editing. KZ: Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the National Natural Science Foundation of China under Grant 62306001.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1

    Gawlikowski J. Tassi C. R. N. Ali M. Lee J. Humt M. Feng J. et al . (2023). A survey of uncertainty in deep neural networks. Artif. Intellig. Rev. 56, 15131589. 10.1007/s10462-023-10562-9

  • 2

    Samek W. Montavon G. Lapuschkin S. Anders C. J. Müller K.-R. (2021). Explaining deep neural networks and beyond: A review of methods and applications. Proc, IEEE109, 247278. 10.1109/JPROC.2021.3060483

  • 3

    Wu J. Huang Y. Gao M. Gao Z. Zhao J. Shi J. et al . (2023). Exponential information bottleneck theory against intra-attribute variations for pedestrian attribute recognition. IEEE Trans. Inform. Forens. Secur. 18, 56235635. 10.1109/TIFS.2023.3311584

  • 4

    Zhang A. Li X. Gao Y. Niu Y. (2022). Event-driven intrinsic plasticity for spiking convolutional neural networks. IEEE Trans. Neural Netw. Learn. Syst. 33, 19861995. 10.1109/TNNLS.2021.3084955

  • 5

    Zhang A. Shi J. Wu J. Zhou Y. Yu W. (2023). Low latency and sparse computing spiking neural networks with self-driven adaptive threshold plasticity. IEEE Trans. Neural Netw. Learn. Syst. 112. 10.1109/TNNLS.2023.3300514

  • 6

    Zhang W. Li P. (2019). Information-theoretic intrinsic plasticity for online unsupervised learning in spiking neural networks. Front. Neurosci. 13:31. 10.3389/fnins.2019.00031

Summary

Keywords

artificial neural networks, information theory, information bottleneck, deep learning—artificial intelligence, deep neural networks (DNNs)

Citation

Zhang A, Zhang Q and Zhao K (2024) Editorial: Information theory meets deep neural networks: theory and applications. Front. Neurosci. 18:1448517. doi: 10.3389/fnins.2024.1448517

Received

13 June 2024

Accepted

27 June 2024

Published

15 July 2024

Volume

18 - 2024

Edited and reviewed by

Shuangming Yang, Tianjin University, China

Updates

Copyright

*Correspondence: Anguo Zhang

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics