ORIGINAL RESEARCH article

Front. Artif. Intell.

Sec. Machine Learning and Artificial Intelligence

Volume 8 - 2025 | doi: 10.3389/frai.2025.1618149

This article is part of the Research TopicNeural-Symbolic NLP: Bridging Theory and PracticeView all articles

Deriving Equivalent Symbol-based Decision Models from Feedforward Neural Networks

Provisionally accepted
Sebastian  SeidelSebastian Seidel1Uwe  M. BorghoffUwe M. Borghoff2*
  • 1KNDS Deutschland GmbH & Co. KG, Munich, Germany
  • 2University of the Bundeswehr Munich, Neubiberg, Germany

The final, formatted version of the article will be published soon.

Artificial intelligence (AI) has emerged as a transformative force across industries, driven by advances in deep learning and natural language processing, and fueled by large-scale data and computing resources. Despite its rapid adoption, the opacity of AI systems poses significant challenges to trust and acceptance. This work explores the intersection of connectionist and symbolic approaches to artificial intelligence, focusing on the derivation of interpretable symbolic models, such as decision trees, from feedforward neural networks (FNNs). Decision trees provide a transparent framework for elucidating the operations of neural networks while preserving their functionality. The derivation is presented in a step-by-step approach and illustrated with several examples. A systematic methodology is proposed to bridge neural and symbolic paradigms by exploiting distributed representations in FNNs to identify symbolic components, including fillers, roles, and their interrelationships. The process traces neuron activation values and input configurations across network layers, mapping activations and their underlying inputs to decision tree edges. The resulting symbolic structures effectively capture FNN decision processes and enable scalability to deeper networks through iterative refinement of subpaths for each hidden layer.To validate the theoretical framework, a prototype was developed using KERAS .h5-data and emulating TENSORFLOW within the JAVA JDK/JAVAFX environment. This prototype demonstrates the feasibility of extracting symbolic representations from neural networks, enhancing trust in AI systems, and promoting accountability.

Keywords: artificial neural networks, connectionism, Decision Trees, Explainable AI, Symbolic AI models, symbolism

Received: 25 Apr 2025; Accepted: 07 Jul 2025.

Copyright: © 2025 Seidel and Borghoff. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Uwe M. Borghoff, University of the Bundeswehr Munich, Neubiberg, Germany

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.