HYPOTHESIS AND THEORY article
Front. Comput. Sci.
Sec. Human-Media Interaction
Volume 7 - 2025 | doi: 10.3389/fcomp.2025.1575168
This article is part of the Research TopicEmbodied Perspectives on Sound and Music AIView all 8 articles
Neural Audio Instruments: Epistemological and Phenomenological Perspectives on Musical Embodiment of Deep Learning
Provisionally accepted- 1Department of Music, College of Arts, Media and Design, Northeastern University, Boston, Massachusetts, United States
- 2Division of Data Science and AI, Department of Computer Science and Engineering, Chalmers University of Technology, Gothenburg, Sweden
- 3Department of Computer Science and Engineering, Faculty of Information Technology, University of Gothenburg, Gothenburg, Sweden
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Neural Audio is a category of deep learning pipelines which output audio signals directly, in realtime scenarios of action-sound interactions. In this work, we examine how neural audio-based artificial intelligence, when embedded in digital musical instruments (DMIs), shapes embodied musical interaction. While DMIs have long struggled to match the physical immediacy of acoustic instruments, neural audio methods can magnify this challenge, requiring data collection, model training and deep theoretical knowledge that appear to push musicians toward symbolic or conceptual modes of engagement. Paradoxically, these same methods can also foster more embodied practices, by introducing opaque yet expressive behaviors that free performers from rigid technical models and encourage discovery through tactile, real-time experimentation.Drawing on established perspectives in DMI embodiment literature, as well as emerging neuralaudio-focused efforts within the community, we highlight two seemingly conflicting aspects of these instruments: on one side, they inherit many "disembodying" traits known from DMIs; on the other, they open pathways reminiscent of acoustic phenomenology and soma, potentially restoring the close physical interplay often missed in digital performance.
Keywords: Neural Audio, Digital musical instruments, Neural Audio Instruments, embodied interaction, music performance, artificial intelligence, deep learning, Latent space
Received: 11 Feb 2025; Accepted: 23 Jul 2025.
Copyright: © 2025 Zappi and Tatar. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Victor Zappi, Department of Music, College of Arts, Media and Design, Northeastern University, Boston, 02115, Massachusetts, United States
Kıvanç Tatar, Division of Data Science and AI, Department of Computer Science and Engineering, Chalmers University of Technology, Gothenburg, Sweden
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.