ORIGINAL RESEARCH article
Front. Comput. Sci.
Sec. Human-Media Interaction
Volume 7 - 2025 | doi: 10.3389/fcomp.2025.1575202
This article is part of the Research TopicEmbodied Perspectives on Sound and Music AIView all 12 articles
Exploring Gestural Affordances in Audio Latent Space Navigation
Provisionally accepted- 1Queen Mary University of London, London, United Kingdom
- 2University of the Arts London, London, United Kingdom
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
The latent space of generative AI models affords unique creative possibilities and broad design space for AI-enhanced digital music instruments. While interface designs for latent space navigation typically rely on sound-producing gestures that involve bodily motions and movements, the underlying subjective perception of these gestures remains underexplored. To understand how musicians perceive sound-producing gestures and tailor performance techniques in audio latent space, we present a user study workshop with an AI-enhanced digital music instrument with a tablet interface. Eighteen musicians were recruited to test out open-ended gestures and tasked to create musical scores. We report how they use sound-producing gestures in the latent space and develop performance techniques. We contribute findings from an embodied music cognition perspective of how subjective perception of gestures shapes musicians' technique development in audio latent space navigation. We discuss the implications of new gestural affordances discovered by participants in our workshop, aiming to elucidate new opportunities for digital musical instruments with audio latent space navigation.
Keywords: AI Musical Instruments, New interfaces for musical expression, neural audio synthesis, Sound and music computing, embodied music cognition, Digital musical instruments
Received: 12 Feb 2025; Accepted: 15 Sep 2025.
Copyright: © 2025 Zheng, Xambó Sedó and Bryan-Kinns. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Shuoyang Jasper Zheng, shuoyang.zheng@qmul.ac.uk
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.