ORIGINAL RESEARCH article
Front. Syst. Neurosci.
This article is part of the Research TopicAdvancing brain-computer interfaces through fNIRS and EEG integrationView all articles
Real-World Evaluation of Deep Learning Decoders for Motor Imagery EEG-based BCIs
Provisionally accepted- 1University of Salerno, Fisciano, Italy
- 2King's College London, London, United Kingdom
- 3University of Kindu, Kindu, Democratic Republic of Congo
- 4Universita del Salento, Lecce, Italy
- 5ISPT-Kin, Kinshasa, Democratic Republic of Congo
- 6Universitat Klagenfurt, Klagenfurt am Wörthersee, Austria
- 7Universite de Kinshasa, Kinshasa, Democratic Republic of Congo
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Motor Imagery (MI) Electroencephalography (EEG)-based control in online Brain-Computer Interfaces demands issuing decisions on short temporal context. However, most published Deep Learning (DL) EEG decoders are introduced and validated offline on public datasets with longer windows, leaving their real-time suitability uncertain. We investigate this gap by testing 10 representative DL decoders—covering convolutional neural networks (CNNs), filter-bank CNNs, temporal-convolutional networks (TCNs), attention and Transformer hybrids—under a soft real-time protocol with 2-second windows. We quantify performance with accuracy, sensitivity, precision, miss-as-neutral (MANR), false-alarm rate (FAR), information-transfer rate (ITR), and workload. To relate behavior to physiology, we examine lateralization indices, C3 vs C4 mu-band power, and MI-versus-Neutral topographic contrasts. Our study reveals ranking shifts between offline and online BCI operation and a marked rise in inter-subject variability. Compact spectro-temporal CNN backbones with lightweight temporal context (TCN or dilations) sustain performance more consistently under short-time windows, whereas deeper attention and Transformer stacks are more sensitive to subject and session differences. This work provides a reproducible benchmark and actionable guidance for online-first EEG decoder design and calibration that remains reliable under real-world, short-time constraints.
Keywords: Electroencephalography, Brain-Computer Interfaces, Motor Imagery, deep learning, digital signal processing
Received: 03 Oct 2025; Accepted: 21 Nov 2025.
Copyright: © 2025 Sedi Nzakuna, D'Auria, Paciello, Gallo, Kamavuako, Lay-Ekuakille and KYAMAKYA. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Vincenzo Paciello, vpaciello@unisa.it
Kyandoghere KYAMAKYA, kyandoghere.kyamakya@aau.at
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
