Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med., 11 November 2025

Sec. Translational Medicine

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1693921

A Dual-input deep learning architecture for classification and latency estimation in ABR signals


Youssef DarahemYoussef Darahem1Oguz YilmazOguz Yilmaz2Halil B. SaldirimHalil B. Saldirim3Berna MutluBerna Mutlu3Hasan F. AtesHasan F. Ates4Bahadir K. Gunturk
Bahadir K. Gunturk1*
  • 1Department of Computer Engineering, Istanbul Medipol University, Istanbul, Türkiye
  • 2Department of Audiology, Bakircay University, Izmir, Türkiye
  • 3Department of Health Sciences, Istanbul Medipol University, Istanbul, Türkiye
  • 4Department of Artificial Intelligence and Data Engineering, Ozyegin University, Istanbul, Türkiye

Introduction: Auditory brainstem response (ABR) is an objective neurophysiological evaluation designed to measure the electrical activity originating from the auditory nerve and brainstem in response to auditory stimulation. This assessment objectively records synchronous neural activity as it propagates along the auditory pathway. It is characterized by several distinct waves, most notably waves I, III, and V. Wave V plays a central clinical role since its presence and latency are routinely used to assess a patient's hearing status. However, manual identification and localization of wave V are time consuming and subjective. Previous work has explored automated detection methods to reduce this burden.

Methods: In this paper, we make two primary contributions. First, we propose a multi-task deep learning pipeline that simultaneously (i) detects the presence of wave V and (ii) predicts its latency, thus eliminating the need for separate manual interpretation steps and enhancing clinical usability. Second, inspired by the audiologist's practice of comparing responses at multiple click sound intensities—specifically, using responses at high intensities, where waves are more prominent, as reference—we introduce a paired-signal approach. Each input to our deep learning model consists of the test signal together with its corresponding 80 dB reference from the same recording session. This provides the model with richer contextual information, and we show that the paired-signal approach improves over the single input approach. For multi-task learning, we design a network that consists of a backbone and two branches, one for latency prediction and the other for classification of whether wave V exists or not. We first train a latency-prediction network and then freeze its feature extraction layers to initialize a classification branch. Finally, we fine-tune the entire network using a joint loss function that balances classification and regression objectives.

Results: Experimental results demonstrate that our joint model1 outperforms conventional single-task approaches. For classification, it achieves an F1-score of 0.92; for latency regression, it attains an R2 of 0.90.

Discussion: Our findings highlight the promise of convolutional neural networks for enhancing ABR analysis and underscore their potential to streamline clinical workflows in the diagnosis of auditory disorders.

1 Introduction

Auditory Brainstem Response (ABR) test is a widely used electrophysiological test method for detecting hearing loss, especially in infancy and individuals who cannot cooperate with behavioral tests (1). ABRs are recorded non-invasively via scalp electrodes and reflect the bioelectric activity of specific components of the auditory pathway following an auditory stimulus (2). Its clinical value lies in the objective assessment of hearing thresholds and functional integrity of the auditory system (3). An ABR signal has a typical structure with 7 positive waves. Three known waves (I, III, and V) are more prominent and therefore clinically the most useful (4). These waveforms are typically labeled through visual inspection. The lowest intensity level at which wave V can be observed gives the hearing threshold. An example is shown in Figure 1. The figure shows the ABR signals for six different intensity levels. The lowest intensity level where wave V is observed is 10 dB, indicating the hearing threshold of the ear.

Figure 1
Line graph showing amplitude versus time with lines for different intensities ranging from 0 to 80 dB. Peaks are marked with Roman numerals at several points, including I, III and V. The legend on the right indicates line colors for each intensity level.

Figure 1. ABR signal examples for a normal hearing ear at different sound intensity (dB) levels. The signal amplitude is measured in μVolts. Note that, to display all signals on the same plot, a signal is offset by 0.3 μV from the previous one, starting with the 0 dB signal.

Interpretation of an ABR test is performed by trained clinicians through visual inspection of ABR waveforms. Interpretation is known to vary greatly from clinician to clinician (5); and this subjectivity has encouraged the development of automated ways to interpret and analyze ABR signals. The initial work uses signal processing (6, 7) and traditional machine learning (8) techniques. In recent studies, deep learning based approaches have been explored. Some studies are designed to predict the latency (position) of the ABR waves. In (9), convolutional recurrent neural networks are used to localize waves I, III, and V using a dataset of 482 ABR waveforms recorded at 80 dB. In another study (10), an attention mechanism is used to predict the latency of wave V, achieving 95.89% accuracy and a maximum error of 0.1 ms (10).

Some other studies are designed to detect whether a wave (commonly, wave V) exists in an ABR signal. In (11), the so-called Wide&Deep and Light-MLP models, which use time-domain and frequency-domain features and incorporate demographic variables such as age, sex, and pure-tone thresholds, are used to detect the existence of wave V. The Wide&Deep model achieved a classification accuracy of 91.0%, while the Light-MLP model achieved 95.4%. Another study used a deep convolutional neural network to classify ABR waveforms into “clear response”, “inconclusive”, or “response absent”, achieving 92.9% accuracy along with high sensitivity and specificity on a public dataset (12). In (13), a convolutional neural network (CNN) based model is applied to standardized ABR waveform images to classify hearing loss, achieving 85% accuracy and contributing to earlier detection with improved diagnostic efficiency and objectivity (13). In (14), computer vision based approach is taken to detect Waves I, III, and V from ABR waveform images.

A summary and comparison of these recent deep learning approaches for automated ABR analysis is provided in Table 1. Unlike previous works focusing on wave V localization or existence classification, our approach jointly addresses both tasks using a unified deep learning framework. We propose a multi-task convolutional neural network that simultaneously classifies the presence of wave V and predicts its latency. This joint modeling helps the network learn richer representations and reduces overfitting, especially in low signal-to-noise scenarios.

Table 1
www.frontiersin.org

Table 1. Comparison of recent deep learning studies for automated ABR analysis.

Our method is further inspired by clinical practice, where audiologists often analyze ABR signals at multiple stimulus intensities, particularly higher intensities such as 80 dB, where wave V is more prominent. To emulate this reasoning process, our model receives as input a pair of ABR signals: the target signal and a reference signal from the same patient at a higher intensity level. This pairing strategy helps the model make better-informed decisions, particularly when wave V is difficult to detect at lower intensities.

To train and evaluate our models, we collected an in-house dataset comprising 16,436 auditory brainstem response (ABR) signals, of which 11,353 were recorded using the standard click stimulus. These recordings include data from both normal-hearing and hearing-impaired individuals. Each signal is paired with expert-annotated wave V latencies, providing high-quality labels for both classification (presence/absence) and regression (latency estimation) tasks. To the best of our knowledge, this is the first study to leverage multi-intensity signal pairs along with multi-task training for ABR analysis at this scale and level of precision.

In this paper, our contributions are as follows:

• We propose a novel approach that combines classification of wave V existence and localization into a single multi-task learning model.

• We introduce a signal-pairing technique, where each input signal is paired with a high-intensity (specifically, 80 dB) reference signal to provide contextual information inspired by clinical procedures.

• We employ a multi-task training approach that enables joint learning of both classification and latency prediction tasks.

• We investigate the effects of different loss functions on the training dynamics and the final model performance.

2 Methods

2.1 Data

The electrophysiological tests were performed in a sound-treated double-walled booth of Istanbul Medipol University—Mega Hospital with a bed and curtains to provide darkness to favor sleep. All data were collected with a commercial ABR software module (Interacoustics, software version 4.2.0.8) running on an Interacoustics Eclipse EP25 platform (hardware version 3.4.4). The click stimuli at alternating polarity, calibrated in dB normalized hearing level (nHL), were presented via insert phones. After cleaning the skin's surface with Nuprep gel, recording electrodes(Ag/AgCI) were placed on the forehead (vertex), the chin (ground) and the mastoids (i.e., reference electrode). Before starting to record, it was ensured that the impedance values of the electrodes were below 3–5 kOhms. Band-pass filtered from 100 Hz to 3,000 Hz using filter slopes of 12 dB/octave, and digitized with a 16-bit resolution. The sampling frequency is 15 kHz, which is about 0.067 milliseconds; the recording starts 1 ms before the click impulse and continues up to 20 ms after the impulse. An artifact rejection level of ±40 μ V was applied. The maximum intensity level was determined as 100 dBnHL, and waves I, III, and V were assessed at 80 dBnHL. Two runs, each consisting of averaged responses from 2,000 sweeps, were obtained at each presentation level, and thresholds were established using a 20-10 dB down and 5-10 dB up with steps that considered the last visible wave V as the threshold. Sample ABR recordings for an ear are shown in Figure 2.

Figure 2
Line graph showing signal values in microvolts over time in milliseconds. Two lines represent 60 dB (blue solid) and 80 dB (red dashed) sound levels. Wave V is marked around 6 ms.

Figure 2. ABR signal examples, recorded at 60 dB and 80 dB.

Ethical approval for this study was obtained from the Ethics Committee for Non-Interventional Research of Istanbul Medipol University (Protocol number: E-10840098-772.02-785) on 26/01/2023.

2.2 Data Prepossessing

Our dataset comprises recordings from 934 patients (609 male, 325 female), including 1,717 ears. The type of hearing was classified into five groups: normal hearing, sensorineural hearing loss (SHL), profound hearing loss, conductive hearing loss (CHL), and neuropathy. Table 2 summarizes the overall distribution.

Table 2
www.frontiersin.org

Table 2. Distribution of hearing types in the dataset.

In total, 16,436 ABR signals were recorded; about 31% of these waveforms lacked a discernible wave V, suggesting potential auditory pathway abnormalities. To prepare the data for the machine learning model, the first 250 samples were taken, which correspond to a time range from -1 ms to about 15.7 ms, beyond which the signal is irrelevant for wave V.

The dataset was then split into 70% training and 30% validation sets, preserving the ratio of wave V absence and presence. The split was done based on the ear; that means, it was ensured that signals from an ear are either in the training set or the testing set during the split. The final distributions with respect to the presence of wave V and with respect to the wave V positions are shown in Figures 3, 4, respectively.

Figure 3
Two pie charts compare data availability in training and validation sets. The training set (n=11,507) shows 69.8% for wave V existing and 30.2% for missing wave V. The validation set (n=4,929) shows 67.4% for wave V existing and 32.6% for missing wave V.

Figure 3. Distribution of wave V presence vs absence.

Figure 4
Density plot comparing two datasets: Training (in blue, n=8,031) and Validation (in orange, n=3,322). Both curves show a peak around 120 on the X-axis (Wave V Location Indices). Training mean is indicated at 123.4 and Validation mean at 122.2, marked by dashed lines. Y-axis represents the density.

Figure 4. Distribution of wave V locations.

2.3 Model architecture

In this study, we investigated convolutional neural networks (CNNs) (15) for simultaneous classification of wave V existence and regression of its latency in ABR signals. Our design process was iterative, starting with a simple 1D CNN and progressing toward a more advanced architecture that includes ResNet (16) blocks to improve performance and robustness. The final design is shown in Figure 5. The design includes a CNN backbone, followed by a regressor for wave V location and a classifier for wave V existence. The regressor head essentially has four ResNet-like layers. Each layer has a combination of convolution, batch normalization, and ReLU layers with skip connections. The standard residual blocks are modified to include temporal dilation rates of 2, 4, and 8, enabling multiscale feature extraction over varying latency spans. An adaptive average pooling layer is applied after the final residual block to aggregate temporal features before a fully connected layer.

Figure 5
Flowchart of a neural network architecture for processing paired ABR signals. It includes convolutional blocks with Batch Norm, ReLU activations, and fully connected layers. The CNN Backbone processes inputs through multiple layers, adaptive pooling, and concatenation, leading to a Position Regressor and Existence Classifier.

Figure 5. Multi-task model architecture.

The classification head uses a series of convolutional blocks, each combining convolution, batch normalization, and ReLU layers. To take advantage of the shared representations, intermediate feature maps from the regression head are concatenated to the convolution blocks before feeding them to the classification layer.

2.4 Paired signal approach

To enhance the model's ability to capture meaningful patterns, we introduced a paired-signal approach, in which an ABR recording (test signal) is paired with its corresponding high-intensity (specifically, 80 dB) reference signal from the same experiment. The paired signal is input to the model, which is designed to predict whether wave V exists in the test signal and its position. High-intensity ABR signals generally exhibit clearer waveforms, providing structural cues that facilitate the interpretation of lower-intensity responses. By pairing signals recorded from the same ear and session, we integrate complementary information that enhances wave V classification and latency prediction.

2.5 Loss functions for regression

To train the regression head, we investigated several loss functions. Two of those are the traditional mean squared error (MSE) and mean absolute error (MAE). For a batch size of N, the true value yi, and the predicted value ŷi, the MSE loss and the MAE loss are

MSE=1Ni=1N(y^iyi)2,    (1)

and

LMAE=1Ni=1N|y^i-yi|,    (2)

respectively.

In addition, we used a weighted squared error loss function, where deviation from the true value was penalized differently:

LW=1Ni=1Nwi(y^i-yi)2.    (3)

For the weights wi, we tested two functions. The first one is denoted as LW-sigmoid, where the weights are based on a sigmoid function:

wi=1.0+σ(|y^i-yi|-m).    (4)

The second one is denoted as LW-exponential, where the weights are based on an exponential function:

wi=1.0+exp(-k·(|y^i-yi|-m)),    (5)

with m and k being some scalar values. The behaviors of these functions (LMSE, LMAE, LW-sigmoid, LW-exponential) are plotted in Figure 6 for m = 3 and k = 0.5. In the case of LW-sigmoid, the function behaves similarly to the MAE loss for small deviations, but for large deviations, it penalizes more than the MAE loss. The exponential loss has the opposite behavior: It penalizes more than the MAE loss when the deviation is small, but behaves like the MAE loss as the deviation gets larger.

Figure 6
Line graph comparing different loss functions against absolute error. The x-axis represents absolute error from 0 to 10, while the y-axis represents loss values from 0 to 20. Lines represent L_MAE (blue), L_MSE (green, dotted), L_W-sigmoid (red, dashed), and L_W-exponential (purple, dash-dotted). L_MSE rises steeply, while L_MAE increases linearly. L_W-sigmoid and L_W-exponential display gradual increases.

Figure 6. Comparison between different loss functions for the regression task.

2.6 Loss functions for classification

To train the classification head to detect the presence or absence of wave V in an ABR signal, we employed the binary cross-entropy (BCE) loss, a standard choice for binary classification tasks. For each sample xi with ground-truth label yi∈{0, 1} (where 1 denotes the presence of wave V and 0 denotes the absence) and predicted probability ŷi, the BCE loss is defined as:

LBCE=-1Ni=1N[yilog(y^i)+(1-yi)log(1-y^i)],    (6)

which essentially measures how well the predicted probabilities match the true labels.

2.7 Training process

To maximize clinical utility, we employed a multitask learning framework (17) that simultaneously predicts wave V latency (regression) and classifies its presence (binary classification). Our training procedure consisted of three stages. First, we trained the latency prediction branch to convergence. Next, we froze its convolutional backbone and shared layers, and used these pre-trained features to initialize the classification branch (the right side of Figure 5). Finally, we unfroze all network parameters and fine-tuned the entire model using a combined loss function, which is a weighted sum of classification loss and regression loss functions. In the experiments, the weight of the classification loss is set to double the weight of the regression loss. This multi-task approach allows the model to learn shared representations that benefit both tasks, enhancing the overall performance and robustness of the system.

2.8 Hardware configuration

Inference speed was evaluated on the system described in Table 3. On the GPU, the model achieves an average inference time of 3.31 ms per paired ABR signal (median: 3.14 ms, std: 0.58 ms), corresponding to a throughput of 303 samples/second. On CPU, inference takes 5.37 ms on average (median: 5.29 ms, std: 0.55 ms), corresponding to 186 samples/second. These results confirm that the model is suitable for near real-time clinical use.

Table 3
www.frontiersin.org

Table 3. Hardware specifications used for training and testing.

3 Results

The performance of the proposed model is evaluated on the test set using several metrics. To show the effectiveness of the multi-task approach, we also evaluated single-task CNN and single-task ResNet approaches.

The single-task CNN consists of CNN layers followed by a fully connected layer that ends with a regressor for wave V latency prediction or a binary classifier for classifying the existence of wave V. The model is separately trained for the regression and regression tasks, and compared with the proposed multi-task approach.

The single-task ResNet model is essentially the regressor side of the multi-task model shown in Figure 5. It consists of a CNN backbone followed by multiple layers of residual blocks. It is trained for the regression task and compared with proposed multi-task approach.

For all models, we investigated the effect of having a paired input instead of a single input. The input layer is modified to take a paired or single input, keeping the rest of the network unchanged. Each model is separately trained for the same number of epochs to ensure a fair comparison.

The models with trained weights are available at https://github.com/youssefdarahem/ABR_analysis_model.

3.1 Performance metrics

To measure regression performance, we used three performance metrics. The first two are the mean absolute error (MAE) and the coefficient of determination (R2), which are commonly used metrics in regression problems. The third metric is the “accuracy for a given error tolerance”, which is the percentage of predictions within a specified error margin (tolerance) of the ground truth value. As tolerance increases, accuracy improves due to a wider margin of error. High accuracy at low tolerances indicates superior model performance.

To measure classification performance, we use standard classification metrics, including accuracy, precision, recall, F1 score, and area under the curve (AUC).

3.2 Results and comparisons

The regression results are shown in Table 4 (The regression loss function used for the results in this table is the weighted squared error with exponential decay). Three different models are evaluated. For each model, single-input and paired-input versions are trained and tested. For each model, it is seen that the paired-input version produces better R2 and MAE results compared to the single-input version, supporting the hypothesis that the use of multiple ABR signals helps to distinguish the locations of the waves. Regarding the accuracy for different tolerances, it is observed that the paired-input versions provide more gains for smaller tolerances.

Table 4
www.frontiersin.org

Table 4. Regression (prediction of wave V latency) results.

Considering all metrics, the best performing model is the proposed multi-task model, with an R2 value of 0.90, MAE value of 1.72 index error, which is about 0.11 ms. Representative model outputs are visualized in Figure 7, demonstrating accurate Wave V localization even in low-intensity or noisy recordings, thanks to the paired-input context.

Figure 7
Two line graphs show auditory brainstem response (ABR) signals at intensities of forty and eighty decibels. The left graph displays lower amplitude variations than the right. Both graphs compare actual and predicted Wave V timings, marked by different lines, with error values inset; the left shows an error of 0.08 milliseconds, and the right shows an error of 0.01 milliseconds.

Figure 7. Example model predictions for a low and a high intensity ABR signal. Each plot includes the zoomed-in region of Wave V. Ground-truth Wave V latency is marked with a red vertical line; model prediction is shown in green.

The classification performances are summarized in Table 5. The proposed model outperformed the single-task CNN model in all metrics, including accuracy, precision, recall, and F1 score, demonstrating the benefits of deeper feature extraction and regularization. These results underscore the value of advanced neural architectures in improving ABR signal classification. For both models, the paired-input version produces better results compared to the single-input version.

Table 5
www.frontiersin.org

Table 5. Classification (prediction of wave V existence) results.

4 Analysis

We have done analyses on the effects of loss functions, architectural choices, and click intensities.

4.1 Effect of loss function

On the regression head of the model, we investigated the use of alternative loss functions. The accuracy results for different tolerance values are shown in Figure 8. In all cases, the MSE loss leads to the worst performance. The MAE loss provides marked improvement over the MSE loss, especially in the low error tolerance scenarios (i.e., 0.06 ms and 0.12 ms). The best performance overall is achieved with the weighted loss that uses an exponential function and behaves somewhere between MSE and MAE loss as shown in Figure 6.

Figure 8
Bar graph showing accuracy percentages across different error tolerances in milliseconds (0.06, 0.12, 0.18, 0.24, 0.30). Four methods are compared: L-MSE (green), L-MAE (blue), L-W-sigmoid (red), and L-W-exponential (purple). Accuracy increases with error tolerance, reaching above 90% at 0.30 milliseconds for all methods.

Figure 8. Accuracy achieved with different loss functions.

4.2 Effect of intermediate-level feature sharing

A regular multi-task network has a common backbone followed by multiple separate heads. In this work, we investigated a variation where there is sharing in the intermediate layers. Specifically, as shown in Figure 5, an intermediate tensor from the regression head is fed to the classifier head, transferring important features. This, in effect, reduces the need for a more complex classifier head and leads to better performance with limited data. We tested connections from different intermediate levels and achieved the best performance when low-level features were transferred. The network is trained with and without the connection, and the results are shown in Table 6.

Table 6
www.frontiersin.org

Table 6. Effect of the intermediate connection in the proposed model.

4.3 Effect of stimulus volume

As the stimulus volume decreases, wave V becomes more difficult to detect in the ABR signal. In the previous section, we reported results without making any distinction of the stimulus intensity. Here, we analyze the performance of low-intensity and high-intensity groups. We divided the test data into two groups, namely, high intensity (≥60 dB) and low intensity (< 60 dB) groups. Figure 9 shows the wave V localization results; as expected, the accuracy is significantly higher for the high-intensity group. A similar trend is also observed for the classification performance, as shown in Table 7.

Figure 9
A bar chart compares accuracy percentages at different error tolerances for two intensity levels: greater than or equal to 60 decibels (green) and less than 60 decibels (orange). At 0.06 milliseconds, the accuracy difference is 10.2 percent. At 0.12 milliseconds, it is 4.5 percent, and at 0.30 milliseconds, it is 3.8 percent. Green bars consistently show higher accuracy.

Figure 9. Wave V localization performance for different intensity (dB) levels.

Table 7
www.frontiersin.org

Table 7. Classification performance by intensity (dB) level.

These results also support the idea of a paired input approach. A high-intensity reference signal, which has prominent wave characteristics, provides important information for localizing and classifying the low-intensity signal.

5 Conclusions

In this study, we presented a multi-task deep learning model to address two critical challenges in ABR analysis: The detection of wave V latency and classification of its presence. By unifying classification and regression into a single architecture and optimizing both tasks together, we achieve greater robustness and precision than separate, single-task models. We presented a paired-input approach, and investigated various loss functions for training.

Key findings include the following.

Paired-input approach: The paired-input strategy provides the network with complementary information: the high-intensity waveform acts as a clear reference in which Wave V is readily visible, guiding the interpretation of noisier, low-intensity recordings. Clinically, audiologists routinely compare high-intensity and low-intensity responses when identifying wave V; our method emulates this practice and embeds it directly into the model.

Incorporating high-dB reference signals alongside target ABR signals enhanced the model's ability to detect subtle patterns, improving localization accuracy by 6% and classification performance, particularly in low-intensity (≤60 dB) scenarios.

Loss functions: To encourage high precision around clinically relevant latency thresholds, we evaluated two new loss functions in addition to the MAE and MSE losses. The new loss functions place relatively more weight on either small errors or large errors during training.

The loss function, which is tailored to focus on tight error margins, yields more reliable latency estimates for diagnostic purposes. The model achieves around 88% accuracy for an error margin of 0.12 ms, and 81% accuracy for an error margin of 0.06 ms.

Multi-task model with common backbone and shared layers: Our multi-task architecture leverages shared features for presence detection and latency regression. After pretraining the regression branch, we fine-tuned the full network with a combined loss, allowing the classifier and regressor to reinforce each other's learned representations. This synergy yields notable gains, especially in localization performance, while aligning closely with clinical workflows by delivering both outputs in a single inference pass.

The model achieved an F1 score of 0.92 for classification and an R2 of 0.90 for latency prediction, surpassing single-task models.

Automated analysis of ABR signals offers an objective and efficient way of diagnosing auditory disorders. Future work may explore broader datasets, real-time implementation, and integration with clinical devices. This work underscores the potential of deep learning in advancing electrophysiological diagnostics, bridging the gap between computational models and practical audiological care.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

Ethical approval for this study was obtained from the Ethics Committee for Non-Interventional Research of Istanbul Medipol University (Protocol number: E-10840098-772.02-785) on 26/01/2023. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants' legal guardians/next of kin.

Author contributions

YD: Methodology, Software, Writing – original draft, Writing – review & editing. OY: Data curation, Supervision, Writing – review & editing. HS: Data curation, Resources, Writing – review & editing. BM: Data curation, Resources, Writing – review & editing. HA: Funding acquisition, Supervision, Writing – review & editing. BG: Funding acquisition, Methodology, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This study was supported by TUBITAK under the Grant Number 123E478.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^The trained model is available at https://github.com/youssefdarahem/ABR_analysis_model.

References

1. Young A, Cornejo J, Spinner A. Auditory Brainstem Response. Treasure Island (FL): StatPearls Publishing. (2025). Available online at: https://www.ncbi.nlm.nih.gov/books/NBK564321/ (Accessed January 12, 2023).

Google Scholar

2. Jewett D, Williston J. Auditory evoked far fields averaged from the scalp of humans. Brain. (1971) 94:681–96. doi: 10.1093/brain/94.4.681

PubMed Abstract | Crossref Full Text | Google Scholar

3. van Zanten G, Versnel H, van der Stoep N, Koopmans W, Hoetink A. Short-latency evoked potentials of the human auditory system. In:Naz S, , editor. Auditory System. Rijeka: IntechOpen. (2022). doi: 10.5772/intechopen.102039

Crossref Full Text | Google Scholar

4. Atcherson SR, Lim TJ, Moore PC, Minaya CP. Comparison of auditory brainstem response peak measures using ear lobe, mastoid, and custom ear canal reference electrodes. Audiol Res. (2012) 2:e3. doi: 10.4081/audiores.2012.e3

PubMed Abstract | Crossref Full Text | Google Scholar

5. Vidler M, Parker D. Auditory brainstem response threshold estimation: subjective threshold estimation by experienced clinicians in a computer simulation of the clinical test. Int J Audiol. (2004) 43:417–29. doi: 10.1080/14992020400050053

PubMed Abstract | Crossref Full Text | Google Scholar

6. Bradley AP, Wilson WJ. Automated analysis of the auditory brainstem response. In: Proceedings of the 2004 Intelligent Sensors, Sensor Networks and Information Processing Conference, 2004. Melbourne, VIC: IEEE. (2004). p. 541–545.

Google Scholar

7. Delgado RE, Ozdamar O. Automated auditory brainstem response interpretation. IEEE Eng Med Biol Magaz. (1994) 13:227–37. doi: 10.1109/51.281682

Crossref Full Text | Google Scholar

8. Davey R, McCullagh P, Lightbody G, McAllister G. Auditory brainstem response classification: a hybrid model using time and frequency features. Artif Intell Med. (2007) 40:1–14. doi: 10.1016/j.artmed.2006.07.001

PubMed Abstract | Crossref Full Text | Google Scholar

9. McKearney RM, Simpson DM, Bell SL. Automated wave labelling of the auditory brainstem response using machine learning. Int J Audiol. (2024) 63:966–74. doi: 10.1080/14992027.2024.2404537

PubMed Abstract | Crossref Full Text | Google Scholar

10. Ji J, Wang X, Jing X, Zhu M, Pan H, Jia D, et al. ABR-attention: an attention-based model for precisely localizing auditory brainstem response. IEEE Trans Neural Syst Rehabilitat Eng. (2024) 32:3179–88. doi: 10.1109/TNSRE.2024.3445936

PubMed Abstract | Crossref Full Text | Google Scholar

11. Liang S, Xu J, Liu H, Liang R, Guo Z, Lu M, et al. Automatic recognition of auditory brainstem response waveforms using a deep learning-based framework. Otolaryngol-Head Neck Surg. (2024) 171:1165–71. doi: 10.1002/ohn.840

PubMed Abstract | Crossref Full Text | Google Scholar

12. McKearney RM, MacKinnon RC. Objective auditory brainstem response classification using machine learning. Int J Audiol. (2019) 58:224–30. doi: 10.1080/14992027.2018.1551633

PubMed Abstract | Crossref Full Text | Google Scholar

13. Ma J, Seo JH, Moon IJ, Park MK, Lee JB, Kim H, et al. Auditory brainstem response data preprocessing method for the automatic classification of hearing loss patients. Diagnostics. (2023) 13:3538. doi: 10.3390/diagnostics13233538

PubMed Abstract | Crossref Full Text | Google Scholar

14. Majidpour A, Jameel SK, Majidpour J, Bagheri H, Rashid TA, Nazeri A, et al. Detection of auditory brainstem response peaks using image processing techniques in infants with normal hearing sensitivity. Biomed Signal Process Control. (2023) 86:105117. doi: 10.1016/j.bspc.2023.105117

Crossref Full Text | Google Scholar

15. O'Shea K, Nash R. An Introduction to Convolutional Neural Networks. (2015). Available online at: http://arxiv.org/abs/1511.08458

Google Scholar

16. He K, Zhang X, Ren S, Sun J. Deep Residual Learning for Image Recognition. (2015). Available online at: https://arxiv.org/abs/1512.03385

Google Scholar

17. Zhang Y, Yang Q. A Survey on multi-task learning. IEEE Trans Knowl Data Eng. (2022) 34:5586–609. doi: 10.1109/TKDE.2021.3070203

Crossref Full Text | Google Scholar

Keywords: auditory brainstem response (ABR), deep learning, convolutional neural networks (CNNs), transfer learning, machine learning (ML)

Citation: Darahem Y, Yilmaz O, Saldirim HB, Mutlu B, Ates HF and Gunturk BK (2025) A Dual-input deep learning architecture for classification and latency estimation in ABR signals. Front. Med. 12:1693921. doi: 10.3389/fmed.2025.1693921

Received: 27 August 2025; Accepted: 22 October 2025;
Published: 11 November 2025.

Edited by:

Siquan Wang, Columbia University, United States

Reviewed by:

Anandhan E. Dhanasingh, MED-EL, Austria
Fatih Ozyurt, Firat Universitesi Muhendislik Fakultesi, Türkiye

Copyright © 2025 Darahem, Yilmaz, Saldirim, Mutlu, Ates and Gunturk. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Bahadir K. Gunturk, YmtndW50dXJrQG1lZGlwb2wuZWR1LnRy

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.