- 1The Chinese University of Hong Kong, Shenzhen, Guangdong, China
- 2Engineering Research Center, National Health Data Institute, Shenzhen, China
- 3Mobile Health Ministry of Education, China Mobile Joint Laboratory, Xiangya Hospital, Central South University, Changsha, China
- 4National Clinical Research Center for Geriatric Disorders, Xiangya Hospital, Changsha, China
- 5Institute of Artificial Intelligence, North China University of Technology, Tangshan, China
Introduction: This work proposes an explainable deep learning framework to transform complex biosignal dynamics into interpretable health assessments. The core of our approach is the PhysioGraph Inference Network (PGIN), which combines temporal graph reasoning with probabilistic modeling to capture dynamic inter sensor dependencies under physiological priors.
Methods: To further enhance adaptability, an Adaptive Health State Inference Mechanism (AHSIM) is introduced to adjust diagnostic granularity based on uncertainty and signal entropy.
Results and discussion: Evaluations on four biosensing datasets show that our framework achieves superior diagnostic accuracy (up to 92.48%) and AUC (up to 93.65%), outperforming several transformer-based baselines such as RoBERTa and T5. Furthermore, the model provides transparent uncertainty estimates, making it suitable for deployment in clinical and wearable scenarios. By integrating physiological semantics and model interpretability, our framework bridges the gap between black-box AI and trustworthy biomedical intelligence.
1 Introduction
The increasing demand for accurate, real-time health diagnostics and biomedical monitoring has led to a surge in biosensing technologies. However, while biosensors can capture complex physiological signals with high fidelity, the interpretation of such data remains a significant challenge. Traditional analysis methods often fail to provide both precise predictions and transparent reasoning, which are crucial for clinical decision-making and patient trust (Ansari and Cho, 2008). Not only do healthcare professionals require reliable interpretations of biosensing data, but they also need explanations that can be communicated clearly to patients (Pantelopoulos and Bourbakis, 2008). Moreover, as biosensing systems become more integrated into portable and wearable devices, the demand for interpretable, automated, and scalable analytical tools becomes even more pressing (Asada et al., 2003). This necessitates frameworks that are not only technically robust but also capable of offering actionable insights in a clear and understandable manner. Hence, developing an explainable deep learning framework tailored to the complexities of biosensing data is both timely and essential for advancing biomedical engineering and enhancing real-time diagnostic capabilities (Jegan and Anusuya, 2018).
Initial efforts to interpret biosensing data focused on rule-based systems that relied on predefined thresholds and logical reasoning to analyze physiological signals. These methods utilized structured frameworks to encode domain knowledge, enabling transparent and traceable decision-making processes (Wang and Katz, 2010). For instance, systems designed for ECG or EEG analysis could identify anomalies by comparing signal patterns against expert-defined criteria (Mohanty and Kougianos, 2006). While these approaches provided interpretability and consistency, they were often limited in their ability to handle noisy or high-dimensional data. Furthermore, their reliance on static rules made them inflexible when applied to diverse patient populations or evolving sensor technologies (Chan et al., 2006). The need for manual updates to maintain these systems further constrained their scalability, underscoring the necessity for more adaptive and automated solutions.
To enhance adaptability, researchers began employing statistical models capable of learning directly from labeled biosensing datasets. These methods, including support vector machines, decision trees, and ensemble techniques, demonstrated improved performance by identifying patterns and correlations within the data (Mendelson, 2005). By incorporating feature engineering, these models could extract relevant physiological descriptors, thereby improving accuracy in tasks such as arrhythmia detection and glucose monitoring (Junfeng et al., 2005). However, the reliance on handcrafted features often limited their ability to capture complex, nonlinear relationships inherent in biosensing signals. Additionally, as model complexity increased, their interpretability diminished, raising concerns about their applicability in critical healthcare scenarios (Nandakumar et al., 2010). These limitations highlighted the need for more advanced frameworks that could balance performance with transparency.
The advent of deep learning introduced a paradigm shift in biosensing data analysis by enabling end-to-end learning from raw signals. Neural network architectures, such as convolutional and recurrent networks, excelled in automatically extracting hierarchical features, achieving state-of-the-art results in applications like seizure prediction and sleep stage classification (Bronzino, 2005). Pretrained models further enhanced generalization by leveraging knowledge from related datasets, reducing the dependency on large labeled datasets (Ibrahim et al., 2015). Despite these advancements, the opaque nature of deep learning models posed significant challenges for clinical adoption, as their decision-making processes were difficult to interpret (Lee et al., 2012). Efforts to improve transparency, such as attention mechanisms and saliency maps, have shown promise but remain insufficient for widespread clinical validation (Heileman et al., 2013). Thus, while deep learning has revolutionized biosensing data interpretation, integrating explainability into these models remains a critical area of ongoing research.
While recent deep learning approaches have demonstrated promising performance in biosignal interpretation, several limitations persist that hinder their clinical deployment. First, many state-of-the-art models are developed with a focus on prediction accuracy but offer limited interpretability, making them less suitable for use in high-stakes medical environments where transparency is critical. Second, most existing models rely heavily on large labeled datasets and handcrafted feature engineering, which limits their generalizability across patient populations with diverse physiological profiles and varying sensor configurations. Third, traditional models often struggle with noisy or incomplete data streams, which are common in real-world wearable sensing environments. Finally, although attention mechanisms and gradient-based saliency maps have been introduced to improve interpretability, they generally provide post hoc explanations that lack consistency and physiological grounding. These challenges underscore the need for novel frameworks that not only maintain high performance but also ensure domain-aligned interpretability, uncertainty quantification, and robust reasoning under data variability.
Despite these advancements, a precise research gap remains insufficiently addressed in the current literature. Most existing biosensing interpretation frameworks prioritize predictive performance while neglecting the underlying interpretability required for medical decision-making. In particular, deep learning models often operate as black boxes, lacking mechanisms to trace diagnostic outcomes back to physiologically meaningful features. Furthermore, few models incorporate explicit uncertainty quantification, which is critical for clinical deployment in high-risk settings. While attention-based and saliency-based explanation techniques have gained popularity, they provide limited insight into dynamic signal interdependencies and rarely align with domain-specific physiological structures. In addition, current methods generally treat multimodal signals as flat feature vectors, failing to model inter-sensor relationships and temporal evolution in a structured manner. These limitations create a significant barrier to building trustworthy and generalizable biosensing intelligence. In this work, I aim to bridge these gaps by introducing a graph-based, uncertainty-aware, and physiologically informed diagnostic framework that not only achieves state-of-the-art performance but also supports real-time and interpretable health assessments across diverse sensing environments. To overcome the limitations of symbolic rigidity, the feature dependency of traditional machine learning, and the opacity of deep learning models, I propose an explainable deep learning framework for biosensing data interpretation tailored to biomedical engineering and real-time diagnostics. This framework seeks to harmonize the predictive strength of deep learning with the clarity and traceability essential for clinical contexts. It leverages model-agnostic interpretability techniques, such as SHAP and counterfactual explanations, alongside domain-informed architectural choices to ensure that every prediction can be contextualized and justified. Not only does this framework support real-time analysis of complex biosignals, but it also enables transparent reasoning that can be validated and trusted by medical professionals. Furthermore, it is designed for integration into wearable and point-of-care devices, facilitating scalable deployment in diverse healthcare environments. As such, our approach represents a step toward interpretable, data-driven health intelligence capable of supporting decision-making across a spectrum of biomedical applications.
• The framework incorporates modular interpretability layers that align deep neural outputs with domain-specific signal characteristics, enhancing transparency in real-time diagnostics.
• It is optimized for multimodal biosensing data and generalizes well across different health monitoring tasks, ensuring efficient performance in both clinical and wearable settings.
• Empirical evaluation on public and proprietary datasets shows improved diagnostic accuracy with explainability scores outperforming existing attention- and saliency-based methods.
2 Related work
2.1 Explainable approaches in biosensing interpretation
The integration of deep learning techniques into biosensing has significantly enhanced the ability to interpret complex biomedical signals, including electrochemical, optical, acoustic, and electrophysiological data (Bronzino, 2005). Explainable models have been developed to address the need for transparency in clinical applications, employing methods such as attention mechanisms, gradient-based saliency maps, and model-agnostic explanations (Jegan and Anusuya, 2018). Attention-based architectures have been utilized to identify critical spectral or temporal features in spectrophotometric biosensing, enabling domain experts to verify alignment with biochemical signatures (Ibrahim et al., 2015). Gradient visualization techniques, such as Grad-CAM and Integrated Gradients, have been applied to trace the influence of sensor perturbations on diagnostic outputs, aiding in the detection of spurious correlations (Lee et al., 2012). Model-agnostic approaches, including LIME, provide localized interpretability by generating surrogate models, though they often lack insights into global decision structures (Perumal and Hashim, 2014). Challenges arise as data dimensionality increases, complicating the scalability of interpretability methods and imposing computational burdens (Heileman et al., 2013). Pre-computed explanation templates and lightweight surrogate models have been proposed to address these issues, particularly in real-time applications (Tittl et al., 2019). Robustness to sensor noise is a critical focus, with methods enforcing consistency in explanations under signal perturbations to ensure biomedical plausibility (Vashistha et al., 2018). Human-in-the-loop systems have been explored, allowing clinicians to iteratively refine model parameters based on interpretability feedback (Zhang et al., 2011). Hybrid frameworks that combine symbolic physiological models with learned components are emerging, offering explanations grounded in biomechanical relationships while maintaining high diagnostic performance (Nguyen et al., 2015).
2.2 Real-time deep diagnostics architectures
The demand for real-time health diagnostics has driven the development of deep learning architectures optimized for low-latency performance, particularly in edge and mobile platforms (Bellagambi et al., 2017). Lightweight neural networks, such as MobileNet variants and depthwise separable convolutions, have been employed to enable efficient on-device inference (Armghan et al., 2023). Convolutional-LSTM hybrids capture spatiotemporal dependencies in biosignal streams while balancing computational efficiency (Nisha and Meeral, 2021). Early-exit classifiers allow preliminary predictions with partial computation, escalating to full models only when necessary (Qureshi et al., 2023). Data compression techniques, including feature hashing and dimensionality reduction, preprocess biosensor outputs to reduce input size and improve inference speed (Schackart III and Yoon, 2021). Parallel processing pipelines ensure continuous throughput by simultaneously handling biosignal windows and prediction outputs (Esmaeili et al., 2023). Event-driven architectures, coupled with shallow networks, have been implemented for ultralow-latency scenarios such as arrhythmia detection (Sadak et al., 2022). Edge-cloud hybrid systems offload computationally intensive tasks to nearby servers, maintaining functionality through fallback submodels on the device (Boulogeorgos et al., 2020). Quantized neural networks integrated into microcontroller hardware accelerators leverage integer arithmetic to minimize latency (Guo et al., 2022). Sensor fusion modules combining multimodal data, such as ECG and PPG, enhance diagnostic reliability while preserving real-time responsiveness (Rong et al., 2023). Validation studies often focus on optimizing end-to-end latency from sensor acquisition to diagnostic output, ensuring compliance with real-time constraints under limited power budgets (Raji et al., 2022).
2.3 Frameworks combining interpretability and real-time sensing
The intersection of interpretability and real-time diagnostic capability has led to the development of frameworks that balance transparency and low-latency performance in biomedical applications (Bronzino, 2005). Attention-augmented lightweight networks have been designed to produce both rapid classifications and interpretable importance maps in a single forward pass (Jegan and Anusuya, 2018). Multi-task learning approaches incorporate auxiliary interpretability objectives, such as alignment with clinical feature importance, to ensure transparent decision-making pathways without increasing computational demands (Ibrahim et al., 2015). Hybrid models utilize shallow interpretative networks to flag critical signal regions, relaying them to deeper modules only when necessary (Lee et al., 2012). Distillation-based pipelines create compact student models that replicate both the predictive accuracy and explanation patterns of larger teacher networks (Perumal and Hashim, 2014). Temporal segmentation methods isolate salient events within streaming biosignal data, feeding only these segments to more expressive models to balance latency and interpretability (Heileman et al., 2013). Symbolic components derived from physiological knowledge have been integrated into neural networks, constraining behavior to align with clinically validated patterns while accelerating inference (Tittl et al., 2019). Pipeline frameworks deliver concise explanation outputs, such as highlighted waveform regions, within the time constraints of continuous monitoring systems (Vashistha et al., 2018). User studies evaluating clinician trust and diagnostic accuracy indicate that transparent systems improve situational awareness and error detection compared to opaque models (Zhang et al., 2011). Adaptive explanation fidelity has been explored, providing minimal explanations in routine cases and deeper insights for uncertain or atypical patterns (Nguyen et al., 2015). These advancements demonstrate the potential of combining interpretability and real-time responsiveness to deliver actionable and trustworthy diagnostics in biosensing applications (Bellagambi et al., 2017).
Recent advances in biomedical signal modeling and intelligent health diagnostics have also highlighted the importance of integrating formal mathematical frameworks into practical sensing systems. For example, Evirgen et al. (2025) proposes a fractional-order modeling approach to represent physiological systems with long-term memory and time-dependent characteristics, which is particularly relevant to biosignals exhibiting complex temporal dynamics. In a related direction, Shams et al. (2024) presents an uncertainty-aware classification framework that leverages hybrid decision mechanisms to improve robustness in physiological signal interpretation under noisy conditions. Work by Shams (2024) explores hybrid intelligent modeling strategies that fuse data-driven learning with expert knowledge, offering enhanced interpretability in automated health diagnostics. Complementing this, Evirgen et al. (2023) provides a bioengineering perspective on explainable AI for biosensing, focusing on human-centered model transparency and clinical applicability. From a theoretical standpoint, Shams et al. (2021) investigates nonlinear dynamics and stability properties in sensor-integrated physiological systems, providing insights into the structural behavior of biosensor networks. Meanwhile, Rufai and Ramos (2020) contributes foundational analysis on the mathematical control and bifurcation properties of real-time biosensor frameworks. These studies collectively demonstrate the growing relevance of hybrid models, uncertainty quantification, and physiological structure alignment in the development of next-generation diagnostic frameworks. Our work builds upon these themes by combining graph-based reasoning, probabilistic latent modeling, and domain-informed interpretability, thereby extending this body of research with a unified and deployable solution tailored for real-time health monitoring.
3 Methods
3.1 Overview
The integration of biosensing data into health diagnostics has fundamentally transformed the measurement, interpretation, and application of physiological signals in clinical and personalized medical contexts. This study proposes a comprehensive framework that integrates multimodal biosensor signals—such as ECG, EMG, and PPG—for reliable and interpretable health state inference. The proposed approach addresses key challenges in modeling biosignal dynamics, constructing domain-aware inference mechanisms, and developing a diagnostic framework that leverages temporal and multivariate physiological measurements.
This section outlines the methodological pipeline in detail. Section 3.2 formalizes the biosensing problem within a symbolic framework, defining the structure of biosignal data, their latent representations, and the classification problem for diagnostic purposes. This formalization enables precise characterization of the statistical structures underlying biosignal variability and health-related anomalies, providing a foundation for the development of principled models. The symbolic definitions introduced in this section enhance the interpretability of biosensing dynamics and offer generalizable abstractions applicable across multimodal biosensor platforms.
Section 3.3 introduces the proposed model, termed the PhysioGraph Network, which is a domain-specific inference engine designed to process biosensor streams under constraints of interpretability, robustness, and physiological plausibility. The architecture combines temporal graph reasoning with deep probabilistic encodings to capture latent dependencies among physiological signals. Unlike conventional deep learning models that operate on generic feature representations, the PGIN explicitly embeds structural priors such as known physiological correlations (cardiac-respiratory coupling) and medically validated signal ontologies (ECG-PPG synchrony). This ensures that the learned latent space aligns with causal and interpretable pathways, thereby improving sample efficiency and clinical trust. This integration results in a sample-efficient model that aligns with clinically relevant causal pathways. The architectural innovation lies in its fusion of physiological connectivity priors with data-driven learning mechanisms, producing interpretable latent representations suitable for medical interpretation and downstream decision-making.
In our model, physiological priors are represented through dynamic graph structures that encode known signal dependencies—for instance, respiratory-cardiac coupling or synchronized vascular responses. These priors influence edge formation via learned similarity functions that emphasize biologically meaningful interactions. The temporal smoothness regularization further reinforces stable, gradual changes in connectivity, consistent with real-world physiological transitions. Through this design, prior knowledge guides both graph topology and latent trajectory evolution. Section 3.4 details the reasoning and diagnostic framework built upon the PhysioGraph Network. This framework, referred to as the Adaptive Health State Inference Mechanism, employs a hierarchical inference strategy tailored to biosignal irregularities and inter-patient variability. By utilizing symbolic health state embeddings and real-time physiological context encoding, the framework extracts temporal signatures predictive of acute or chronic health events. Additionally, a dynamic adjustment mechanism is introduced to modulate diagnostic granularity based on estimated uncertainty and signal entropy, ensuring adaptability for deployment in both high-risk clinical environments and resource-constrained wearable platforms.
3.2 Preliminaries
In addition to traditional physiological sensors such as ECG, PPG, and EMG, recent advancements in biomedical sensing have expanded to include vision-based modalities. RGB cameras are commonly used to extract facial cues such as micro-expressions, pulse signals via remote photoplethysmography, and even subtle respiratory patterns through chest movement tracking. These sensors are particularly useful in non-contact settings such as intensive care units or telemedicine applications. Depth cameras, on the other hand, provide three-dimensional spatial information that is valuable for capturing posture, gait, and body orientation—factors closely associated with musculoskeletal and neurological assessments. Unlike 2D cameras, depth-based sensing reduces background noise and improves robustness to lighting variations. Wearable camera sensors, such as body-mounted or head-mounted vision devices, offer egocentric perspectives that facilitate context-aware health monitoring, including activity recognition, fall detection, and behavioral analysis in daily living environments. While these camera-based modalities introduce rich semantic information, they also pose unique challenges such as privacy concerns, real-time processing demands, and the need for advanced feature extraction techniques. Incorporating such diverse sensors into a unified analytical framework requires model adaptability to heterogeneous data streams and robust fusion mechanisms. Our framework is designed with modular encoding layers that can be extended to process visual modalities alongside traditional biosignals, thereby enhancing its versatility across multimodal health monitoring scenarios.
Let
where
The generation of these observed signals is assumed to be governed by a latent physiological state
where
Each sensor is associated with an observation model conditioned on the latent state (Equation 3):
where
To capture the temporal dynamics of the signals, a sequence-level functional mapping is introduced (Equation 4):
where
where
To model inter-sensor dependencies, a dynamic graph
where
The diagnostic task is formulated as a mapping
where
To enhance model interpretability and flexibility, an intermediate embedding space
and the diagnostic mapping is redefined as (Equation 9)
This modular design separates the feature extraction and classification processes, facilitating interpretability and adaptability.
Physiological signals are inherently non-stationary and may exhibit concept drift over time. To address this, an online calibration function is incorporated (Equation 10):
where
Uncertainty quantification is critical for reliable diagnostics. A confidence score function
where
This formulation establishes the theoretical foundation for the diagnostic modeling and inference strategies developed in subsequent sections. It provides a structured approach to handling biosensor data, ensuring robustness and interpretability across diverse patient populations and operational conditions.
The mapping from high-dimensional biosensor streams to interpretable latent trajectories in our framework is grounded in several theoretical assumptions inspired by both probabilistic modeling and physiological signal theory. One core assumption is that the evolution of physiological states can be described by a temporally dependent latent variable
I assume that each biosensor signal
I further posit that inter-sensor relationships are not static but dynamically modulated by latent physiological interactions. This is modeled through a time-varying graph
3.3 PhysioGraph inference network
To process and interpret complex, multivariate biosensing data for diagnostic purposes, I introduce the PhysioGraph Inference Network (PGIN). This model is designed to explicitly encode inter-sensor physiological dependencies, time-varying latent dynamics, and context-aware signal modulation. PGIN unifies temporal graph convolution, latent variable modeling, and uncertainty-aware decision heads into an end-to-end learnable system (Figure 1).
Figure 1. Overview of the methodological pipeline for biosensing data integration in health diagnostics. The process begins with the collection of high-dimensional, heterogeneous physiological signals from biosensors. Subsequently, signal cleaning and normalization are performed to remove noise, align temporal data, and standardize measurements. Finally, temporal and multivariate feature engineering extracts meaningful patterns and dynamics for downstream diagnostic inference.
3.3.1 Multimodal encoder architecture
The PGIN begins by projecting the biosensing input sequence
where
To capture inter-sensor dependencies, a dynamic physiological graph
To ensure robustness under noisy or missing sensor data, PGIN leverages a smooth edge construction mechanism combined with graph regularization. The use of SoftSym normalization stabilizes adjacency matrices across time, while edge weights are learned to down-weight unreliable sensor inputs. During training, masked dropout simulates sensor dropout, and a temporal smoothness loss
where
Figure 2. Overview of the PhysioGraph Inference Network (PGIN) architecture. Multivariate biosensing data is processed through sensor-specific encoders to generate latent node embeddings. These embeddings are propagated through a temporal graph convolution module, which captures inter-sensor dependencies and latent physiological states. The architecture integrates multimodal encoding, dynamic graph construction, and temporal modeling for robust physiological inference.
3.3.2 Graphical propagation layer
To model temporal evolution, the PGIN employs a latent state transition model. The latent physiological state
where
where
with
The graph convolutional layer not only aggregates information across sensors at each time step but also produces temporally-aware embeddings that serve as inputs to the latent state transition module. This facilitates seamless integration between spatial (inter-sensor) and temporal (sequential) modeling, capturing both instantaneous and evolving physiological patterns.
3.3.3 Uncertainty-aware decision mechanism
The diagnostic prediction is generated from a sampled latent path
where
The full learning objective combines supervised classification loss, latent regularization, and graph smoothness constraints (Equation 22):
The total loss integrates multiple objectives to jointly optimize diagnostic accuracy, uncertainty modeling, and interpretability. The cross-entropy term
PGIN is optimized end-to-end using stochastic backpropagation through time, with reparameterization applied to latent variables. The architecture ensures that physiological structure, signal uncertainty, and temporal trends are integrated into a coherent and interpretable diagnostic model. This model design lays the algorithmic foundation for the diagnostic strategy introduced in Section 3.4, where hierarchical reasoning and context-sensitive adjustment mechanisms further enhance robustness and deployment versatility.
Unlike conventional black-box AI models that often lack clinical transparency, our framework incorporates several design choices that align directly with biomedical reasoning principles. The PhysioGraph Inference Network (PGIN) utilizes structural priors that reflect established physiological dependencies—for example, signal correlations observed between cardiac, respiratory, and muscular systems. These relationships are encoded into time-varying graph structures, enabling the model to propagate information in a way that mirrors real-world inter-organ and inter-signal dynamics. This structural encoding ensures that the learned latent representations are both physiologically grounded and interpretable by medical professionals.
The use of variational latent state modeling enables probabilistic inference over health trajectories. This design allows the framework to communicate uncertainty about its predictions in a quantifiable manner, which is critical in clinical settings where overconfident misclassification can have severe consequences. Instead of producing deterministic outputs, the model provides confidence-aware predictions that account for signal ambiguity and data quality.
The hierarchical inference strategy embedded in the Adaptive Health State Inference Mechanism (AHSIM) further enhances alignment with diagnostic practice. By modulating decision granularity based on contextual factors such as signal entropy and latent state change, the system mimics the clinical reasoning process in which coarse screening precedes fine-grained differentiation. These mechanisms work in concert to translate raw biosensor inputs into structured, explainable, and clinically relevant outputs. As a result, the framework serves as a bridge between data-driven AI and domain-aligned biomedical intelligence, offering both predictive performance and interpretability required for real-world deployment.
3.4 Adaptive health state inference mechanism
Building upon the latent structure encoded by the PhysioGraph Inference Network (PGIN), I propose the Adaptive Health State Inference Mechanism (AHSIM). This strategy leverages temporally structured latent representations, integrates uncertainty-aware decision mechanisms, and adapts contextual rule flows to enable robust, personalized, and clinically actionable health diagnostics (Figure 3).
3.4.1 Temporal memory integration
Let
Figure 4. Schematic representation of the biosensor signal observation framework and latent state modeling. The latent physiological state, z(t), governs the dynamics of the biosensors, s1 through sK, which generate observed signals over time. Each sensor captures univariate or multivariate signals, xk(t), influenced by the latent state and stochastic variability.
The state vector
where
which modulates the contribution of each timestep to the global inference state (Equation 27):
3.4.2 Uncertainty modulation framework
I apply a decision gating module to map the aggregated state to a final diagnostic outcome (Equations 28, 29):
where
3.4.3 Hierarchical decision logic
AHSIM also supports hierarchical health reasoning. Let
To accommodate deployment-specific tolerances, I design an adaptive decision thresholding (Equations 34, 35):
The overall optimization combines multi-resolution diagnostic objectives (Equation 36):
4 Experimental setup
4.1 Dataset
Biosensor Signal Analysis Dataset Sutar et al. (2025) comprises a diverse range of physiological signals collected from wearable biosensors, including electrodermal activity (EDA), skin temperature, and accelerometer data. These multimodal data streams were gathered from a heterogeneous population under different activity conditions such as resting, walking, and stress-induced tasks. Each data sample is time-stamped, enabling fine-grained temporal analysis. The dataset emphasizes variability in sensor placement, motion artifacts, and real-world conditions to challenge signal processing robustness. Preprocessing steps include low-pass filtering and normalization, ensuring consistency across recordings. The dataset serves as a benchmark for evaluating machine learning algorithms in biosignal classification, anomaly detection, and multimodal fusion tasks, particularly under non-ideal recording environments. Real-Time Health Monitoring Dataset Liu and Du (2024) contains continuous recordings of physiological parameters including heart rate, oxygen saturation, and respiration rate, captured from real-time remote monitoring systems deployed in outpatient settings. It includes both baseline and abnormal health states, annotated by clinical professionals. Each record spans several hours, enabling exploration of long-term health trends and real-time anomaly detection algorithms. The dataset includes contextual metadata such as age, gender, and clinical history, enabling patient-specific modeling. It is suitable for research in predictive diagnostics, real-time monitoring models, and wearable system calibration. Biomedical Sensor Data Collection Suthar et al. (2023) offers synchronized multimodal recordings from multiple biosensors including EEG, EMG, ECG, and PPG signals. The dataset targets laboratory-controlled experimental sessions with varied stimuli to elicit distinct physiological responses. High-resolution annotations mark significant events, enabling the study of reaction-based models and supervised learning approaches. It is tailored for evaluating the temporal resolution of biosignal analytics and exploring signal interdependencies. This dataset provides an ideal ground for sensor fusion studies, as each subject was recorded under identical experimental protocols. Deep Learning Biosensing Dataset Geng et al. (2022) was designed specifically for training deep neural networks on raw sensor inputs. It includes a large volume of labeled biosensor data collected from wearable systems across multiple sessions and subjects. Data types include ECG, body temperature, and galvanic skin response, with synchronized labels for health conditions such as fatigue, dehydration, and stress. The dataset provides extensive training-validation-test splits, data augmentation protocols, and baseline benchmarks, fostering reproducibility in deep learning model development. It also supports unsupervised and semi-supervised learning paradigms, addressing the challenge of limited labeled biomedical data.
The datasets collectively span a wide range of biosignals, including ECG, EEG, EMG, PPG, EDA, skin temperature, respiration rate, and accelerometry. To maintain clinical relevance and consistency across datasets, preprocessing steps were applied uniformly. These included bandpass filtering (0.5–40 Hz for ECG), z-score normalization across temporal windows, wavelet-based motion artifact removal, and synchronization of multimodal inputs at 100 Hz. Missing values were interpolated or masked during training to simulate real-world conditions, ensuring robustness under incomplete data.
4.2 Experimental details
All experiments were conducted using PyTorch on NVIDIA A100 GPUs with 40 GB memory. The training and evaluation pipelines were implemented using a modular architecture to ensure reproducibility and scalability. For all datasets, signals were first resampled to a unified frequency of 100 Hz, and each segment was normalized using z-score normalization across the temporal axis. A sliding window of 5 s with 50% overlap was applied to construct fixed-length input samples. Signal augmentation techniques, including Gaussian noise injection, random cropping, and temporal warping, were applied during training to improve generalization.
I adopted a deep convolutional architecture inspired by ResNet-18 as the base encoder. The input signals were treated as multi-channel time-series inputs, with each modality occupying one channel. The model comprises four residual blocks followed by global average pooling and a fully connected layer. Batch normalization and ReLU activation were applied after each convolution. To support multi-modal fusion, a modality attention module was incorporated before the final classifier. For sequence-level tasks, a Bi-LSTM layer was appended after feature extraction to capture temporal dependencies. For classification, a softmax layer was used, and for regression tasks, a linear layer predicted the physiological value.
The models were trained using the Adam optimizer with an initial learning rate of 1e-3, which was decayed by a factor of 0.1 every 10 epochs. A batch size of 64 was used across all experiments. Early stopping with a patience of 20 epochs was employed to prevent overfitting. Cross-entropy loss was used for classification tasks, while mean squared error was adopted for regression problems. For datasets with class imbalance, I applied weighted loss based on inverse class frequency. Each experiment was repeated five times with different random seeds, and I reported the average and standard deviation of all metrics.
To improve the transparency, reproducibility, and replicability of our experimental procedures, I provide a detailed summary of all key parameter settings in Table 1. This includes information on the optimizer choice, learning rate scheduling, batch size, window segmentation, and hardware configurations. In particular, all experiments were conducted using the Adam optimizer with an initial learning rate of 1e-3, and early stopping was employed to prevent overfitting. I also specify architectural components such as the backbone ResNet-18 variant, Bi-LSTM sequence modeling layers, and modality attention mechanisms used for multimodal fusion. These parameters were consistently applied across all datasets unless otherwise noted. By disclosing these details, I aim to facilitate reproducibility and support future comparative studies on biosensing data interpretation frameworks.
For evaluation, I employed accuracy, precision, recall, F1-score, and area under the ROC curve (AUC) for classification. For regression tasks, Pearson correlation coefficient and RMSE were reported. To ensure fairness, all baseline models were re-implemented under identical preprocessing and training conditions. In cases where published code was available, I directly adopted the official implementations and hyperparameters. Otherwise, I followed the best practices as described in the respective papers and tuned parameters through grid search.
Ablation studies were conducted to analyze the contributions of different components including modality attention, sequence modeling, and signal augmentation. I also performed leave-one-subject-out cross-validation (LOSO) for inter-subject generalization analysis and 5-fold cross-validation for intra-subject experiments. All experimental protocols and model checkpoints are released for reproducibility. The overall experimental setup adheres closely to standards adopted in top-tier conferences such as NeurIPS, ICLR, and AAAI, ensuring scientific rigor and comparability with state-of-the-art benchmarks.
4.3 Comparison with SOTA methods
Table 2 presents the performance comparison of our proposed method with several state-of-the-art (SOTA) baselines, including TextCNN, BERT, RoBERTa, ALBERT, DistilBERT, and T5, on the Biosensor Signal Analysis and Real-Time Health Monitoring Datasets. Our method achieves the best performance across all evaluation metrics, with a notable accuracy of 92.48% and 91.87% on the two datasets respectively, surpassing the strongest baseline (RoBERTa) by approximately 2.34% and 3.87%. The F1 score, recall, and AUC also reflect consistent gains, with improvements ranging from 1.8% to 3.5% over the second-best performing model. These results indicate that our model not only captures high-level semantic representations effectively but also retains critical temporal and multimodal dependencies that transformer-based models such as BERT and T5 tend to overlook in biosignal domains. While transformer-based architectures generally excel at textual representation learning, their direct application to continuous biomedical signals appears suboptimal due to lack of inductive bias for temporal signal patterns and noise-resilience, especially under conditions of signal artifacts and sensor drift. Additionally, lighter models such as DistilBERT suffer from significant performance drops, reinforcing that simple model compression leads to loss of nuanced physiological information necessary for reliable predictions in biosensor analysis.
Table 2. Comparison of ours with SOTA methods on biosensor signal analysis and real-time health monitoring datasets.
Table 3 further evaluates the same set of models on the Biomedical Sensor Data Collection and Deep Learning Biosensing Datasets. Once again, our approach consistently outperforms all SOTA models across every metric. Specifically, our model achieves an accuracy of 91.88% and 90.97%, reflecting a significant edge of over 2%–3% compared to the next-best method, T5. Our method’s superiority is particularly prominent in AUC scores, registering 93.01 and 91.60 respectively, which underscores our system’s strong capability in distinguishing subtle health-related classes under varying distributions and session-level variability. The robustness of our model arises from its integrated modality attention mechanism, which dynamically weighs each biosignal modality, allowing it to downplay noisy or uninformative channels. In contrast, traditional NLP-based methods such as BERT and RoBERTa lack explicit support for channel-wise feature fusion, limiting their flexibility when dealing with heterogeneous sensor inputs. Furthermore, our framework incorporates sequential modeling layers such as Bi-LSTM, which captures temporal evolution of physiological signals more effectively than position-encoded transformers that assume discrete token boundaries. This is crucial in scenarios such as reaction-based EMG or fatigue-related ECG patterns, where signal latency and transitional phases hold diagnostic importance. Therefore, while the performance gap may appear moderate in numeric terms, it represents substantial real-world gains in biosignal monitoring, where even minor improvements in recall or AUC can translate into reduced false alarms or improved early detection rates.
Table 3. Comparison of ours with SOTA methods on biomedical sensor data collection and deep learning biosensing datasets.
Across all four datasets, the performance margins suggest that our approach generalizes well across different biosensing scenarios—ranging from stress detection to long-term health monitoring. These gains can be attributed to several design factors: (1) the incorporation of temporal convolution and sequence modeling, which captures both local and global signal transitions; (2) the modality-specific attention mechanism, which enhances feature fusion across diverse signal types; and (3) extensive use of data augmentation, which regularizes the model and improves robustness to real-world noise. While models like T5 and RoBERTa show strong baseline performance due to their pretraining on large-scale corpora, their adaptation to biosignal domains is limited by their reliance on discrete token structures and lack of alignment with the continuous and often noisy nature of sensor data. Our model, being specifically tailored for the biosensing context, demonstrates the importance of designing architectures that are sensitive to both temporal and physiological signal characteristics. Moreover, the consistent superiority across both inter-subject (Real-Time Health Monitoring) and intra-session (Deep Learning Biosensing) datasets suggests that our method is both robust and scalable. This highlights its potential for real-world deployment in healthcare applications requiring high reliability and minimal latency, and further supports the utility of incorporating physiological priors into deep learning architectures for biosignal analysis.
4.4 Ablation study
To evaluate the contributions of the proposed components in our PhysioGraph Inference Network (PGIN), I conducted an ablation study across four datasets. The results are presented in Table 4 and 5. Three ablated variants were analyzed: (1) w/o Temporal Graph Convolution, which removes the temporal graph convolution layer responsible for capturing inter-sensor dependencies; (2) w/o Latent Variable Modeling, which excludes the latent variable modeling framework for temporal dynamics; and (3) w/o Uncertainty-Aware Decision Heads, which omits the uncertainty-aware decision mechanism. The complete model, referred to as Ours, includes all components and serves as the baseline for comparison.
Table 4. Ablation study results on ours model across biosensor signal analysis and real-time health monitoring datasets.
Table 5. Ablation study results on ours model across biomedical sensor data collection and deep learning biosensing datasets.
The removal of the temporal graph convolution layer (w/o Temporal Graph Convolution) resulted in the most significant performance degradation across all datasets. For example, on the Biosensor Signal Analysis Dataset, accuracy decreased from 92.48% to 89.76%, and the F1 score dropped by over 3%. This indicates the importance of modeling inter-sensor dependencies for accurate representation learning. Excluding the latent variable modeling framework (w/o Latent Variable Modeling) led to a moderate reduction in performance, particularly in recall and F1 scores on datasets with strong temporal dependencies, such as the Biomedical Sensor Data Collection. This highlights the role of latent dynamics in capturing temporal variations. The absence of the uncertainty-aware decision heads (w/o Uncertainty-Aware Decision Heads) showed a smaller but consistent impact, particularly in terms of reduced robustness and increased variance across multiple runs. This suggests that the uncertainty-aware mechanism enhances the model’s reliability under noisy and heterogeneous conditions.
These results validate the architectural choices in the PhysioGraph Inference Network. The temporal graph convolution layer effectively captures inter-sensor relationships, the latent variable modeling framework enhances temporal dynamics, and the uncertainty-aware decision heads improve robustness and generalization. The full model consistently achieves superior performance across all metrics and datasets, demonstrating the synergistic benefits of integrating these components.
4.5 Comparison with additional baseline systems
To ensure a fair and comprehensive evaluation of our proposed method, I conducted experiments comparing its performance with several conventional and state-of-the-art baseline models. The selected baselines include shallow neural networks (TextCNN), recurrent architectures (BiLSTM), and transformer-based models (BERT, RoBERTa, and T5). All models were trained under the same preprocessing pipeline, data splits, and evaluation metrics to ensure consistency. Hyperparameters for each baseline were either taken from published implementations or tuned using grid search to reflect optimal performance. Table 6 summarizes the results on two representative datasets: Biosensor Signal Analysis and Biomedical Sensor Data Collection. As shown, our proposed model consistently outperforms all baselines across multiple metrics, particularly in diagnostic accuracy and AUC. This reinforces the utility of incorporating physiological priors, temporal graph reasoning, and uncertainty-aware inference in biosensing diagnostics.
4.6 Robustness under noisy or missing inputs
To further validate the robustness of the proposed PGIN model, we conducted a series of controlled degradation experiments to simulate real-world sensor corruption. Specifically, two types of perturbations were introduced: (1) additive Gaussian noise with zero mean and 0.1 standard deviation applied to randomly selected sensor channels, and (2) random masking of 10%, 20%, and 30% of sensor channels during inference to mimic sensor dropout or transmission loss.
Table 7 presents the performance of PGIN under varying levels of corruption on the Biosensor Signal Analysis Dataset. Compared to RoBERTa and BiLSTM baselines, PGIN exhibits significantly more stable accuracy, F1 score, and AUC values across all corruption levels. This robustness is attributed to the model’s temporal graph regularization term and the masked dropout strategy employed during training, which jointly enforce smooth and plausible inter-sensor dependency representations. The results confirm that our model maintains interpretability and predictive reliability even under noisy or incomplete data conditions.
Table 7. Robustness evaluation under sensor noise and missing data (biosensor signal analysis dataset).
4.7 Clinician interpretability evaluation
In particular, we found that the latent trajectory visualizations provided a valuable temporal context for understanding patient states. For example, clinicians noted that sharp changes in the trajectory often coincided with labeled events such as onset of arrhythmia or stress episodes, allowing them to retrospectively validate the physiological plausibility of the model’s internal reasoning. Several participants commented that the shape and smoothness of the latent trajectory over time helped them distinguish between stable and deteriorating patient conditions. This suggests that the learned latent space is not only algorithmically informative, but also aligned with clinical intuitions. The ability to cross-reference latent states with observable biosignal dynamics enhances the transparency and trustworthiness of the diagnostic process. This supports the use of such trajectories as verifiable and interpretable representations in real-world medical practice.
To evaluate whether the interpretability of our model translates into clinically actionable insights, we conducted a human-in-the-loop usability study involving six board-certified clinicians from cardiology and critical care departments. Each participant was presented with 50 biosignal cases sampled from the Real-Time Health Monitoring Dataset and asked to perform diagnostic classification under two conditions: (1) using only the raw model predictions, and (2) using model predictions augmented with interpretability cues such as attention heatmaps, uncertainty scores, and latent physiological trajectories.
The results are summarized in Table 8. Clinicians achieved significantly higher diagnostic accuracy and expressed greater confidence and trust when interpretability outputs were available. Specifically, average diagnostic accuracy increased by 12.5%, while perceived trust in the AI system improved by over 40%. Additionally, decision-making time slightly decreased, indicating that interpretability support also improves efficiency. Qualitative feedback highlighted the value of attention maps and latent state visualizations in understanding anomalous biosignal patterns, especially in ambiguous cases. These findings indicate that the interpretability features of our framework go beyond analytical visualization and serve as meaningful tools that support clinical reasoning and foster trust in AI-assisted diagnostics.
Table 8. Clinician usability study comparing model-only vs. model + interpretability outputs (N = 6 clinicians).
5 Conclusions and future work
This work proposes an interpretable and uncertainty-aware deep learning framework for biosensing-based health diagnostics. By introducing the PhysioGraph Inference Network (PGIN), we enable dynamic modeling of physiological signal relationships using graph-based structures grounded in clinical priors. The Adaptive Health State Inference Mechanism (AHSIM) further enhances the framework’s decision-making process by integrating temporal memory, uncertainty quantification, and diagnostic granularity adjustment. Extensive evaluations across four diverse datasets demonstrate that our approach outperforms existing baselines in terms of accuracy, robustness, and explainability, making it suitable for deployment in both clinical and wearable settings.
Beyond algorithmic performance, the model’s interpretability components are designed to generate clinically meaningful outputs. Latent physiological trajectories produced by the model can be visualized and cross-referenced with biosignal events, offering diagnostic cues that align with expert knowledge. Attention mechanisms and modality-aware encoders allow clinicians to trace decision factors back to source signals, enabling hybrid human-AI workflows that support transparency and reduce diagnostic risk.
Despite these contributions, the framework has several limitations. The symbolic abstraction of latent states may not capture all nuances in atypical physiological conditions. Additionally, while the model is adaptable to different sensor types, further validation is needed across larger, more heterogeneous patient populations and devices. Future work will focus on integrating structured physiological knowledge graphs, improving edge deployment efficiency, and conducting human-in-the-loop studies to better assess clinical utility in real-world practice.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
ZY: Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Software, Validation, Formal analysis, Writing – original draft, Writing – review and editing. WH: Formal analysis, Investigation, Data curation, Methodology, Supervision, Project administration, Resources, Validation, Visualization, Writing – original draft, Writing – review and editing. HZ: Funding acquisition, Visualization, Supervision, Writing – original draft, Writing – review and editing.
Funding
The author(s) declared that financial support was not received for this work and/or its publication.
Conflict of interest
The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Ansari, M. Z., and Cho, C. (2008). “Design and analysis of a high sensitive microcantilever biosensor for biomedical applications,” in 2008 international conference on BioMedical engineering and informatics (IEEE), 2, 593–597. doi:10.1109/bmei.2008.308
Armghan, A., Logeshwaran, J., Sutharshan, S., Aliqab, K., Alsharari, M., and Patel, S. K. (2023). Design of biosensor for synchronized identification of diabetes using deep learning. Results Eng. 20, 101382. doi:10.1016/j.rineng.2023.101382
Asada, H. H., Shaltis, P., Reisner, A., Rhee, S., and Hutchinson, R. C. (2003). Mobile monitoring with wearable photoplethysmographic biosensors. IEEE Eng. Med. Biol. Magazine 22, 28–40. doi:10.1109/memb.2003.1213624
Bellagambi, F. G., Baraket, A., Longo, A., Vatteroni, M., Zine, N., Bausells, J., et al. (2017). Electrochemical biosensor platform for tnf-α cytokines detection in both artificial and human saliva: heart failure. Sensors Actuators B Chem. 251, 1026–1033. doi:10.1016/j.snb.2017.05.169
Boulogeorgos, A.-A. A., Trevlakis, S. E., Tegos, S. A., Papanikolaou, V. K., and Karagiannidis, G. K. (2020). Machine learning in nano-scale biomedical engineering. IEEE Trans. Mol. Biol. Multi-Scale Commun. 7, 10–39. doi:10.1109/tmbmc.2020.3035383
Bronzino, J. (2005). “Biomedical engineering: a historical perspective,” in Introduction to biomedical engineering (Elsevier), 1–29.
Chan, L. L., Cunningham, B. T., Li, P. Y., and Puff, D. (2006). A self-referencing method for microplate label-free photonic-crystal biosensors. IEEE Sensors J. 6, 1551–1556. doi:10.1109/jsen.2006.884547
Duarte, J. M., and Berton, L. (2023). A review of semi-supervised learning for text classification. Artif. Intell. Rev. 56, 1–69. doi:10.1007/s10462-023-10393-8
Esmaeili, F., Cassie, E., Nguyen, H. P. T., Plank, N. O., Unsworth, C. P., and Wang, A. (2023). Utilizing deep learning algorithms for signal processing in electrochemical biosensors: from data augmentation to detection and quantification of chemicals of interest. Bioengineering 10, 1348. doi:10.3390/bioengineering10121348
Evirgen, F., Özköse, F., Yavuz, M., and Özdemir, N. (2023). Real data-based optimal control strategies for assessing the impact of the omicron variant on heart attacks. AIMS Bioeng. 10, 218–239. doi:10.3934/bioeng.2023015
Evirgen, F., Uçar, S., and Özdemir, N. (2025). Mathematical analysis and optimal control of a caputo fractional diabetes system with parameter identification. J. Comput. Appl. Math. 117151. Available online at: https://www.sciencedirect.com/science/article/pii/S037704272500665X.
Geng, H., Pedersen, S. V., Ma, Y., Haghighi, T., Dai, H., Howes, P., et al. (2022). Noble metal nanoparticle biosensors: from fundamental studies toward point-of-care diagnostics. Accounts Chem. Res. 55, 593–604. doi:10.1021/acs.accounts.1c00598
Guo, Z., Tian, R., Xu, W., Yip, D., Radyk, M., Santos, F. B., et al. (2022). Highly accurate heart failure classification using carbon nanotube thin film biosensors and machine learning assisted data analysis. Biosens. Bioelectron. X 12, 100187. doi:10.1016/j.biosx.2022.100187
Heileman, K., Daoud, J., and Tabrizian, M. (2013). Dielectric spectroscopy as a viable biosensing tool for cell and tissue characterization and analysis. Biosens. Bioelectron. 49, 348–359. doi:10.1016/j.bios.2013.04.017
Ibrahim, F., Thio, T. H. G., Faisal, T., and Neuman, M. (2015). The application of biomedical engineering techniques to the diagnosis and management of tropical diseases: a review. Sensors 15, 6947–6995. doi:10.3390/s150306947
Jegan, R., and Anusuya, K. (2018). Biosensor-based feature extraction and physiological parameters measurement for biomedical applications. Int. J. Biomed. Eng. Technol. 28, 67–80. doi:10.1504/ijbet.2018.094430
Junfeng, L., Xiang, D., Xinglong, Y., and Dongsheng, W. (2005). Data analysis of surface plasmon resonance biosensor based on phase detection. Sensors Actuators B Chem. 108, 778–783. doi:10.1016/j.snb.2004.12.094
Kowsari, K., Meimandi, K., Heidarysafa, M., Mendu, S., Barnes, L. E., and Brown, D. E. (2019). Text classification algorithms: a survey. 10, 150. doi:10.3390/info10040150
Lee, S. H., Sung, J. H., and Park, T. H. (2012). Nanomaterial-based biosensor as an emerging tool for biomedical applications. Ann. Biomed. Eng. 40, 1384–1397. doi:10.1007/s10439-011-0457-4
Liu, L., and Du, K. (2024). A perspective on computer vision in biosensing. Biomicrofluidics 18, 011301. doi:10.1063/5.0185732
Mendelson, Y. (2005). “Biomedical sensors,” in Introduction to biomedical engineering (Elsevier), 505–548.
Min, S., Lewis, M., Hajishirzi, H., and Zettlemoyer, L. (2021). Noisy channel language model prompting for few-shot text classification. Annu. Meet. Assoc. Comput. Linguistics. Available online at: https://aclanthology.org/2022.acl-long.365/.
Minaee, S., Cambria, E., and Gao, J. (2020). Deep learning–based text classification. ACM Comput. Surv. Available online at: https://dl.acm.org/doi/abs/10.1145/3439726.
Mohanty, S. P., and Kougianos, E. (2006). Biosensors: a tutorial review. IEEE Potentials 25, 35–40. doi:10.1109/mp.2006.1649009
Nandakumar, V., Bishop, D., Alonas, E., LaBelle, J., Joshi, L., and Alford, T. L. (2010). A low-cost electrochemical biosensor for rapid bacterial detection. IEEE Sensors J. 11, 210–216. doi:10.1109/jsen.2010.2055847
Nguyen, H. H., Park, J., Kang, S., and Kim, M. (2015). Surface plasmon resonance: a versatile technique for biosensor applications. Sensors 15, 10481–10510. doi:10.3390/s150510481
Nisha, S. S., and Meeral, M. N. (2021). “Applications of deep learning in biomedical engineering,” in Handbook of deep learning in biomedical engineering (Elsevier), 245–270.
Pantelopoulos, A., and Bourbakis, N. (2008). “A survey on wearable biosensor systems for health monitoring,” in 2008 30th annual international conference of the IEEE engineering in medicine and biology society (IEEE), 4887–4890.
Perumal, V., and Hashim, U. (2014). Advances in biosensors: principle, architecture and applications. J. Applied Biomedicine 12, 1–15. doi:10.1016/j.jab.2013.02.001
Qureshi, R., Irfan, M., Ali, H., Khan, A., Nittala, A. S., Ali, S., et al. (2023). Artificial intelligence and biosensors in healthcare and its clinical relevance: a review. IEEE Access 11, 61600–61620. doi:10.1109/access.2023.3285596
Raji, H., Tayyab, M., Sui, J., Mahmoodi, S. R., and Javanmard, M. (2022). Biosensors and machine learning for enhanced detection, stratification, and classification of cells: a review. Biomed. Microdevices 24, 26. doi:10.1007/s10544-022-00627-x
Rong, G., Xu, Y., and Sawan, M. (2023). Machine learning techniques for effective pathogen detection based on resonant biosensors. Biosensors 13, 860. doi:10.3390/bios13090860
Rufai, M. A., and Ramos, H. (2020). Numerical solution of bratu’s and related problems using a third derivative hybrid block method. Comput. Appl. Math. 39, 322. doi:10.1007/s40314-020-01372-8
Sadak, O., Sadak, F., Yildirim, O., Iverson, N. M., Qureshi, R., Talo, M., et al. (2022). Electrochemical biosensing and deep learning-based approaches in the diagnosis of covid-19: a review. IEEE Access 10, 98633–98648. doi:10.1109/access.2022.3207207
Schackart III, K. E., and Yoon, J.-Y. (2021). Machine learning enhances the performance of bioreceptor-free biosensors. Sensors 21, 5519. doi:10.3390/s21165519
Shams, M. (2024). On a stable multiplicative calculus-based hybrid parallel scheme for nonlinear equations. Mathematics 12, 3501. doi:10.3390/math12223501
Shams, M., Rafiq, N., Kausar, N., Agarwal, P., Park, C., and Mir, N. A. (2021). On highly efficient derivative-free family of numerical methods for solving polynomial equation simultaneously. Adv. Differ. Equations 2021, 465. doi:10.1186/s13662-021-03616-1
Shams, M., Kausar, N., Araci, S., and Oros, G. I. (2024). Artificial hybrid neural network-based simultaneous scheme for solving nonlinear equations: applications in engineering. Alexandria Eng. J. 108, 292–305. doi:10.1016/j.aej.2024.07.078
Soni, S., Chouhan, S., and Rathore, S. (2022). Textconvonet: a convolutional neural network based architecture for text classification. Appl. Intelligence Bost. 53, 14249–14268. doi:10.1007/s10489-022-04221-9
Sutar, P., Pethe, A., Kumar, P., Tripathi, D., and Maity, D. (2025). Hydrogel innovations in biosensing: a new frontier for pancreatitis diagnostics. Bioengineering 12, 254. doi:10.3390/bioengineering12030254
Suthar, J., Alvarez-Fernandez, A., Osarfo-Mensah, E., Angioletti-Uberti, S., Williams, G. R., and Guldin, S. (2023). Amplified eqcm-d detection of extracellular vesicles using 2d gold nanostructured arrays fabricated by block copolymer self-assembly. Nanoscale Horizons 8, 460–472. doi:10.1039/d2nh00424k
Tittl, A., John-Herpin, A., Leitis, A., Arvelo, E. R., and Altug, H. (2019). Metasurface-based molecular biosensing aided by artificial intelligence. Angew. Chem. Int. Ed. 58, 14810–14822. doi:10.1002/anie.201901443
Vashistha, R., Dangi, A. K., Kumar, A., Chhabra, D., and Shukla, P. (2018). Futuristic biosensors for cardiac health care: an artificial intelligence approach. 3 Biotech. 8, 358. doi:10.1007/s13205-018-1368-y
Wang, J., and Katz, E. (2010). Digital biosensors with built-in logic for biomedical applications—biosensors based on a biocomputing concept. Anal. Bioanalytical Chemistry 398, 1591–1603. doi:10.1007/s00216-010-3746-0
Wang, G., Li, C., Wang, W., Zhang, Y., Shen, D., Zhang, X., et al. (2018). Joint embedding of words and labels for text classification. Annual Meeting of the Association for Computational Linguistics. Available online at: https://arxiv.org/abs/1805.04174.
Keywords: explainable deep learning, PhysioGraph inference network (PGIN), adaptive health state inference mechanism (AHSIM), diagnostic accuracy, uncertainty estimation
Citation: Yang Z, Huang W and Zhang H (2026) An explainable deep learning framework for biosensing data interpretation in biomedical engineering and real-time health diagnostics. Front. Bioeng. Biotechnol. 13:1688586. doi: 10.3389/fbioe.2025.1688586
Received: 19 August 2025; Accepted: 15 December 2025;
Published: 12 February 2026.
Edited by:
Xiaolin Huang, Nanchang University, ChinaCopyright © 2026 Yang, Huang and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Zheng Yang, cHVjaDI4MzdAb3V0bG9vay5jb20=
Zheng Yang1,2*