- Department of Computer Science, Carlos III University of Madrid, Madrid, Spain
Introduction: Electrocardiograms (ECGs) have long served as the standard method for cardiac monitoring. While ECGs are highly accurate and widely validated, they require direct skin contact, are sensitive to motion artifacts, and are not always practical for continuous or unobtrusive monitoring, limiting their generalization to real-world, dynamic environments. However, radar-based technologies offer a novel, non-invasive alternative for acquiring cardiac signals without direct contact. This improves both hygiene and patient comfort, making it especially attractive for medical applications. Despite these benefits, it may raise privacy concerns, inadvertently revealing personal attributes such as sex and age. This study investigates, for the first time, how such demographic information can be inferred from radar-acquired cardiac signals.
Methods: To address this question, we developed a machine learning framework to predict demographic attributes from radar-based cardiac signals. These signals were transformed into scalograms—a time-frequency representation—and then classified using a Convolutional Neural Network (CNN). Given the lack of prior studies on demographic inference from radar-based cardiac signals, the generalization capabilities of existing approaches remain untested in this context. Moreover, the small size of available datasets further limits model performance. To mitigate these issues, we applied data augmentation using a Conditional Wasserstein Generative Adversarial Network (cWGAN), which generated synthetic scalograms conditioned on class labels. Notably, there are very few prior studies on data augmentation specifically for this type of signal. This strategy aimed to enhance model accuracy and generalization by enriching the training data.
Results: Our experiments demonstrate that data augmentation significantly improves model performance. The trained model achieved an accuracy of 78.40% in predicting the sex of individuals and 72.83% accuracy in classifying them into two age groups (18–29 and 30–65 years), despite the dataset being limited to only 30 subjects.
Discussion: These findings reveal a potential privacy risk associated with radar-based biometric systems. The ability to infer sensitive demographic information from physiological signals could have serious implications, particularly in secure applications such as electronic passports (e-passports), where access to RFID chip data often depends on such personal attributes. Therefore, while radar technologies offer promising advantages, their deployment must consider and address the associated privacy challenges.
1 Introduction
The significance of biosignals in contemporary society, particularly within the healthcare system, has grown steadily in recent years. This trend is driven by advancements in technology, including the development of sophisticated sensors and wearable devices or data capture through the Internet of Things. These innovations allow for continuous and real-time monitoring of patients, leading to early detection of anomalies and more effective management of chronic diseases. Furthermore, the application of artificial intelligence (AI) and machine learning algorithms to biosignal data enhances the accuracy of diagnoses and prognoses, paving the way for personalized medicine. Beyond traditional medical applications, biosignals are also being utilized in emerging fields such as human-computer interaction, neuroprosthetics, mental health monitoring, and security. The ability to interpret biosignals accurately is transforming how healthcare providers approach patient care, emphasizing preventive measures and improving overall health outcomes.
Biosignals play a crucial role in various contemporary applications. Among the most common, electrocardiograms (ECGs) record the heart’s electrical activity, essential for diagnosing cardiac diseases. Electroencephalograms (EEGs) capture brain activity, aiding in neurological diagnoses. Electromyograms (EMGs) measure electrical impulses in muscles, assessing muscle function and diagnosing neuromuscular conditions. Other biosignals, such as galvanic skin responses (GSRs), measure changes in skin conductance due to emotional or physiological arousal. Respiratory signals, along with blood pressure and oxygen saturation (SpO2) measurements, are fundamental in monitoring cardiovascular and respiratory health [1].
Within these signals, cardiac monitoring is essential in modern healthcare. Specifically ECGs are instrumental in diagnosing various cardiac conditions, including arrhythmias, myocardial infarctions, and heart rhythm abnormalities. By capturing the depolarization and repolarization of cardiac muscle cells, ECGs provide critical insights into heart function and health. In clinical practice, ECGs are obtained through electrodes placed on the skin, typically on the chest, limbs, and sometimes the torso. These electrodes detect the tiny electrical impulses generated by the heart with each heartbeat. The resulting waveform reflects different phases of the cardiac cycle and abnormalities in these patterns can indicate specific cardiac disorders. Beyond diagnosis, ECGs are also widely used in research to study cardiac function under various conditions and interventions. Additionally, they have proven valuable in fields such as emotion recognition [2], and in security applications, where they serve as a potential biometric identifier due to their unique characteristics and the possibility of continuous authentication without requiring user cooperation [3, 4]. Additionally, research in fields such as human-computer interaction and wearable technology uses cardiac signals to improve user experience, monitor stress levels, and optimize performance.
However, obtaining an ECG always requires physical contact, which presents several inherent limitations. Firstly, contact-based methods such as ECG require specialized equipment for signal acquisition, which can be cumbersome and restrict accessibility and portability. Moreover, direct physical contact with the user may raise concerns regarding comfort, hygiene, and potential skin issues, especially in sensitive populations like premature babies undergoing continuous monitoring [5]. Furthermore, these devices are susceptible to movement artifacts, which can compromise measurement accuracy, particularly in dynamic environments. In contrast, contactless methods using radar technology provide an interesting alternative. Doppler radar operates by emitting electromagnetic waves towards the chest area, where they interact with the surface of the skin and underlying tissues. As the heart beats, it causes subtle movements in the chest surface. These movements alter the frequency of the reflected waves due to the Doppler effect, where the frequency shift is directly related to the velocity of the chest movements induced by the heartbeat. By analyzing these frequency shifts over time, Doppler radar can extract detailed information about cardiac activity, including heart rate and rhythm enabling non-intrusive cardiac motion detection and continuous monitoring without the need for physical contact [6]. This approach enhances user comfort and convenience; however, it currently faces some limitations. One of them is the susceptibility of radar signals to noise and interference from various environmental sources, which can compromise the reliability of measurements. As a result, achieving clinical-grade accuracy remains a ongoing challenge, particularly in real-world scenarios where uncontrolled environmental factors and patient mobility can affect the precision of radar-based health monitoring. Furthermore, there are significant challenges in radar signal processing and the application of machine learning, especially with advanced deep learning techniques. Researchers are focused on identifying the most effective methods for extracting vital information from radar data and analyzing it robustly, ensuring functionality across a wide range of situations [7].
However, embedded within these previously described biosignals are implicit personal details, which may inadvertently disclose sensitive information about individuals like sex or age. While this data are essential for medical insights, their unintended disclosure poses significant privacy risks. Unauthorized access or misuse of such information could lead to discrimination, compromised personal privacy, or even targeted exploitation in various contexts [8], highlighting the importance of robust data protection measures in biosignal processing and storage. According to major privacy regulations such as the General Data Protection Regulation (GDPR) [9] in Europe and the California Consumer Privacy Act (CCPA) [10] in the United States, personal information includes not only direct identifiers but also data that can be reasonably linked to an individual, including inferred attributes used for profiling. There are also concrete security implications: for example, sex and age are used in deriving access keys for RFID chips embedded in e-passports, so their unauthorized inference via remote sensing could reduce key entropy and facilitate brute-force attacks [11]. Furthermore, such inferred data can be exploited to segment individuals into demographic or consumer groups, potentially enabling manipulative advertising, discriminatory pricing, or other unauthorized profiling practices, as specified in Section (K) of the CCPA. The potential leakage of personal information, alongside other medically relevant data, has been explored in several biometric modalities, including ECG [12], EEG [13], photoplethysmography (PPG) [14, 15], and gait signatures [16]. These studies demonstrate the unintended disclosure of sensitive attributes, such as age and sex. However, to the best of our knowledge, no comprehensive investigation has yet addressed this concern for cardiac signals derived from radar technology, representing a significant gap in the study of this type of signal.
The mechanism behind this potential privacy leakage lies in the inherent encoding of demographic information within the morphological and spectral features of cardiac signals. For instance, differences in chest anatomy, hormonal influences on heart rate variability, or vascular compliance associated with aging can subtly alter the frequency components and temporal structure of the radar-extracted signal. When processed using advanced machine learning models, these patterns can be exploited to infer attributes such as sex and age—even if such information was not explicitly targeted. This raises significant privacy concerns, especially in security-sensitive environments such as biometric authentication systems or electronic passports. Technically, radar-based monitoring still faces challenges such as noise susceptibility, difficulty in isolating the cardiac component from respiration and motion artifacts, and a lack of large, labeled datasets—factors that complicate the generalization and reliability of deep learning models trained on this data. Furthermore, traditional signal processing approaches often fall short in capturing the complex time-frequency dynamics needed to robustly infer demographic features, motivating the need for deep learning and data augmentation solutions. These challenges and opportunities motivate the present study, which aims to systematically investigate the presence and implications of such demographic leakage.
The objective of this article is to investigate whether this information (sex and age) is implicitly encoded in cardiac signals extracted via radar. In other words, our main question is: Is there a leakage of age and sex, along with other medical attributes, in the heart signal extracted from radar? To this end, we aim to estimate both variables from this type of signal. Understanding this potential correlation could provide insights into novel applications and implications for biometric identification and motivate the exploration and development of different security methods that could improve authentication and privacy protection.
The key contributions of this paper are as follows:
1. We present the first study that demonstrates the leakage of demographic information (sex and age) from radar-acquired cardiac signals. While prior works have explored this in ECG and PPG signals, no such analysis has been conducted using radar technology. Our study addresses this gap and provides empirical evidence supporting the presence of demographic markers in radar signals.
2. We propose a novel data augmentation approach based on Conditional Wasserstein GANs (cWGAN), specifically tailored to scalograms of radar-extracted heartbeat signals. Unlike previous methods that rely on ECG as a generative input, our approach directly synthesizes radar-based scalograms conditioned on class labels, enabling better balance and generalization in small datasets.
3. Our experiments show that the use of cWGAN-generated scalograms significantly improves classification performance. For sex classification, accuracy increases from 72.80% to 78.40% (+5.6%), and for age group classification from 65.60% to 72.83% (+7.23%). Frame-level accuracy, FAR, FRR, precision, recall, and F1-score all show consistent improvements with data augmentation.
4. We ensure robust evaluation using leave-one-subject-out cross-validation, preventing data leakage and simulating realistic scenarios where the model must generalize to unseen individuals.
5. We highlight the privacy implications of our findings, particularly in biometric applications where demographic leakage from cardiac signals could compromise personal security.
The rest of this article is structured as follows. The Related Work section reviews the existing scientific literature on the use of radar-based heart signals, age and sex estimation through biosignals, and data augmentation in this field. The Materials and Methods section provides detailed information about the dataset, preprocessing steps, and the methodology used for age group and sex estimation with radar data. It also explains the data augmentation techniques employed to enhance the training dataset. The Results section presents the study’s findings, evaluating the performance of the proposed methods using various metrics. Finally, the article concludes with a discussion of the results and their implications for future research.
2 Related work
2.1 Radar-based heart signal applications
Within the field of cardiac monitoring, one of the most widely utilized signals is the ECG, which finds application across various domains such as clinical monitoring [17–19], diagnosis of cardiac conditions [20–23], sports medicine [24], stress monitoring [25], emotion recognition [2, 26] or biometric identification [27, 28].
Non-invasive health monitoring solutions, such as radar-based systems, are emerging as alternatives to traditional methods like PPG and ECG, which generally require contact-based sensors [29]. These non-contact radar technologies allow for continuous and unobtrusive physiological monitoring, offering a comfortable solution for patients in both clinical and everyday environments. With applications ranging from vital sign detection [30] and respiratory monitoring [31] to blood pressure estimation [32] and cardiac monitoring [33], radar systems represent a significant advancement in the context of non-invasive health technology. However, despite the advancements in non-contact monitoring, research on radar-based cardiac signal analysis remains limited. A promising approach within this field is the use of Microwave Doppler sensors, which enable the extraction of heartbeat and individual feature quantities through time-frequency analysis without requiring direct skin contact. These sensors are sensitive enough to detect minute vibrations on the chest surface caused by heartbeats after isolating these vibrations from other movements such as respiration or artifacts. To address this issue, some studies [34, 35] employ Butterworth filters to attenuate lower frequencies associated with respiration. Conversely, [36] proposes Wavelet Packet Decomposition (WPD) as an alternative method, achieving less than 2% error for respiration and 3.5% for heart rate, thereby enhancing the accuracy of vital sign detection compared to Bandpass filters and Peak Detection methods. Furthermore, in [37], various techniques were assessed to identify the most effective approach for extracting cardiac signals from radar data, with Discrete Wavelet Transform (DWT) demonstrating superior performance, even surpassing WPD.
Many studies that work with this type of radar-obtained signal use Machine Learning techniques to extract information [38]. Applications can be quite diverse, such as estimating the patient’s heart rate [39–43] or heart rate variability [44–46], with some studies also applying this process to animals [47–49]. In [50] they use commodity Wi-Fi devices to capture the heart signal of subjects, and in [51] the heart signal of different subjects who are simultaneously in the same environment is captured using radar. In addition to detecting heart rate, work has also been done on the detection or monitoring of breath rate, as in [52] or more generally reviewed in [53]. Another field that has received much attention is the monitoring of both heartbeats [54–58] and vital signs [59–61]. The possible applications of these types of signals in security have also begun to be investigated, for example, with the aim of authenticating people in a convenient, contactless, and continuous manner although the results are still scarce. In [62], they create an authentication method based on the fiduciary analysis of the heart signal (which involves using specific reference points in the signal as references). In [63], they perform authentication by separating the signal into heartbeats and classifying each one, while in [64], they transform the signal into a scalogram to then extract features and classify using Deep Convolutional Neural Networks (DCNNs), achieving an accuracy of 98.5% in a sample composed of only 4 people. In [65], they attempt identification in a set of 20 subjects using the spectral distribution of the signal and obtain an Error Rate (EER) of 3.48%. On the other hand, the use of these types of signals for other purposes has been explored, such as identification from respiratory signals [66], gait recognition [67], event recognition [68], emotion recognition [69], and multiperson spatial tracking [70]. The main limitation in the development of this area, however, is the lack of datasets that collect this type of signal. Currently, those that exist do not usually have a large number of subjects, which sometimes prevents presenting more solid results.
2.2 Age and sex estimation
The field where estimation of age and sex has been most extensively researched is through facial images [71]. However, there is also notable work using other signals such as ECG [12].
Various methods have been employed to estimate sex from ECG, with the use of CNNs being among the most frequent approaches. This is demonstrated in [72] achieving 90.4% accuracy. In [73], clustering of heartbeats into 10 groups followed by training separate classifiers for each group aimed to maximize performance, achieving 94.4% accuracy. On the other hand, estimated age from ECG reflects the overall health status of the cardiovascular system. When significantly above chronological age, it may indicate the presence of abnormalities or disorders [12]. In [74], they attempt to predict sex and age group (young/old) from ECG using CNNs, achieving accuracies of 86.82% for sex and 82.97% for age group. Most current studies utilize Deep Learning to estimate cardiac age, employing CNNs with ECG signals as network inputs, as demonstrated in [75] or in [72] where an average error of 6.9 5.6 years in age estimation was reported. In [76], a neural network regression model was used to predict chronological age from ECG with the same mean error of 6.9 5.6 years, while in [77], a Deep Learning model achieved a mean error of 6.899 years. The advantage of estimating age and sex from ECG lies in the availability and size of datasets, sometimes exceeding 700,000 patients as seen in [72].
However, when it comes to estimating age and sex from radar-extracted cardiac signals, there is limited research. [78] estimates sex based on the act of transitioning from sitting to standing, and [79] estimates it using gait features, both studies using radar to capture body movement rather than heartbeats. Similarly, [80] uses radar to assess body movement in elderly individuals to estimate their age and fall risk. In [81], authors do study sex differences in relation to radar-detected cardiac signals, finding that the reference signal-to-noise ratio is significantly higher in males than females, possibly due to chest anatomy differences, making it somewhat more challenging to detect cardiac signals with this technique in females compared to males. Therefore, to the best of our knowledge, no prior studies have attempted to estimate sex or age from radar-detected cardiac signals.
2.3 Data augmentation
Data augmentation is a technique used to improve the performance of machine learning and deep learning models by artificially expanding the size and diversity of training datasets. In the context of biosignals, this can be particularly beneficial as datasets are often limited due to challenges in collecting large amounts of certain types of medical data. However, given the wide variability among classes of these data, data augmentation techniques can vary significantly depending on the biosignal being studied.
When working with ECG data, studies often employ data augmentation techniques that rely on signal transformations or a combination thereof, rather than utilizing deep learning methods. For instance, in [82], researchers develop an algorithm that optimizes the order and parameters of these transformations. These methods include time masking, which zeros out parts of the signal, SpecAugment, which masks components in both the time and frequency domains of a signal transformed using Short-Time Fourier Transform (STFT), Discriminative Guided Warping, using Dynamic Time Warping to align a source ECG with a representative reference signal distinct from other classes, and SMOTE, an oversampling strategy that generates synthetic examples of the minority class by interpolating minority samples to address class imbalance in ECG prediction problems. Similarly, studies like [83–85] employ a combination of transformations to augment the training set, whereas [86] augments heartbeats specifically to mitigate class imbalance within their dataset.
A common scenario when applying data augmentation techniques is to work with images as inputs, for which there are multiple strategies to increase the training data, among which the use of GANs is one of the main ones [87]. This situation is frequently encountered when working with EEG, where different data augmentation techniques can yield good results [88]. In [89] new samples are obtained after training a GAN to regenerate a part of the image that has been erased, demonstrating that this method can improve emotion recognition. In [90], they use a similar procedure but employing a Conditional Wasserstein GAN (cWGAN), also improving the results in emotion classification, just like in [91], where they use a multiple generator cWGAN for the same purpose. In [92], a conditional Boundary Equilibrium GAN is employed to generate new samples of EEG data, eye movement data, and their direct concatenations. Meanwhile, in [93], GANs are utilized to augment one-dimensional data. On the other hand, [94] demonstrates the use of a Wasserstein GAN to enhance Ground-penetrating radar data, achieving superior results compared to auxiliary classifier GANs and other traditional methods. Finally, in [95], a Wasserstein GAN with confidence loss is applied to augment a dataset of plant images. Similarly, [96] presents an alternative strategy for augmenting ECG signals based on segmentation and rearrangement of time-series data. This method, although not based on deep generative models, allows the generation of realistic and well-structured synthetic signals with minimal computational cost. Another recent approach [97] focuses on emotion recognition from ECG signals using a unimodal system. Given the scarcity of labeled affective ECG datasets, the authors propose a multi-filtering augmentation method, which generates new data by applying diverse filtering techniques to enhance signal quality showing a significant performance boost.
Wasserstein GANs (WGANs) are a type of GAN that utilize the Wasserstein distance, also known as Earth Mover’s Distance (EMD), as a loss function to improve the training stability and performance of GANs [98]. Traditional GANs often suffer from issues such as mode collapse and vanishing gradients, which can make training difficult and unstable. The Wasserstein distance provides a smoother and more meaningful measure of the difference between the generated data distribution and the real data distribution, helping to mitigate these issues. Gradient penalty and spectral norm are two enhancements used to enforce the Lipschitz constraint in WGANs, further improving training stability. Gradient penalty adds a penalty term to the loss function to constrain the gradient norm of the critic’s output with respect to its input to be close to 1. This is done by interpolating between real and generated data points and penalizing deviations from this gradient norm, with the penalty term typically calculated as , where are interpolated samples, is the critic’s output, and is a coefficient determining the penalty strength. Spectral norm, on the other hand, normalizes the weight matrices of the neural network layers by their spectral norm, which is the largest singular value of the weight matrix. By constraining the spectral norm to be at most 1, the Lipschitz continuity of the critic function is ensured without the need for weight clipping or gradient penalties, making training more stable and reducing the risk of exploding or vanishing gradients.
An example of this technique is used in the study [99], where researchers working with accelerometer data for human activity recognition demonstrated that augmenting training data using cWGAN is more effective than using a conditional GAN. Regarding data augmentation for heart signals obtained through radar technology, we found only two studies. In [100] they use an attention-based WGAN with Gradient penalty to generate new I/Q signals, feeding the generator with ECGs. The objective is to classify heart signals into two groups: low heart rate (less than 60 beats per minute) or high. In [101], the authors utilize the decomposition provided by the Discrete Wavelet Transform to augment the data. However, they train the model with the newly created samples and test it on the original samples, which may lead to potential data leakage.
In summary, while previous research has shown the feasibility of using radar signals for vital sign detection and even biometric tasks, there is a lack of studies focused on privacy risks—specifically the unintended leakage of demographic information such as age and sex. Moreover, existing works often rely on small datasets and do not leverage advanced data augmentation techniques tailored to this signal type. Most GAN-based approaches in biosignal processing either require complementary signals (e.g., ECG) or are not designed for conditional generation of demographic categories. These gaps motivate our proposed method, which directly addresses the challenges of limited data, signal ambiguity, and demographic inference through an end-to-end deep learning and data augmentation pipeline.
3 Materials and methods
This section describes the methods employed throughout the study. A graphical summary of the workflow for this process is presented in Figure 1. The first step after capturing the I/Q quadrature signals was to divide them into 10 s windows. Subsequently, we performed preprocessing on each window, as detailed in Section 3.2, to extract a signal corresponding to thoracic movement. From this signal, we isolated the component related to the heartbeat using the Maximal Overlap Discrete Wavelet Transform (MODWT), eliminating the influence of respiration and other noise. Next, we subdivided the 10 s heartbeat signal into overlapping frames of 4 s each and generated a scalogram for each frame. These scalograms, along with additional scalograms artificially generated using a cWGAN, were then used to train a CNN that classified each scalogram into male or female categories or age groups. Finally, to derive a single prediction for each 10 s window, we conducted a voting process based on the predictions provided by the CNN for each one of its frames.
Figure 1. Flowchart for the age group and sex estimation process. Created using Canva, licensed under Free Content License.
3.1 Data
The study utilized a publicly available dataset from [102], collected at University Hospital Erlangen, Germany, involving 30 healthy participants (14 males, 16 females, with an average age of 30.7 years). A radar system, optimized with a focal point at approximately 40 cm, employed Six-Port technology to capture chest movements in a contactless manner. The system featured a bi-static antenna setup with transmitting and receiving antennas set at angles and a laser for alignment purposes. The dataset included recordings from five distinct scenarios: a resting scenario, Valsalva maneuver, apnea, tilt-up, and tilt-down. In this study, we only used recordings from the resting scenario, where participants reclined for at least 10 min in a relaxed position with calm breathing. The data consisted of In-Phase and Quadrature (I/Q) signals sampled at 2,000 Hz from the 30 different recordings, each approximately 10 min long. Figures 2a and 2b show the distribution of subjects by sex and age. Due to the highly imbalanced age data, we divided ages into two groups: 18 to 29 years and 30 to 65 years. The dataset is accessible at [103].
Figure 2. Distribution of subjects by sex and age. (a) Distribution by Sex. (b) Distribution by Age.
Although the dataset used in this study comprised only 30 individuals, this sample size was comparatively large within the current state of radar-based cardiac signal research. As reviewed in [104], most datasets incorporating radar signals—both public and private—were significantly smaller, with a median of just 12 subjects and an interquartile range of 5 to 24. Among the few public datasets available (only 10 in total), only one included more subjects than ours; however, it contained only children under the age of 13, making it unsuitable for our study. Therefore, the field of radar-based cardiac signal analysis remains in an early stage, and datasets of a size comparable to those available for other signals—such as ECG or PPG—are not yet available. Consequently, our use of 30 adult participants represents a substantially above-average sample size in this emerging research domain and marks a meaningful step forward in demonstrating the feasibility of demographic attribute estimation from radar signals at a scale larger than typically seen in previous work. Furthermore, we ensured strict subject-wise separation between training and testing data to avoid data leakage and to enhance the generalizability of our results. It is also reasonable to assume that, with larger datasets encompassing a broader diversity of subjects and recording conditions, the model’s generalization capability would further improve, potentially leading to even higher classification performance. This design strengthens the validity of our findings and provides a valuable contribution to a field still constrained by limited data availability.
3.2 Preprocessing
The raw cardiac signals used as input to our models consisted of I/Q components. These complex-valued signals were recorded at a sampling frequency of 2,000 Hz and captured subtle chest movements in a contactless manner. Using these I/Q signals provided by the dataset, the initial step involved segmenting them into non-overlapping 10 s windows, a window size that aligns with common practices in the state-of-the-art [105]. Subsequently, ellipse fitting was applied to the I/Q data points within each window, following the methodology outlined in [106]. This process enables arc tangent demodulation using the parameters derived from the fitted ellipses [107]. The outcome of this procedure represents the movement observed in the region of interest (the thorax), which encompasses various components, notably including the patient’s movements, respiratory patterns, and cardiac motion. To isolate the cardiac motion signal, we employed the Maximal Overlap Discrete Wavelet Transform (MODWT). MODWT is particularly effective in this context due to its ability to perform multiresolution analysis without the downsampling step present in classical discrete wavelet transforms (DWT). This property ensures translation invariance, which is essential when time alignment of features such as heartbeats must be preserved across frames. Radar signals from the thorax are highly non-stationary and contain overlapping frequency components from respiration, subtle body movements, and cardiac activity. MODWT offers a data-driven decomposition that adaptively separates these sources by their dominant frequency ranges, outperforming traditional bandpass filters. In [37], this approach has been evaluated against alternative methods, demonstrating superior performance. Subsequently, the cardiac motion signal within each window was segmented into overlapping 4 s frames with a stride of 0.5 s for classification, similar to the methodology outlined in [63] for heartbeat detection. Each frame was then transformed into its scalogram, represented as a black-and-white image with dimensions of pixels. The CNN produces predictions at the frame level. To obtain the final decision for each 10 s window, frame-level probabilities are aggregated into a single probability distribution, from which the window-level prediction is derived. This voting scheme ensures more robust and stable classifications by combining evidence from multiple frames.
3.3 Data augmentation
In accordance with findings from other studies [108–110], we performed data augmentation using a GAN [111], specifically a cWGAN. This kind of WGAN incorporates conditional information: both the generator and the discriminator (or critic) receive additional input, typically in the form of class labels or other conditioning variables. The architectures of our generator and critic can be seen in Tables 1, 2 respectively. We used Adam as the optimizer for the generator, and RMSprop, as recommended in [98], for the critic. Additionally, we implemented gradient penalty with a coefficient of 10 to stabilize the training process, which helps enforce the Lipschitz constraint by ensuring the gradients have a norm close to 1. For both tasks, predicting sex and age respectively, each data sample was assigned a single label.
Table 1. Architecture summary of the generator in the cWGAN with 2 labels and 100 as input dimension.
The cWGAN addressed the challenge of limited and imbalanced radar datasets by synthetically generating realistic scalograms conditioned on class labels (e.g., sex or age group). By modeling the data distribution and introducing controlled variability, the cWGAN enriched the training set with representative samples that improved generalization and reduced overfitting. This approach is particularly effective when collecting more real radar data is impractical due to time, cost, or ethical constraints. By training our cWGAN on the scalograms of patients dedicated to the training set, we were able to generate new scalograms labeled in each of the two categories: male and female in the case of sex, and 18–29 and 30–65 in the case of age. We chose this strategy instead of oversampling or undersampling to preserve the original distribution and avoid overfitting. Although other methods such as SMOTE or class weighting were considered, the cWGAN was selected for its ability to generate label-consistent and structurally realistic samples.
Our cWGAN was trained for 151 epochs with a batch size of 32. Early stopping was not used, but both generator and critic losses were monitored throughout training. The generator used a learning rate of , while the critic (or discriminator) used a smaller learning rate of . For every generator update, the critic was updated five times. The latent space dimension was set to 100. We conditioned the generation on two labels (e.g., male/female or age group), which were embedded and concatenated to both the generator input and the critic input. The training was carried out on scalograms corresponding to the training set, and a gradient penalty with coefficient 10 was applied to enforce the Lipschitz constraint and promote stable convergence.
3.4 Classification
We utilized a CNN for the classification task, as these models are frequently used in computer vision tasks, particularly as feature extractors from images across various fields [112, 113]. They concatenate hierarchical layers of convolutional filters to automatically learn meaningful representations from raw pixel data. These learned features capture hierarchical patterns such as edges, textures, and shapes, which are essential for understanding and classifying visual content. This feature extraction process from the scalograms is carried out in the convolutional layers of the CNN, while the fully connected layers placed at the end serve to classify the scalograms based on the previously obtained features.
The structure of our CNN is shown in Figure 3 and consisted of 5 convolutional layers followed by 4 fully connected layers. The output was a number corresponding to the predicted class for the input scalogram. After each convolutional layer, we applied batch normalization to normalize the input and mitigate internal covariate shift [114]. The Rectified Linear Unit (ReLU) was used as the activation function to introduce non-linearity, which aids in learning complex patterns in the data. Max pooling was applied after each convolutional block to progressively reduce spatial dimensions and retain the most relevant features. The kernel sizes for the five convolutional layers were 5, 3, 13, 5, and 2 respectively, all with stride 1 and no padding. The input to the network consisted of 3-channel scalograms with fixed dimensions, and the size of the flattened feature map after the final convolutional block was 2048. The fully connected (MLP) part of the network contained layers of size 1024, 256, 64, and 1, with ReLU activations and dropout (p = 0.5) applied after each layer except the final output. The model was trained using the Adam optimizer with a learning rate of 0.001 and binary cross-entropy loss. With this model, we were able to predict the class of each scalogram, and thus of each frame. To obtain a prediction for each window, we then used hard voting based on the predictions of all the frames within that window. Nevertheless, we did not conduct an optimization process for the selected CNN architecture; therefore, alternative architectures might have yielded superior final results.
3.5 Privacy threat model
Radar-based cardiac sensing enables unobtrusive acquisition of physiological signals, often without requiring physical contact or active participation from the subject. While typically used for clinical monitoring or biometric authentication, this modality also allows for the potential inference of sensitive demographic attributes from the collected signals.
In our threat model, we considered a passive adversary who can covertly acquire radar reflections from a subject without requiring physical access to the sensing system. The adversary may place a radar device in the environment (e.g., in smart home devices, hospital rooms, or public kiosks) to continuously collect cardiac signals from individuals situated within sensing range. Data acquisition can occur in everyday scenarios where individuals are seated or stationary, and does not require user awareness or consent. The adversary is assumed to have access to trained deep learning models capable of inferring demographic attributes such as age or sex from radar-derived cardiac signals. By applying repeated predictions the adversary can aggregate results via majority voting to improve classification accuracy. Although individual predictions may be affected by noise, the accumulation of repeated inferences over extended periods enables reliable profiling of individuals based on latent features contained in the physiological signal.
3.6 Experimental setup
To prevent data leakage, training and testing sets were split at the subject level, ensuring that all frames from a given individual were assigned exclusively to either the training or the test set. Specifically, for each experiment, 80% of the subjects were randomly assigned to the training set and the remaining 20% to the test set, while maintaining a balanced distribution of sex or age groups as needed. Frames from the training subjects were used to train the CNN, including both real scalograms and synthetic scalograms generated by a cWGAN trained only on the training subjects’ data, while the test set contained only real scalograms. Consequently, neither the CNN nor the cWGAN was exposed to any data from the test subjects during training. This subject-wise splitting strategy guarantees that the CNN learns meaningful physiological features without any contamination from the test data.
All experiments were conducted on a workstation running Ubuntu 22.04.4 LTS. The system was equipped with a 12th Gen Intel Core i7-12700KF CPU running at 5.00 GHz, 32 GB of DDR4 RAM, and an NVIDIA GeForce RTX 3080 Ti GPU with 12 GB of VRAM. The code was executed using Python 3.11.7, PyTorch 2.3.0, and CUDA 12.2. For wavelet-based signal processing, the functions modwt and modwtmra from MATLAB’s Wavelet Toolbox were invoked via the Python-MATLAB engine. Key parts of the implementation, including model architecture, MODWT filtering, data generation with cWGAN and splitting functions will be made publicly available for full reproducibility [115].
4 Results
The aim of this article is to investigate whether personal data, specifically sex and age group, can inadvertently leak alongside other medically relevant information in cardiac signals obtained through radar technology. To address this issue, we will explore the feasibility of predicting these two variables from the radar-derived signals in a substantial number of cases. For the train-test split in the experiments, since the goal is to estimate two attributes specific to each person, age group and sex, we are separating the data by individuals. This ensures that signals from the same individual do not appear in both the training and testing sets, preventing possible data leakage, which may have occurred in the use of other biosignals [116].
4.1 Metrics
To evaluate whether cardiac signals obtained via radar transmit sensitive patient information, such as sex or age, in addition to heart-specific data, we will employ the following metrics in the classification process. Accuracy measures the overall effectiveness of the model by calculating the proportion of correctly classified instances out of the total. False Acceptance Rate (FAR) indicates how often unauthorized individuals are incorrectly accepted by the system, reflecting the rate of false positives among all actual negatives. On the other hand, False Rejection Rate (FRR) measures how frequently authorized individuals are incorrectly rejected, showing the rate of false negatives among all actual positives. Precision evaluates the proportion of true positive predictions among all positive predictions made by the model, while recall assesses the proportion of true positive instances correctly identified out of all actual positives. Finally, the F1-score combines precision and recall into a single metric, offering a balanced measure of the model’s performance in identifying positive instances [117].
4.2 Sex
To determine whether sex information is being inadvertently transmitted through the cardiac signals, we attempted to predict the sex of patients set aside for testing by training the classification model on the remaining ones. The dataset consists of 30 patients, with 16 females and 14 males. We reserved 6 patients (3 males and 3 females chosen randomly) for testing, and used the data of the other 24 patients for training. To enhance the robustness of the training process, we applied data augmentation by generating new scalograms that do not correspond to any specific patient but are labeled by class, following the procedure described in Section 3.3. In Figure 4, we present a comparison between real and generated samples, showing a random subset of real scalograms from the training set alongside synthetic scalograms produced by the cWGAN. For a quantitative assessment, we computed the Fréchet Inception Distance (FID) [118] between real and synthetic scalograms for the sex classification task, obtaining a value of 88.74. To adapt the grayscale scalograms to Inception-v3, the single channel was replicated three times and the images were resized to match the network’s input size. Although relatively high, this FID may be influenced by the domain mismatch between our grayscale scalograms and the RGB images used to pretrain Inception-v3, which can amplify the measured distance. To evaluate the impact of the amount of additional data, we performed the classification process four times by varying this amount, expressed as the percentage increase in scalograms added to the training set. The average accuracy results for sex prediction are shown in Figure 5. In this figure, we can observe that, except when increasing data by 25% or 150%, data augmentation with the cWGAN generally benefits the model’s efficiency in all other cases. Among these, optimal results appear to be achieved with a 75% data increase, where the test set window accuracy for sex prediction reaches nearly 72%. To evaluate whether the observed improvement with data augmentation is statistically significant, we repeated the classification experiment 34 times under two conditions: (i) without data augmentation, and (ii) with a 75% increase in the training set using synthetic scalograms. For each run, we computed the average test window accuracy. To compare both distributions, we applied Welch’s t-test, obtaining a t-statistic of 2.6169, a p-value of 0.0146, and a 95% confidence interval for the difference in means of [0.1080, 0.0130]. These findings indicate that the improvement in classification accuracy due to data augmentation is statistically significant, reinforcing the hypothesis that synthetically generated data enhances the model’s ability to predict sex.
Given the limited number of subjects in the dataset, we can perform a leave-one-out classification for each of the 6 test patients, using the other five as additional training material. In this way, for each of these 6 patients dedicated for test, a model is trained with the remaining 29 patients plus the generated scalograms. The goal of the data augmentation process was to generate samples conditioned on specific attributes, thereby enhancing the model’s ability to control and manipulate the output based on desired characteristics. Table 3 presents the results obtained using this methodology, comparing the classification performance with only the original data to the performance with a 75% increase in data through augmentation. The application of data augmentation led to substantial improvements across all metrics. The classification accuracy for 10 s windows (i.e., after aggregating frame-level predictions into a single decision per window) rose significantly, from 72.80% to 78.40%, reflecting a 5.6% improvement. This enhancement demonstrates the model’s increased capability to accurately classify sex due to the added variability in the data. Frame-level accuracy (i.e., the performance of the CNN when predicting individual 4 s frames) also benefited, showing a 4.11% increase, which indicates a consistent trend of enhanced performance even at finer time granularities. In terms of error rates, both the FAR and FRR declined, with reductions of 6.02% and 5.69%, respectively, after applying data augmentation. Precision and recall also improved, with precision increasing by 6.02% and recall by 5.69%. These results suggest that the augmented dataset not only makes the model more accurate but also promotes a better balance between detecting true positives and minimizing false negatives. This balance is further evidenced by a 5.57% increase in the F1-score. Additionally, in Figure 6, we can observe the confusion matrix for the case where data augmentation is increased by 75%. It is important to note that we are classifying 10 s windows, which the numbers in the figure correspond to. In this figure, we can see that for the model, women are easier to identify (86.19% accuracy), whereas men present more difficulties (71.13% accuracy).
Additionally, we can analyze the accuracy of sex prediction for each window in relation to the patient, as depicted in Figure 7, where we see that there is a male whose windows are mostly classified by the model as belonging to female, thereby reducing the average results.
Given the limited number of subjects in the dataset, we can perform a leave-one-out classification for each of the 6 test patients, using the other five as additional training material. In this way, for each of these 6 patients dedicated for test, a model is trained with the remaining 29 patients plus the generated scalograms. The goal of the data augmentation process was to generate samples conditioned on specific attributes, thereby enhancing the model’s ability to control and manipulate the output based on desired characteristics. Table 3 presents the results obtained using this methodology, comparing the classification performance with only the original data to the performance with a 75% increase in data through augmentation. The application of data augmentation led to substantial improvements across all metrics. The classification accuracy for 10 s windows (i.e., after aggregating frame-level predictions into a single decision per window) rose significantly, from 72.80% to 78.40%, reflecting a 5.6% improvement. This enhancement demonstrates the model’s increased capability to accurately classify sex due to the added variability in the data. Frame-level accuracy (i.e., the performance of the CNN when predicting individual 4 s frames) also benefited, showing a 4.11% increase, which indicates a consistent trend of enhanced performance even at finer time granularities. In terms of error rates, both the FAR and FRR declined, with reductions of 6.02% and 5.69%, respectively, after applying data augmentation. Precision and recall also improved, with precision increasing by 6.02% and recall by 5.69%. These results suggest that the augmented dataset not only makes the model more accurate but also promotes a better balance between detecting true positives and minimizing false negatives. This balance is further evidenced by a 5.57% increase in the F1-score. All metrics reported, except for frame-level accuracy, are computed at the window level. Additionally, in Figure 6, we can observe the confusion matrix for the case where data augmentation is increased by 75%. It is important to note that we are classifying 10 s windows, which the numbers in the figure correspond to. In this figure, we can see that for the model, women are easier to identify (86.19% accuracy), whereas men present more difficulties (71.13% accuracy).
4.3 Age
Similarly to the sex analysis, we explored the extent to which personal information, such as age, might be conveyed through the available cardiac signals. Given the small dataset of only 30 subjects, age presents a more significant challenge, as illustrated in Figure 2b. There are age groups with no patients, and the distribution is clearly imbalanced. To mitigate this issue, we categorized the data into two groups: individuals aged 18 to 29 and those aged 30 or older. Given the imbalance in distribution, grouping them into a larger number of categories would have resulted in some groups having very few elements. Despite this grouping, the data distribution remains markedly imbalanced, as shown in Figure 8a. Therefore, data augmentation was crucial to achieve a more balanced training set. As illustrated in Figure 8b, after increasing the original dataset by 75%, the scalogram dataset for age classification became balanced across the two groups. Specifically, synthetic scalograms were generated until both classes had approximately equal sample sizes (around 14,000 each). This approach was preferred over oversampling or undersampling to preserve the original distribution and reduce the risk of overfitting. While other techniques such as SMOTE or class weighting were considered, the cWGAN was ultimately chosen due to its capacity to produce label-consistent and structurally realistic synthetic samples. In Figure 4, we also provide a comparison between real and synthetic scalograms for the age classification task. As in the sex classification analysis, we quantitatively assessed the similarity between real and synthetic scalograms for age by computing the FID, which resulted in a value of 39.56.
Figure 8. Distribution of frames in the training set by age group. (a) Without data augmentation. (b) After data augmentation of 75%.
Similar to our approach with sex, we can investigate the impact of the amount of new scalograms added to the training set on the classification of test patients by age group. This analysis is illustrated in Figure 9, where the best results are obtained by training the classifier with an additional 75% of generated data. To create this plot, we performed classification four times, depicting the average results. In this figure, it is observed that the use of data augmentation does notably improve the performance of the classification algorithm between the two groups. Furthermore, the accuracy takes on a concave shape, reaching its maximum when increasing the data by 75% and decreasing thereafter. On the other hand, in Table 4, we observe the metrics comparing the results after applying data augmentation and without applying it, following the procedure described for sex using leave-one-out classification with the aim of maximizing the training data. The inclusion of 75% data augmentation resulted in a marked improvement across all classification metrics. The window-level accuracy rose from 65.60% to 72.83%, an increase of 7.23%, while frame-level accuracy similarly increased by 6.48%, reaching 69.50%. In terms of error rates, both the FAR and the FRR exhibited notable decreases, dropping from 25.73% to 18.54% and from 34.73% to 27.46%, respectively. This reduction in error rates indicates that the model becomes more reliable with data augmentation, with fewer incorrect classifications on both acceptance and rejection. Precision and recall showed notable improvements, with increases of 7.19% and 7.27%, respectively, compared to the initial values. The F1-score also rose by 8.85%, reflecting a better balance between precision and recall. As in the case of sex, we evaluated the statistical significance of the observed improvement by repeating the classification experiment 34 times and applying Welch’s t-test. The results yielded a t-statistic of 2.4017, a p-value of 0.0225, and a 95% confidence interval of [0.1067, 0.0087], indicating that the performance gain from data augmentation is statistically significant in the age group classification task.
In Figures 10a, 10b, we can respectively observe the confusion matrix and the accuracy when predicting the age group for each person’s windows, ordered by age. In Figure 10a, we observe that the model exhibits a slight bias, tending to classify a higher number of windows into the most frequent age group in the dataset, which is 18 to 29 years. Additionally, one of the six individuals in the test set, a 35-year-old woman, is being classified as under 30. It is worth clarifying that the age detected through heart signals does not uniquely correspond to a person’s chronological age but rather expresses the general state of the cardiovascular system [12]. In certain cases, this may cause individuals with ages near the boundaries of age groups to be misclassified. This error might not be attributed to the model in some cases, but rather to the characteristics and condition of their particular cardiovascular system.
Figure 10. Performance metrics for age group classification. (a) Confusion matrix for classification in age group. (b) Accuracy in age group classification by person.
4.4 Explainability
Despite the strong performance often achieved by deep learning models, their black-box nature can limit interpretability and hinder trust in their outputs. In our context, where cardiac radar signals are analyzed using convolutional neural networks (CNNs) to infer demographic attributes, understanding the rationale behind the model’s decisions is crucial—both for scientific insight and for fostering confidence in the system. To this end, we employed Gradient-weighted Class Activation Mapping (Grad-CAM) [119] to generate visual explanations of the CNN’s predictions. Grad-CAM produces heatmaps that highlight the regions of the input scalogram that most strongly influence the output, enabling the identification of time-frequency patterns the model deems most relevant for distinguishing between classes such as sex or age group. Incorporating such explainability not only improves the transparency of our system but also provides a means for qualitative assessment and validation of the model’s behavior.
To analyze the internal decision-making process of our CNN classifier, we integrated Grad-CAM into the fourth convolutional layer of the network. We trained the CNN multiple times (five repetitions) with the augmented dataset to ensure robustness, and then applied Grad-CAM to the test set to identify the regions of the input scalograms that most influenced the model’s predictions. For each correctly classified instance, we extracted class-specific activation maps and aggregated them separately for each class (male and female). This yielded averaged heatmaps representing the most salient time-frequency patterns used by the network to distinguish between the two categories. The results for both sex and age classification tasks are presented in Figure 11. In general, the network consistently allocates more attention to the low-frequency regions of the scalogram, which appear to contain the most discriminative information for class separation, while higher frequencies receive comparatively less focus. This observation suggests that future work could benefit from emphasizing this lower frequency range, which seems to play a decisive role in the model’s decisions. In both classification tasks, we also observe class-specific differences in the regions of highest activation. These differences are particularly pronounced in the sex classification task, where the network tends to focus on slightly wider frequency bands for female subjects. As previously suggested in the literature [81], this might be attributed to morphological differences in the chest between male and female individuals, which could influence the cardiac radar signal and, consequently, the frequency content of the resulting scalograms. Similarly, in the age classification task, differences in activation patterns may reflect known physiological changes in cardiac dynamics with age. For instance, cardiac features such as Heart Rate Variability (HRV) have been shown to vary with age [120], which may contribute to the distinct attention patterns observed between the two age groups.
Figure 11. Grad-CAM heatmaps showing CNN attention regions on scalograms. (a) Age group classification. (b) Sex clasification.
4.5 Use of transfer learning
Due to the limited number of patients in our dataset, Transfer Learning is one of the most common approaches to improve model training in scenarios with scarce data [121]. Transfer Learning is particularly valuable in cases with restricted data availability, as it transfers relevant representations from previously trained models, optimizing the training process and potentially enhancing prediction accuracy. Moreover, its applicability extends to various biomedical tasks where acquiring large volumes of labeled data is both costly and challenging [122].
In our study, we evaluated four different pre-trained networks by applying fine-tuning to each. This fine-tuning process involves freezing the parameters of the network’s initial layers, thereby retaining the general features learned from a large and diverse dataset, while training only the final layers to adapt to the specific characteristics of our study. The networks evaluated included ResNet50 [123], DenseNet [124], ShuffleNet [125], and VGG-19 [126], all pre-trained on large-scale, real-world image datasets. Despite the potential advantages, the performance results achieved with these networks were not satisfactory. Tables 5, 6 show the classification results by sex and age group using each of these networks with 75% data augmentation applied consistently across all cases.
Table 5. Sex classification metrics with different pre-trained networks, using 75% of data augmentation.
Table 6. Age group classification metrics with different pre-trained networks, using 75% of data augmentation.
The performance of transfer learning models such as ResNet-50, DenseNet, ShuffleNet, and VGG-19 was noticeably lower than that of our custom-designed network. We attribute this performance gap primarily to domain mismatch [127]: these architectures are pre-trained on ImageNet, a dataset of natural images with vastly different structural and spectral characteristics compared to radar-derived cardiac scalograms. While transfer learning is effective when source and target domains are similar, in this case, the feature representations learned from natural textures, shapes, and objects do not transfer well to time-frequency biomedical signals. We also experimented with fine-tuning deeper layers and adding domain-specific dense layers on top of the pre-trained backbones. However, these adjustments did not yield substantial improvements, suggesting that the core convolutional filters learned from natural images were suboptimal for this task. These results indicate that task-specific architecture design and training from scratch are more effective strategies for modeling radar-acquired physiological signals.
4.6 Evaluating threat feasibility through temporal aggregation
To assess the practical feasibility of the privacy threat described in Section 3.5, we conducted an additional experiment based on the hypothesis that an adversary could improve the reliability of demographic inference by aggregating predictions over longer temporal spans. Specifically, we evaluated the classification accuracy of our model when predictions are aggregated from multiple consecutive input segments corresponding to the same individual. Since the model provides a prediction for each individual frame, we generated all possible contiguous subsets of frames for each subject in the test set (6 subjects), where each subset spans at least 7 seconds. For each subset, a final prediction was obtained by applying majority voting over the model’s predictions for the individual frames within that subset. We then grouped these subsets into bins according to their temporal length and computed the average accuracy for each bin. The results are shown in Figure 12a, where each line corresponds to a single test subject and each point represents the average accuracy over all subsets of that specific duration. As can be observed, in the age group classification task, one subject—a 35-year-old female—is consistently misclassified as belonging to the 18–29 age group, regardless of the temporal span. Aside from this case, the overall accuracy increases steadily as the input duration grows, supporting the notion that temporal aggregation significantly boosts inference reliability. In the age classification task, for example, three out of the six test subjects reach 100% accuracy when the aggregated input covers at least 20 s. Figure 12b complements this analysis by showing the cumulative percentage of subjects for whom the model achieves at least 95% accuracy at or before a given input duration. Notably, for both tasks, the model predicts the correct class with high confidence (95%) for 80% of subjects after observing less than 100 s of data. These findings provide empirical support for the proposed privacy threat model. By demonstrating that demographic attributes such as sex and age can be inferred with high confidence from temporally aggregated cardiac radar signals—even in the absence of physical contact or subject cooperation—our results highlight the tangible risks of long-term passive monitoring. The consistent increase in prediction accuracy with longer input durations suggests that an adversary with access to sustained radar observations could reliably construct demographic profiles of individuals over time.
Figure 12. Classification accuracy for both sex and age group as a function of temporal aggregation. (a) Average classification accuracy per test subject over temporally aggregated subsets of varying duration. (b) Cumulative percentage of test subjects for whom the model achieves at least 95% accuracy, as a function of temporal input duration.
5 Discussion & conclusions
The results of this research suggest that radar-extracted cardiac signals, while primarily intended to capture relevant medical features, also inadvertently encode personal information such as sex and age, which could pose privacy concerns. This characteristic, also observed in other biometric data types like facial recognition and ECG, highlights the importance of implementing robust security measures for the storage and management of such data. Specifically, we achieved an accuracy of 78.40% for correctly predicting sex, and 72.83% for predicting the patient’s age group, consistently using 10 s temporal windows. To accomplish this, we employed a model primarily based on a CNN that classifies the scalograms derived from cardiac signals into their respective categories. An important tool for improving these results was data augmentation, which we conducted using a cWGAN, also studying the amount of extra data that most positively influences the model’s performance. Both for sex (Figure 5) and for age (Figure 9), experiments determined that a 75% increase in the original training set offered optimal performance. This increase in accuracy was 5.6% in sex estimation and 7.23% for age (Tables 3, 4). On the other hand, the use of pre-trained networks on large image datasets (Transfer Learning) did not prove beneficial in our case, yielding results that were inferior to those of our own model, which was trained solely on samples obtained from the original dataset. Hyperparameter optimization for both the CNN classifier and the cWGAN model was not performed and remains an open task for future work. Due to the limited size of the dataset, we did not perform a full grid search. Instead, hyperparameters were selected through manual tuning informed by prior literature and validation performance. Future work will include a systematic search strategy, such as Bayesian optimization or nested cross-validation, to improve reproducibility and fairness.
However, while traditional methods such as ECG benefit from larger datasets and have shown high performance in demographic attribute estimation, they still face challenges regarding portability, comfort, and suitability for continuous monitoring. In contrast, radar-based techniques offer a good contactless alternative but are currently limited by smaller datasets, which can affect both model accuracy and generalization capacity. Additionally, due to the novelty of this approach, further studies are required to systematically evaluate its robustness in diverse real-world settings. This comparison can be seen in Table 7, where we compare our results with those obtained through other signals, as we are not aware of previous studies conducting this procedure with this type of signal. In this table, it is observed that the accuracy in estimating sex usually exceeds 90%, and the databases used sometimes include hundreds of thousands of patients.
Overall, the small dataset size can impose certain limitations on the generalization of these results, as dedicating only 6 patients for testing means that the incorrect prediction of just one of them can significantly influence the final results. This is evident in both sex and age cases, where in each, there is a subject consistently misclassified into the wrong class. In the case of sex, this occurs with patient 17 (Figure 7), who is the male with the lowest body weight in the dataset (28 years, 165 cm, 57 kg), while for age, it occurs with a 35-year-old woman, who is near the boundary between both age groups. Furthermore, public datasets for this type of signal remain very limited, with 30 patients representing a number that is actually higher than the average found in other studies [105]. It is also important to note that we are dividing the training set and the test set by patients, ensuring that the model has never seen data from the same patient when predicting the class of a window. This procedure brings the study’s results closer to potential real-world applications.
Our results show that classification accuracy improves significantly when predictions are aggregated over longer temporal windows, indicating that even models with moderate single-frame performance can achieve high reliability given sustained observations. These findings highlight the potential for passive adversaries to infer sensitive demographic attributes through unobtrusive, long-term data collection, emphasizing the need for robust privacy protections. Our concerns align with biometric information protection standards such as ISO/IEC 24745 [128], the General Data Protection Regulation (GDPR) [9], and NIST Special Publication 800-63 [129], which require safeguarding against unauthorized identity disclosure and the inference of secondary attributes without explicit consent. This work underscores the necessity to develop effective countermeasures and privacy-preserving techniques for radar-based cardiac monitoring systems. To mitigate these risks, future work should explore:
• Signal sanitization techniques, such as adversarial perturbations or autoencoder-based anonymization, to suppress demographic features while preserving the primary utility of the signal.
• Differential privacy mechanisms, applied at the feature extraction or model output level, to provide formal privacy guarantees against attribute inference.
• Detection and auditing frameworks capable of identifying unauthorized use of inference models or repeated inference attempts in deployed systems.
As a conclusion of our study, we can affirm the presence of personal information within radar-extracted cardiac signals that could be exploited for identification purposes. As a result, it is crucial to treat these signals with the same level of security and privacy considerations as more traditional biometric data. This is particularly important in the context of biomedical applications, where the misuse of personal health information could have serious ethical and legal implications. A specific example of how data leakage can compromise the security of widely used applications is the use of sex and age information in e-passports. This information is utilized to derive the access key for the RFID chip embedded in the passport, a practice that has been in effect since September 11 [11]. Future research should focus on improving the accuracy and robustness of these methods but also on developing protocols and technologies to safeguard the collected data. By addressing these concerns, we can harness the potential of radar-based cardiac signal analysis for a wide range of medical and healthcare applications while maintaining the trust and confidence of the public.
Therefore, non-invasive monitoring methods, such as radar-based cardiac signal analysis, offer substantial benefits in healthcare by enabling continuous monitoring without the need for direct contact with the patient. This technology is particularly valuable in scenarios where patient mobility or frequent follow-up is challenging, as it provides an efficient, unobtrusive way to track vital signs and other health metrics over time. However, this accessibility also introduces privacy risks, as sensitive biometric information can be derived from these signals, including details about a patient’s sex, age, and possibly unique identifying characteristics. If improperly safeguarded, this information could be exploited in unintended ways, posing risks to patient privacy and autonomy. The trade-off lies in balancing the advantages of improved patient monitoring and personalized care with stringent data protection measures to prevent unauthorized access and misuse of personal health information. Ultimately, ensuring the security of radar-based health data will be essential for achieving patient trust, regulatory compliance, and ethical application in healthcare.
Data availability statement
The authors affirm that this study utilized a pre-existing, publicly available dataset sourced from a previous publication. Specifically, the data employed in this research were obtained from “A dataset of clinically recorded radar vital signs with synchronised reference sensor signals”, originally published by Sven Schellenberger, Kilin Shi, Tobias Steigleder, Anke Malessa, Fabian Michler, Laura Hameyer, Nina Neumann, Fabian Lurz, Robert Weigel, Christoph Ostgathe, and Alexander Koelpin in Scientific Data (2020).
Our use of this dataset adheres to its Creative Commons Attribution 4.0 International (CC BY 4.0) license. Full and proper citation of the original source is provided in the References section of this article.
Ethics statement
This study utilizes a publicly available dataset which was collected in compliance with relevant regulations and approved by the Ethics Committee of the Friedrich-Alexander-Universität Erlangen-Nürnberg (Approval No. 85_15B). Written informed consent to participate in this study was not required from the participants or the participants' legal guardians/next of kin in accordance with the national legislation and the institutional requirements. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author contributions
DF-P: Validation, Software, Supervision, Methodology, Writing – review & editing, Conceptualization, Investigation, Writing – original draft. CC: Funding acquisition, Validation, Conceptualization, Supervision, Writing – review & editing, Methodology. PP-L: Supervision, Methodology, Validation, Conceptualization, Writing – review & editing, Project administration, Funding acquisition.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by grants TED2021-131681B-I00 (CIOMET) and PID2022-140126OB-I00 (CYCAD) from the Spanish Ministry of Science, Innovation and Universities, as well as by the INCIBE under the project VITAL-IoT in the context of the funds from the Recovery, Transformation, and Resilience Plan, financed by the European Union (Next Generation).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that Generative AI was used in the creation of this manuscript. During the preparation of this work the authors used ChatGPT 3.5 in order to detect grammatical errors or typos. After using this tool/service, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Kaniusas E. Physiological Phenomena and Biosignals. Berlin, Heidelberg: Springer (2012). p. 183–282.
2. Hasnul MA, Aziz NAA, Alelyani S, Mohana M, Aziz AA. Electrocardiogram-based emotion recognition systems and their applications in healthcare—a review. Sensors. (2021) 21:5015. doi: 10.3390/s21155015
3. Uwaechia AN, Ramli DA. A comprehensive survey on ecg signals as new biometric modality for human authentication: recent advances and future challenges. IEEE Access. (2021) 9:97760–802. doi: 10.1109/ACCESS.2021.3095248
4. Wenhua Z, Hasan MK, Jailani NB, Islam S, Safie N, Albarakati HM, et al. A lightweight security model for ensuring patient privacy and confidentiality in telehealth applications. Comput Human Behav. (2024) 153:108134. doi: 10.1016/j.chb.2024.108134
5. Beltrão G, Weber R, Hornberger F, Martins W, Tatarinov D, Alaee-Kerahroodi M, et al. Contactless radar-based breathing monitoring of premature infants in the neonatal intensive care unit. Sci Rep. (2022) 12:5150. doi: 10.1038/s41598-022-08836-3
6. Li C, Lubecke VM, Boric-Lubecke O, Lin J. A review on recent advances in doppler radar sensors for noncontact healthcare monitoring. IEEE Trans Microw Theory Tech. (2013) 61:2046–60. doi: 10.1109/TMTT.2013.2256924
7. Fioranelli F, Guendel RG, Kruse NC, Yarovoy A. Radar sensing in healthcare: challenges and achievements in human activity classification & vital signs monitoring. In: International Work-Conference on Bioinformatics and Biomedical Engineering. Springer (2023). p. 492–504.
8. Jain AK, Deb D, Engelsma JJ. Biometrics: trust, but verify. IEEE Trans Biom Behav Identity Sci. (2021) 4:303–23. doi: 10.1109/TBIOM.2021.3115465
9. European Union. Data from: Regulation (eu) 2016/679 of the European parliament and of the council on the protection of natural persons with regard to the processing of personal data and on the free movement of such data (general data protection regulation). GDPR. (2016). Available online at: https://eur-lex.europa.eu/eli/reg/2016/679/oj.
10. Data from: California consumer privacy act of 2018 (cal. civ. code §1798.100 et seq.) (2018). Available online at: https://leginfo.legislature.ca.gov/faces/codes_displayText.xhtml?division=3.&part=4.&lawCode=CIV&title=1.81.5. As amended by Proposition 24 (CPRA), effective Jan.1, 2020; Civil Code Title 1.81.5.
11. Avoine G, Beaujeant A, Hernandez-Castro J, Demay L, Teuwen P. A survey of security and privacy issues in epassport protocols. ACM Comput Surv. (2016) 48:1–37. doi: 10.1145/2825026
12. Ansari MY, Qaraqe M, Charafeddine F, Serpedin E, Righetti R, Qaraqe K. Estimating age and gender from electrocardiogram signals: a comprehensive review of the past decade. Artif Intell Med. (2023) 146:102690. doi: 10.1016/j.artmed.2023.102690
13. Kaushik P, Gupta A, Roy PP, Dogra DP. Eeg-based age and gender prediction using deep blstm-lstm network model. IEEE Sens J. (2018) 19:2634–41. doi: 10.1109/JSEN.2018.2885582
14. Dehghanojamahalleh S, Kaya M. Sex-related differences in photoplethysmography signals measured from finger and toe. IEEE J Transl Eng Health Med. (2019) 7:1–7. doi: 10.1109/JTEHM.2019.2938506
15. Asyrafi H, Sunarno , Faridah , Waruwu M, Wijaya R. Photoplethysmography (PPG) signal-based chronological age estimation algorithm. In: 2022 International Conference of Science and Information Technology in Smart Administration (ICSINTESA). IEEE (2022). p. 128–32.
16. Lu J, Tan YP. Gait-based human age estimation. IEEE Trans Inf Forensics Secur. (2010) 5:761–70. doi: 10.1109/TIFS.2010.2069560
17. Liu J, Liu K, Pan X, Bi K, Zhou F, Lu P, et al. A flexible semidry electrode for long-term, high-quality electrocardiogram monitoring. Adv Compos Hybrid Mater. (2023) 6:13. doi: 10.1007/s42114-022-00596-y
18. Lindsey B, Hanley C, Reider L, Snyder S, Zhou Y, Bell E, et al. Accuracy of heart rate measured by military-grade wearable ecg monitor compared with reference and commercial monitors. BMJ Mil Health. (2025) 171:144–9. doi: 10.1136/military-2023-002541
19. Kim J, Kim E, Kyung Y, Ko H. A study on ECG monitoring embedded systems. In: 2022 13th International Conference on Information and Communication Technology Convergence (ICTC). IEEE (2022). p. 1671–3.
20. Liu X, Wang H, Li Z, Qin L. Deep learning in ecg diagnosis: a review. Knowl Based Syst. (2021) 227:107187. doi: 10.1016/j.knosys.2021.107187
21. Al-Zaiti SS, Martin-Gill C, Zègre-Hemsey JK, Bouzid Z, Faramand Z, Alrawashdeh MO, et al. Machine learning for ECG diagnosis and risk stratification of occlusion myocardial infarction. Nat Med. (2023) 29:1804–13. doi: 10.1038/s41591-023-02396-3
22. Yu H, Guo P, Sano A. Zero-shot ECG diagnosis with large language models and retrieval-augmented generation. In: Machine Learning for Health (ML4H). Cambridge, MA: PMLR (2023). p. 650–63.
23. Boulif A, Ananou B, Ouladsine M, Delliaux S. A literature review: ECG-based models for arrhythmia diagnosis using artificial intelligence techniques. Bioinform Biol Insights. (2023) 17:11779322221149600. doi: 10.1177/11779322221149600
24. Smaranda AM, Drăgoiu TS, Caramoci A, Afetelor AA, Ionescu AM, Bădărău IA. Artificial intelligence in sports medicine: reshaping electrocardiogram analysis for athlete safety—a narrative review. Sports. (2024) 12:144. doi: 10.3390/sports12060144
25. Rabbani S, Khan N. Contrastive self-supervised learning for stress detection from ECG data. Bioengineering. (2022) 9:374. doi: 10.3390/bioengineering9080374
26. Fan T, Qiu S, Wang Z, Zhao H, Jiang J, Wang Y, et al. A new deep convolutional neural network incorporating attentional mechanisms for ECG emotion recognition. Comput Biol Med. (2023) 159:106938. doi: 10.1016/j.compbiomed.2023.106938
27. Melzi P, Tolosana R, Vera-Rodriguez R. ECG biometric recognition: review, system proposal, and benchmark evaluation. IEEE Access. (2023) 11:15555–66. doi: 10.1109/ACCESS.2023.3244651
28. Mitchell AR, Ahlert D, Brown C, Birge M, Gibbs A. Electrocardiogram-based biometrics for user identification–using your heartbeat as a digital key. J Electrocardiol. (2023) 80:1–6. doi: 10.1016/j.jelectrocard.2023.04.001
29. Dong S, Wen L, Ye Y, Zhang Z, Wang Y, Liu Z, et al. A review on recent advancements of biomedical radar for clinical applications. IEEE Open J Eng Med Biol. (2024) 5:707–24. doi: 10.1109/OJEMB.2024.3401105
30. Kebe M, Gadhafi R, Mohammad B, Sanduleanu M, Saleh H, Al-Qutayri M. Human vital signs detection methods and potential using radars: a review. Sensors. (2020) 20:1454. doi: 10.3390/s20051454
31. Vitazkova D, Foltan E, Kosnacova H, Micjan M, Donoval M, Kuzma A, et al. Advances in respiratory monitoring: a comprehensive review of wearable and remote technologies. Biosensors. (2024) 14:90. doi: 10.3390/bios14020090
32. Vysotskaya N, Maul N, Fusco A, Hazra S, Harnisch J, Arias-Vergara T, et al.. Transforming cardiovascular health: a transformer-based approach to continuous, non-invasive blood pressure estimation via radar sensing. In: ICASSP 2024–2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE (2024). p. 2041–5.
33. Boiko A, Martínez Madrid N, Seepold R. Contactless technologies, sensors, and systems for cardiac and respiratory measurement during sleep: a systematic review. Sensors. (2023) 23:5038. doi: 10.3390/s23115038
34. Lopes ASD. Bio-radar applications for remote vital signs monitoring (Master’s thesis). Lisbon: Faculdade de Ciências e Tecnologia, Universidade Nova de Lisboa (2021).
35. Cho HS, Park YJ. Detection of heart rate through a wall using uwb impulse radar. Med Signal Process Biomed Clin Appl. (2018) 2018:4832605. doi: 10.1155/2018/4832605
36. Liu L, Xiao W, Wu J, Xiao S. Wavelet analysis based noncontact vital signal measurements using mm-wave radar. In: Yu Z, Becker C, Xing G, editors. Lecture Notes in Computer Science. Cham: Springer International Publishing (2020). Vol. 12398.
37. Gouveia C, Albuquerque D, Pinho P, Vieira J. Evaluation of heartbeat signal extraction methods using a 5.8 GHz doppler radar system in a real application scenario. IEEE Sens J. (2022) 22:7979–89. doi: 10.1109/JSEN.2022.3156474
38. Krauss D, Engel L, Ott T, Bräunig J, Richer R, Gambietz M, et al. A review and tutorial on machine learning-enabled radar-based biomedical monitoring. IEEE Open J Eng Med Biol. (2024) 5:680–99. doi: 10.1109/OJEMB.2024.3397208
39. Yang ZK, Zhao S, Huang XD, Lu W. Accurate doppler radar-based heart rate measurement using matched filter. IEICE Electron Express. (2020) 17:20200062. doi: 10.1587/elex.17.20200062
40. Ling Z, Zhou W, Ren Y, Wang J, Guo L. Non-contact heart rate monitoring based on millimeter wave radar. IEEE Access. (2022) 10:74033–44. doi: 10.1109/ACCESS.2022.3190355
41. Xiang Y, Guo J, Chen M, Wang Z, Han C. Mae-based self-supervised pretraining algorithm for heart rate estimation of radar signals. Sensors. (2023) 23:7869. doi: 10.3390/s23187869
42. Ling Z, Zhou W, Ren Y, Wang J, Guo L. Non-contact heart rate monitoring based on millimeter wave radar. IEEE Access. (2022) 10:74033–44. doi: 10.1109/ACCESS.2022.3190355
43. Huang IW, Rajbhandary P, Shiu S, Ho JS, Zhu J, Wilson B, et al. Radar-based heart rate sensing on the smart glasses. IEEE Microw Wirel Technol Lett. (2024) 34:829–32. doi: 10.1109/LMWT.2024.3391660
44. Yuan S, Fan S, Deng Z, Pan P. Heart rate variability monitoring based on doppler radar using deep learning. Sensors. (2024) 24:2026. doi: 10.3390/s24072026
45. Uddin SD, Hossain MS, Islam SM, Lubecke V. Heart rate variability-based obstructive sleep apnea events classification using microwave doppler radar. IEEE J Electromagn RF Microw Med Biol. (2023) 7:416–24. doi: 10.1109/JERM.2023.3317304
46. Wang H, Chen J, Zhang D, Lu Z, Wu C, Hu Y, et al.. Contactless radar heart rate variability monitoring via deep spatio-temporal modeling. In: ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE (2024). p. 111–5.
47. Sakamoto T, Iwata I, Minami T, Matsumoto T. Noncontact sensing of animals using radar. IEICE Trans Electron. (2024) 107.C:279–87. doi: 10.1587/transele.2024MMI0002
48. Matsumoto T, Iwata I, Sakamoto T, Hirata S. First noncontact millimeter-wave radar measurement of heart rate in great apes: validation in chimpanzees (pan troglodytes). Am J Primatol. (2024) 86:e23633. doi: 10.1002/ajp.23633
49. Sun G, Kurosawa M, Ninomiya Y, Baba K, Kano Y. Updating “medical radar signal dataset for non-contact respiration and heart rate measurement” with expanded data on laboratory rats under isoflurane anesthesia. Data Brief. (2024) 55:110722. doi: 10.1016/j.dib.2024.110722
50. Shirakami I, Sato T. Heart rate variability extraction using commodity wi-fi devices via time domain signal processing. In: 2021 IEEE EMBS International Conference on Biomedical and Health Informatics (BHI). (2021). p. 1–4.
51. Li Z, Jin T, Guan D, Xu H. Metaphys: contactless physiological sensing of multiple subjects using ris-based 4-d radar. IEEE Internet Things J. (2023) 10:12616–26. doi: 10.1109/JIOT.2023.3252587
52. Shyu KK, Chiu LJ, Lee PL, Tung TH, Yang SH. Detection of breathing and heart rates in UWB radar sensor data using FVPIEF-based two-layer EEMD. IEEE Sens J. (2018) 19:774–84. doi: 10.1109/JSEN.2018.2878607
53. Liebetruth M, Kehe K, Steinritz D, Sammito S. Systematic literature review regarding heart rate and respiratory rate measurement by means of radar technology. Sensors. (2024) 24:1003. doi: 10.3390/s24031003
54. Will C, Shi K, Schellenberger S, Steigleder T, Michler F, Fuchs J, et al. Radar-based heart sound detection. Sci Rep. (2018) 8:11551. doi: 10.1038/s41598-018-29984-5
55. Gaur P, Gupta A. Radar based heart beat detection. In: 2024 IEEE International Conference for Women in Innovation, Technology & Entrepreneurship (ICWITE). IEEE (2024). p. 265–70.
56. Pu J, Zhang H. RF-heartbeat: robust and contactless heartbeat monitoring based on FMCW radar. Authorea Preprints (2023).
57. Darwin S, Raj EFI. A detailed review on embedded based heartbeat monitoring systems. In: Sinha A, Sharma A, Melek LAP, Caviglia D, editors. Smart Embedded Systems and Applications. Boca Raton, FL: CRC Press, Taylor & Francis (2023). p. 79–102.
58. Tchameni MR, Schiffer C, Lücken V, Diewald AR, Schroeder U. Radar based heartbeat estimation in time domain. In: 2024 15th German Microwave Conference (GeMiC). IEEE (2024). p. 129–32.
59. Paterniani G, Sgreccia D, Davoli A, Guerzoni G, Di Viesti P, Valenti ACetal. Radar-based monitoring of vital signs: a tutorial overview. Proc IEEE. (2023) 111:277–317. doi: 10.1109/JPROC.2023.3244362
60. Fan S, Deng Z. Chest wall motion model of cardiac activity for radar-based vital-sign-detection system. Sensors. (2024) 24:2058. doi: 10.3390/s24072058
61. Choi SH, Yoon H. Convolutional neural networks for the real-time monitoring of vital signs based on impulse radio ultrawide-band radar during sleep. Sensors. (2023) 23:3116. doi: 10.3390/s23063116
62. Lin F, Song C, Zhuang Y, Xu W, Li C, Ren K. Cardiac scan: a non-contact and continuous heart-based user authentication system. In: Proceedings of the 23rd Annual International Conference on Mobile Computing and Networking. New York (NY), USA: Association for Computing Machinery (2017). MobiCom ’17. p. 315–28.
63. Shi K, Will C, Weigel R, Koelpin A. Contactless person identification using cardiac radar signals. In: 2018 IEEE International Instrumentation and Measurement Technology Conference (I2MTC). (2018). p. 1–6.
64. Cao P, Xia W, Li Y. Human identification based on radar micro-doppler signatures of the heart using deep learning. Remote Sens. (2019) 11:1220. doi: 10.3390/rs11101220
65. Louro da Silva M, Gouveia C, Albuquerque DF, Plácido da Silva H. Radar-based invisible biometric authentication. Information. (2024) 15:44. doi: 10.3390/info15010044
66. Van NTP, Tang L, Singh A, Minh ND, Mukhopadhyay SC, Hasan SF. Self-identification respiratory disorder based on continuous wave radar sensor system. IEEE Access. (2019) 7:40019–26. doi: 10.1109/ACCESS.2019.2906885
67. Otero M. Application of a continuous wave radar for human gait recognition. In: Signal Processing, Sensor Fusion, and Target Recognition XIV. Bellingham, WA: SPIE (2005). Vol. 5809. p. 538–48.
68. Forouzanfar M, Mabrouk M, Rajan S, Bolic M, Dajani HR, Groza VZ. Event recognition for contactless activity monitoring using phase-modulated continuous wave radar. IEEE Trans Biomed Eng. (2016) 64:479–91. doi: 10.1109/TBME.2016.2566619
69. Gouveia C, Tomé A, Barros F, Soares SC, Vieira J, Pinho P. Study on the usage feasibility of continuous-wave radar for emotion recognition. Biomed Signal Process Control. (2020) 58:101835. doi: 10.1016/j.bspc.2019.101835
70. Pegoraro J, Meneghello F, Rossi M. Multiperson continuous tracking and identification from mm-wave micro-doppler signatures. IEEE Trans Geosci Remote Sens. (2020) 59:2994–3009. doi: 10.1109/TGRS.2020.3019915
71. Di Mascio T, Fantozzi P, Laura L, Rughetti V. Age and gender (face) recognition: a brief survey. In: Methodologies and Intelligent Systems for Technology Enhanced Learning, 11th International Conference 11. Springer (2022). p. 105–13.
72. Attia ZI, Friedman PA, Noseworthy PA, Lopez-Jimenez F, Ladewig DJ, Satam G, et al. Age and sex estimation using artificial intelligence from standard 12-lead ECGs. Circ Arrhythm Electrophysiol. (2019) 12:e007284. doi: 10.1161/CIRCEP.119.007284
73. Cabra Lopez JL, Parra C, Gomez L, Trujillo L. Sex recognition through ECG signals aiming toward smartphone authentication. Appl Sci. (2022) 12:6573. doi: 10.3390/app12136573
74. Ryu JS, Lee S, Chu Y, Koh SB, Park YJ, Lee JY, et al. Deep learning algorithms for estimation of demographic and anthropometric features from electrocardiograms. J Clin Med. (2023) 12:2828. doi: 10.3390/jcm12082828
75. Lopez-Jimenez F, Kapa S, Friedman PA, LeBrasseur NK, Klavetter E, Mangold KE, et al. Assessing biological age: the potential of ecg evaluation using artificial intelligence: JACC family series. Clin. Electrophysiol. (2024) 10:775–89. doi: 10.1016/j.jacep.2024.02.011
76. van der Wall H, Hassing G, Doll R, van Westen G, Cohen A, Selder J, et al. Cardiac age detected by machine learning applied to the surface ecg of healthy subjects: creation of a benchmark. J Electrocardiol. (2022) 72:49–55. doi: 10.1016/j.jelectrocard.2022.03.001
77. Chang CH, Lin CS, Luo YS, Lee YT, Lin C. Electrocardiogram-based heart age estimation by a deep learning model provides more information on the incidence of cardiovascular disorders. Front Cardiovasc Med. (2022) 9:754909. doi: 10.3389/fcvm.2022.754909
78. Wang Z, Meng Z, Saho K, Uemura K, Nojiri N, Meng L. Deep learning-based elderly gender classification using doppler radar. Pers Ubiquitous Comput. (2022) 26:1067–79. doi: 10.1007/s00779-020-01490-4
79. Saleem AA, Siddiqui HUR, Sehar R, Dudley S. Gender classification based on gait analysis using ultrawide band radar augmented with artificial intelligence. Expert Syst Appl. (2024) 249:123843. doi: 10.1016/j.eswa.2024.123843
80. Saho K, Shioiri K, Fujimoto M, Kobayashi Y. Micro-doppler radar gait measurement to detect age-and fall risk-related differences in gait: a simulation study on comparison of deep learning and gait parameter-based approaches. IEEE Access. (2021) 9:18518–26. doi: 10.1109/ACCESS.2021.3053298
81. Shaw V, Pah ND, Rani P, Mahapatra PK, Pankaj D, Kumar DK. Impact of biological sex on radar-measured heart sound quality. In: 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC). (2023). p. 1–5.
82. Raghu A, Shanmugam D, Pomerantsev E, Guttag J, Stultz CM. Data augmentation for electrocardiograms. In: Conference on Health, Inference, and Learning. PMLR (2022). p. 282–310.
83. Nonaka N, Seita J. Data augmentation for electrocardiogram classification with deep neural network. arXiv [Preprint]. arXiv:2009.04398 (2020).
84. Nonaka N, Seita J. Randecg: Data augmentation for deep neural network based ecg classification. In: Annual Conference of the Japanese Society for Artificial Intelligence. Springer (2021). p. 178–89.
85. Pan Q, Li X, Fang L. Data augmentation for deep learning-based ECG analysis. In: Liu C, Li J, editors. Feature Engineering and Computational Intelligence in ECG Monitoring. Singapore: Springer Nature Singapore Pte Ltd (2020). p. 91–111.
86. Do E, Boynton J, Lee BS, Lustgarten D. Data augmentation for 12-lead ECG beat classification. SN Comput Sci. (2022) 3:1–17. doi: 10.1007/s42979-021-00924-x
87. Chlap P, Min H, Vandenberg N, Dowling J, Holloway L, Haworth A. A review of medical image data augmentation techniques for deep learning applications. J Med Imaging Radiat Oncol. (2021) 65:545–63. doi: 10.1111/1754-9485.13261
88. Lashgari E, Liang D, Maoz U. Data augmentation for deep-learning-based electroencephalography. J Neurosci Methods. (2020) 346:108885. doi: 10.1016/j.jneumeth.2020.108885
89. Zhang Z, Liu Y, Zhong S. Ganser: a self-supervised data augmentation framework for EEG-based emotion recognition. IEEE Trans Affect Comput. (2023) 14:2048–63. doi: 10.1109/TAFFC.2022.3170369
90. Luo Y, Lu BL. EEG data augmentation for emotion recognition using a conditional wasserstein GAN. In: 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). (2018). p. 2535–8.
91. Zhang A, Su L, Zhang Y, Fu Y, Wu L, Liang S. EEG data augmentation for emotion recognition with a multiple generator conditional wasserstein GAN. Complex Intell Syst. (2021) 8:3059–71. doi: 10.1007/s40747-021-00336-7
92. Luo Y, Zhu LZ, Lu BL. A GAN-based data augmentation method for multimodal emotion recognition. In: Advances in Neural Networks–ISNN 2019: 16th International Symposium on Neural Networks, ISNN 2019, Moscow, Russia, July 10–12, 2019, Proceedings, Part I 16. Springer (2019). p. 141–50.
93. Zhang Y, Wang Z, Zhang Z, Liu J, Feng Y, Wee L, et al. Gan-based one dimensional medical data augmentation. Soft comput. (2023) 27:10481–91. doi: 10.1007/s00500-023-08345-z
94. Zhao D, Guo G, Ni ZK, Pan J, Yan K, Fang GW. A gans-based data augmentation method for GPR data. IEEE Geosci Remote Sens Lett. (2023) 20:1–5. doi: 10.1109/LGRS.2023.3323981
95. Mi J, Ma C, Zheng L, Zhang M, Li M, Wang MW. A wasserstein GAN with confidence loss for small-sample augmentation. Expert Syst Appl. (2023) 233:120943. doi: 10.1016/j.eswa.2023.120943
96. Safdar MF, Pałka P, Nowak RM, Al Faresi A. A novel data augmentation approach for enhancement of ECG signal classification. Biomed Signal Process Control. (2023) 86:105114. doi: 10.1016/j.bspc.2023.105114
97. Hasnul MA, Ab Aziz NA, Abd Aziz A. Augmenting ecg data with multiple filters for a better emotion recognition system. Arab J Sci Eng. (2023) 48:10313–34. doi: 10.1007/s13369-022-07585-9
98. Arjovsky M, Chintala S, Bottou L. Wasserstein generative adversarial networks. In: International Conference on Machine Learning. PMLR (2017). p. 214–23.
99. Lupión M, Cruciani F, Cleland I, Nugent C, Ortigosa PM. Data augmentation for human activity recognition with generative adversarial networks. IEEE J Biomed Health Inf. (2024) 28:2350–61. doi: 10.1109/JBHI.2024.3364910
100. Yu D, Bouazizi M, Ohtsukil T. Improving heart rate range classification using doppler radar with gan-based data augmentation. In: GLOBECOM 2023 – 2023 IEEE Global Communications Conference. (2023). p. 3885–90.
101. Korti DS, Slimane Z. Advanced human activity recognition through data augmentation and feature concatenation of micro-doppler signatures. Int J Electr Comput Eng Syst. (2023) 14:893–902. doi: 10.32985/ijeces.14.8.7
102. Schellenberger S, Shi K, Steigleder T, Malessa A, Michler F, Hameyer L, et al. A dataset of clinically recorded radar vital signs with synchronised reference sensor signals. Sci Data. (2020) 7:291. doi: 10.1038/s41597-020-00629-5
103. Schellenberger S, Shi K, Steigleder T, Malessa A, Michler F, Hameyer L, et al. Data from: A dataset of clinically recorded radar vital signs with synchronised reference sensor signals (2020). Available online at: https://figshare.com/articles/dataset/A_dataset_of_clinically_recorded_radar_vital_signs_with_synchronised_reference_sensor_signals/12186516.
104. Nocera A, Senigagliesi L, Raimondi M, Ciattaglia G, Gambi E. Machine learning in radar-based physiological signals sensing: a scoping review of the models, datasets and metrics. IEEE Access. (2024) 12:156082–117. doi: 10.1109/ACCESS.2024.3482690
105. Islam SMM, Borić-Lubecke O, Zheng Y, Lubecke VM. Radar-based non-contact continuous identity authentication. Remote Sens. (2020) 12:2279. doi: 10.3390/rs12142279
106. Zakrzewski M, Singh A, Yavari E, Gao X, Boric-Lubecke O, Vanhala J, et al. Quadrature imbalance compensation with ellipse-fitting methods for microwave radar physiological sensing. IEEE Trans Microw Theory Tech. (2014) 62:1400–8. doi: 10.1109/TMTT.2014.2321738
107. Park BK, Boric-Lubecke O, Lubecke VM. Arctangent demodulation with DC offset compensation in quadrature doppler radar receiver systems. IEEE Trans Microw Theory Tech. (2007) 55:1073–9. doi: 10.1109/TMTT.2007.895653
108. Bouallegue G, Djemal R. EEG data augmentation using wasserstein GAN. In: 2020 20th International Conference on Sciences and Techniques of Automatic Control and Computer Engineering (STA). IEEE (2020). p. 40–5.
109. Sundaram S, Hulkund N. Gan-based data augmentation for chest x-ray classification. arXiv [Preprint]. arXiv:2107.02970 (2021).
110. Wang G, Kang W, Wu Q, Wang Z, Gao J. Generative adversarial network (GAN) based data augmentation for palmprint recognition. In: 2018 Digital Image Computing: Techniques and Applications (DICTA). IEEE (2018). p. 1–7.
111. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial networks. Commun ACM. (2020) 63:139–44. doi: 10.1145/3422622
112. Diogo P, Morais M, Calisto FM, Santiago C, Aleluia C, Nascimento JC. Weakly-supervised diagnosis and detection of breast cancer using deep multiple instance learning. In: 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI). IEEE (2023). p. 1–4.
113. Morais M, Calisto FM, Santiago C, Aleluia C, Nascimento JC. Classification of breast cancer in mri with multimodal fusion. In: 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI). IEEE (2023). p. 1–4.
114. Santurkar S, Tsipras D, Ilyas A, Madry A. How does batch normalization help optimization? Adv Neural Inf Process Syst. (2018) 31:2488–98. https://dl.acm.org/doi/10.5555/3327144.3327174
115. Foronda-Pascual D. Data from: Radar-cardiac-age-sex-estimation: code repository for “gender and age estimation from cardiac signals captured via radar using data augmentation and deep learning: a privacy concern.” (2025). Available online at: https://github.com/dani1717/radar-cardiac-age-sex-estimation (Accessed June 09, 2025).
116. Yoshizawa R, Yamamoto K, Ohtsuki T. Investigation of data leakage in deep-learning-based blood pressure estimation using photoplethysmogram/electrocardiogram. IEEE Sens J. (2023) 23:13311–8. doi: 10.1109/JSEN.2023.3272921
117. Ferrer L. Analysis and comparison of classification metrics. arXiv [Preprint]. arXiv:2209.05355 (2022).
118. Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S. GANs trained by a two time-scale update rule converge to a local nash equilibrium. Adv Neural Inf Process Syst. (2017) 30. https://dl.acm.org/doi/10.5555/3295222.3295408
119. Selvaraju RR, Das A, Vedantam R, Cogswell M, Parikh D, Batra D. Grad-cam: Why did you say that? arXiv [Preprint]. arXiv:1611.07450 (2016).
120. Reardon M, Malik M. Changes in heart rate variability with age. Pacing Clin Electrophysiol. (1996) 19:1863–6. doi: 10.1111/j.1540-8159.1996.tb03241.x
121. Zhuang F, Qi Z, Duan K, Xi D, Zhu Y, Zhu H, et al. A comprehensive survey on transfer learning. Proc IEEE. (2020) 109:43–76. doi: 10.1109/JPROC.2020.3004555
122. Kim HE, Cosa-Linan A, Santhanam N, Jannesari M, Maros ME, Ganslandt T. Transfer learning for medical image classification: a literature review. BMC Med Imaging. (2022) 22:69. doi: 10.1186/s12880-022-00793-7
123. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. (2016). p. 770–8.
124. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. (2017). p. 4700–8.
125. Zhang X, Zhou X, Lin M, Sun J. Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. (2018). p. 6848–56.
126. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv [Preprint]. arXiv:1409.1556 (2014).
127. Zhao Z, Alzubaidi L, Zhang J, Duan Y, Gu Y. A comparison review of transfer learning and self-supervised learning: definitions, applications, advantages and limitations. Expert Syst Appl. (2024) 242:122807. doi: 10.1016/j.eswa.2023.122807
128. International Organization for Standardization. Data from: Information security, cybersecurity and privacy protection – biometric information protection. ISO/IEC 24745:2022 (2022). Available online at: https://www.iso.org/es/contents/data/standard/07/53/75302.html?browse=ics.
129. National Institute of Standards and Technology (NIST). Digital identity guidelines. Tech. Rep. NIST Special Publication 800-63-3. U.S. Department of Commerce (2017).
130. Kaushik P, Gupta A, Roy PP, Dogra DP. EEG-based age and gender prediction using deep blstm-lstm network model. IEEE Sens J. (2019) 19:2634–41. doi: 10.1109/JSEN.2018.2885582
131. Lyle JV, Nandi M, Aston PJ. Symmetric projection attractor reconstruction: sex differences in the ecg. Front Cardiovasc Med. (2021) 8:709457. doi: 10.3389/fcvm.2021.709457
132. Lima EM, Ribeiro AH, Paixão GM, Ribeiro MH, Pinto-Filho MM, Gomes PR, et al. Deep neural network-estimated electrocardiographic age as a mortality predictor. Nat Commun. (2021) 12:5117. doi: 10.1038/s41467-021-25351-7
133. Adib A, Zhu WP, Ahmad MO. Age classification based on ECG QRS wave using deep learning. In: 2022 International Conference on Electrical, Computer and Energy Technologies (ICECET) (2022). p. 1–6.
134. Cabra Lopez JL, Parra C, Forero G. A fast deep learning ECG sex identifier based on wavelet RGB image classification. Data. (2023) 8:97. doi: 10.3390/data8060097
135. Kaneko I, Hayano J, Yuda E. How can gender be identified from heart rate data? Evaluation using allstar heart rate variability big data analysis. BMC Res Notes. (2023) 16:5. doi: 10.1186/s13104-022-06270-2
136. Naser JA, Lopez-Jimenez F, Chang AY, Baez-Suarez A, Attia ZI, Pislaru SV, et al.. Artificial intelligence-augmented electrocardiogram in determining sex: correlation with sex hormone levels. In: Mayo Clinic Proceedings. Philadelphia, PA: Elsevier (2023). Vol. 98. p. 541–8.
137. Dias FM, Ribeiro E, Soares Q, Krieger J, Gutierrez MA. Predicting age and sex from reduced lead electrocardiograms using deep learning. Authorea Preprints (2023).
138. Wei L, McHugh JC, Mooney C. A machine learning approach for sex and age classification of paediatric EEGs. In: 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC). (2023). p. 1–4.
Keywords: data augmentation, deep learning, radar-based cardiac signal, privacy, sex and age estimation
Citation: Foronda-Pascual D, Camara C and Peris-Lopez P (2025) Sex and age estimation from cardiac signals captured via radar using data augmentation and deep learning: a privacy concern. Front. Digit. Health 7:1616770. doi: 10.3389/fdgth.2025.1616770
Received: 23 April 2025; Accepted: 1 October 2025;
Published: 14 November 2025.
Edited by:
Xiaofeng Li, Heilongjiang International University, ChinaCopyright: © 2025 Foronda-Pascual, Camara and Peris-Lopez. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Daniel Foronda-Pascual, ZGFuaWVsLmZvcm9uZGFAdWMzbS5lcw==
Pedro Peris-Lopez