Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Digit. Health, 30 September 2025

Sec. Health Informatics

Volume 7 - 2025 | https://doi.org/10.3389/fdgth.2025.1616334

This article is part of the Research TopicAI in Healthcare: Transforming Clinical Risk Prediction, Medical Large Language Models, and BeyondView all 10 articles

MYSTETH—home-based heart monitoring

  • 1Department of Electrical Engineering, Indian Institute of Technology, Kharagpur, West Bengal, India
  • 2Department of Computer Science and Engineering, Netaji Subhas University of Technology, New Delhi, India
  • 3Faculty of Logistics, Molde University College, Molde, Norway

Introduction: The MySteth is an intelligent medical tool designed for cardiac disease screening, utilizing either a stethoscope or smartphone to record heart sounds. Normal heart sounds in healthy individuals consist of “lub” and “dub” noises, while murmurs—additional sounds during heartbeats—can indicate cardiac anomalies such as valve dysfunctions and rapid blood flow, categorized as systolic or diastolic.

Method: MySteth was developed and tested using heart sounds recorded via smartphone and digital stethoscope. For ensuring the clinical validity of the data, all heart sound samples were meticulously annotated by human experts—super-specialized cardiologists with extensive experience in cardiac diagnostics. To achieve high classification accuracy, MySteth employs a hybrid CNN-LSTM model combined with Linear Predictive Coding (LPC) for preprocessing. The study involves classifying recorded heart sounds into normal heartbeats and murmurs, with murmurs further divided into systolic and diastolic categories.

Results: The tool demonstrated an accuracy of 92% in distinguishing normal heartbeats from murmurs, 91% in classifying murmurs into systolic and diastolic types, and 90% in further categorizing systolic murmurs into Ejection Systolic Murmurs (ESM) and Pansystolic Murmurs (PSM). MySteth is accessible and affordable, requiring minimal equipment, as most individuals already own a smartphone, and digital stethoscopes are commonly available. This ease of use facilitates both professional and home-based heart monitoring, especially beneficial in remote areas with limited healthcare access.

Discussion: MySteth is an at-home heart diagnostic tool that leverages deep learning to classify heart sounds into normal, ESM, PSM, and diastolic murmurs. Its user-friendly design and minimal hardware requirements ensure broad adoption across various healthcare settings, facilitating timely and accurate preliminary heart investigations. This capability is crucial in combating the global burden of cardiovascular diseases. MySteth's scalability and ease of deployment underscore its potential in early cardiovascular disease diagnosis, particularly in underserved regions, thereby promoting preventive healthcare.

1 Introduction

The two typical heart sounds in healthy people are a lub and a dub, which happen one after the other with each beating. It's common to refer to the lub as the first heart sound (S1) and a dub as the second heart sound (S2). Additional noises are heard in regular heart sounds (HS), which can be used in pathology diagnosis in circumstances when the heart is aberrant, such as valve dysfunctions and fast blood flow (1). These extra noises, sometimes referred to as murmurs, exhibit distinct traits in relation to heart valve problems, which are circulatory heart illnesses (2). The most common way to categorize cardiac murmurs is by timing; they can be classified as either systolic (3) or diastolic (4), depending on which portion of the heartbeat they occur during.

Murmurs of the heart that are audible during systole are known as systolic murmurs. The most common systolic murmur (5):

1. Ejection-systolic murmurs (ESM): Diamond-shaped or spindle-shaped. The intensity first increases and then decreases during S1.

2. Pansystolic murmurs (PSM): Rectangular shaped. The intensity remains constant during S1.

The murmur heard in the heart during diastole is called diastolic heart murmur. Diastolic murmurs end at or before S1 and begin at or after S2 (6).


Heart murmurs are a problem that affects a large percentage of people worldwide. These murmurs might be an indicator of underlying cardiovascular disorders including valve dysfunctions. About 2.5% of Americans have heart valve disease, with the prevalence rising with age, according to the American Heart Association (7). One of the primary reasons heart murmurs are not timely diagnosed is the lack of access to regular and comprehensive cardiac evaluations, particularly in underserved and rural areas (8). Additionally, the subtle nature of some murmurs can make them difficult to detect without specialized equipment and expertise. The introduction of a home-based preliminary diagnostic tool for heart murmurs could be highly beneficial. Such a tool would enable individuals to monitor their heart sounds regularly, facilitating early detection of abnormalities and prompting timely medical consultations. This proactive approach could significantly reduce the burden of undiagnosed heart conditions, improve patient outcomes, and decrease healthcare costs associated with advanced cardiovascular diseases (9).

With 17.9 million deaths from cardiovascular diseases (CVDs) per year, or 31% of all fatalities globally, CVD is a major public health concern (10). Early detection is key since cardiac disorders can worsen over time and necessitate more involved forms of care. For instance, coronary heart disease, one of the most common cardiac conditions in the United States, can worsen over time and eventually necessitate coronary artery bypass grafting (CABG) (11). Preventative detection of heart diseases is essential, and medical professionals often start by checking the patient's heartbeat and abnormalities. Further tests, such as blood pressure and fasting protein profile tests, are then performed for further analysis (12). Currently, there is no easy method for heart screening at home without specialized medical personnel. Heart health monitoring and the availability of at-home testing options is crucial for promoting heart health awareness. At-home diagnostics can significantly contribute to heart health promotion and better outcomes for those at risk of cardiovascular problems by enabling individuals to adopt proactive measures towards lowering their risk of heart disease (13).

Numerous studies have employed machine learning and deep learning techniques to categorize heartbeat sounds; most of these studies have focused on data from phonocardiography (PCG), a specialized device used for medical diagnostics (1419). However, this technology is not accessible to the average consumer and cannot be performed at home.

Advancements in technology have led to smartphone applications like SensiCardiac (20), Mobile Stethoscope (21), and iStethoscope Pro (22), which allow heart sounds to be conveniently recorded. Some studies have also used Electrocardiogram (ECG) signals, which are obtained from specialized instruments in medical facilities. Some studies have also used audio from electronic stethoscopes and mobile phones (23). Only a few categories have been used to categorize murmurs: artifact, extra-heart sound, extrasystole, murmur, normal heartbeat, moderate, severe, or normal, aberrant. There has never been an attempt to further categorize murmurs into systolic, diastolic, and systolic murmurs as well as ESM and PSM. Because there aren't enough datasets available, the majority of these studies have limitations (24). The models are trained and validated on specific datasets which may not encompass the full variability seen in global populations. Without prior patient information, other classifications of murmurs—such as mitral valve prolapse, mitral regurgitation, and aortic stenosis—cannot be made. These classifications require further tests such as ECG, ultrasound, and cardiac CT (25). The key to reducing healthcare costs from CVD and increasing patient outcomes lies in early detection, prevention, and access to quality health services (26). Unfortunately, emergency rooms and hospitals are overcrowded, while affordable healthcare clinics are scarce. This created the need for the development of in-home health monitoring and CVD management programs (27). Early detection and prevention are crucial because CVD accounts for 17.9 million deaths yearly (28). Medical technology has advanced, but there are still no easily available, user-friendly techniques for doing at-home cardiac screenings (29). This study looks at how Mysteth uses digital stethoscope technology and deep learning methods to offer a quick and easy way to do initial cardiac investigations. The goal is to raise awareness of heart health issues and improve the lives of people who are at risk of CVDs.

This work presents MySteth as an innovative at-home heart diagnostic tool designed to bridge the gap in care by providing a convenient and accessible solution for preliminary heart investigations. While other heart testing options exist, MySteth offers distinct advantages. It is the first screening method of its kind to use deep learning techniques and recorded heartbeat sounds to detect a wide range of heart diseases using a smartphone or digital stethoscope. By employing deep learning, MySteth can classify heart murmurs with greater granularity, distinguishing between various types such as systolic, diastolic, Ejection Systolic Murmurs (ESM), and Pansystolic Murmurs (PSM), which have not been extensively categorized in previous studies. This technique effectively detects prevalent valvular heart conditions, including arrhythmia, mitral regurgitation, and coronary heart disease, at home. By leveraging widely available smartphones and digital stethoscopes, MySteth enhances cardiac diagnostics with precise, real-time analysis that is both accessible and cost-effective, marking a significant advancement in the field.

Recent innovations in wearable diagnostics, such as triboelectric sensors for arteriovenous fistula (AVF) monitoring in hemodialysis patients, highlight the growing feasibility of compact, non-invasive devices for continuous cardiovascular assessment (30). These systems, which utilize triboelectric impedance cardiography (T-ICG) to detect vascular abnormalities like stenosis through changes in signal morphology at key cardiac cycle points, underscore the clinical relevance of acoustic and impedance-based monitoring techniques. Inspired by such developments, our work aims to explore whether similar diagnostic precision can be achieved using more ubiquitous technology, namely, smartphones paired with digital stethoscopes and deep learning models. By enabling at-home screening of cardiac conditions through familiar devices, our approach complements and extends the paradigm of accessible, portable monitoring tools, especially in contexts where specialized equipment and expertise are limited.

The work by Galli et al. (31), which presents a portable, non-invasive ventilation (NIV) system for home and clinical use, offers a valuable reference model. Their device integrates airflow generation with pressure monitoring and remote smartphone-based data transmission, underscoring the importance of user-centered design and technological robustness in remote healthcare devices. For instance, recent advancements in wearable systems for arteriovenous (AV) fistula monitoring in dialysis patients have demonstrated the clinical feasibility and diagnostic value of portable acoustic sensing platforms (32). These systems use similar principles, capturing and analysing vascular sounds, to detect abnormalities such as stenosis, showcasing the real-world applicability of non-invasive auscultatory tools. Building upon such approaches, the MySteth system investigates whether commonly available devices like smartphones, when paired with digital stethoscopes and advanced deep learning models, can replicate and eventually democratize similar diagnostic capabilities.

2 Methods

In our work heartbeat is divided into two categories: murmurs, and normal heart sounds. We next divide the murmurs into systolic and diastolic murmurs. We further classify systolic murmurs into PSM and ESM. We don't need further categorization of diastolic murmurs, as most of the murmurs in this category are pathologic in nature and hence severe (33). The categorization shown in Figure 1 is the one identified by most of the doctors when they first examine a patient using a stethoscope. It is good enough to manifest evidence for a variety of heart disease. This procedure involves three classification steps to progressively refine the detection and categorization of heart sounds.

Figure 1
Flowchart outlining heart sounds classification into normal and murmur types, with murmurs further divided into systolic and diastolic phases, leading to identification of PSM (pan-systolic murmur) and ESM (ejection systolic murmur) patterns aligned with LUB-DUB cycles.

Figure 1. Classification of heart sounds into normal and murmurs, murmurs into systolic and diastolic murmurs, and further classifications to ESM and PSM.

The complete procedure used to perform the classification in Mysteth is explained into two main parts: Data Preprocessing and the MySteth Architecture.

2.1 Data collection, labelling, preprocessing, refining, and data synthesis

This part includes the steps shown in Figure 2, as explained below, which are applied on the original dataset to build a suitable Neural Network Model.

1. Data Collection: The authors used a publicly available Kaggle dataset (https://www.kaggle.com/kinguistics/heartbeat-sounds) to identify murmurs in heartbeat sound audios. The dataset contains 832 distinct heartbeats, of which 480 audios were selected for the use case. This dataset was gathered from the general public via the stethoscope Pro iPhone app and a clinic trial in hospitals using the digital stethoscope, DigiScope. In the original publicly available dataset of 832 heartbeat recordings, the distribution of samples was heavily skewed toward normal heart sounds, with murmurs forming a smaller proportion (only 129 audios). To avoid introducing bias from this imbalance, we selected 480 recordings that included 351 normal heartbeats and 129 murmurs, ensuring that both categories were adequately represented in the initial training set. This balanced selection was essential to prevent the model from underperforming on pathological cases.

2. Data Labelling: The publicly available dataset (https://www.kaggle.com/kinguistics/heartbeat-sounds) was annotated by Dr. Nishant Thakur, a super-specialized cardiologist from Max Hospital, I.P. Extension, Delhi, India, and re-annotated and cross-checked by Dr. Rajat Jain, a super-specialized cardiologist from Safdarjung Hospital, Delhi, India. Since the dataset is publicly available so no ethical approvals were required.

3. Audio Processing and Refining: Raw audios, sampled at 22050Hz, were down sampled to 4 kHz. This down sampling reduces computational load and storage requirements while retaining essential information for heartbeat analysis (34). Only the first 3 s of each audio were preserved to capture a complete cardiac cycle (S1 to S2 to S1), ensuring that the analysis encompasses all critical heart sounds. Audios shorter than 3 s were repeated to reach or exceed the 3-second duration, maintaining consistency in input length for the model. The study transformed audio signals into numerical data through the extraction of distinct features representative of signal characteristics, including amplitude, frequency, and duration, using the librosa library. Librosa is a widely-used Python library for audio analysis, known for its robust feature extraction capabilities, which facilitate effective signal characterization for subsequent classification (35).

4. Data Synthesis: Given the small initial dataset, Gaussian Mixture Models (GMM) were used to increase the dataset size to 10,000 audio vectors. This approach is beneficial as GMMs can generate new, realistic data points by modelling the probability distribution of the existing data, thus enhancing the dataset without additional data collection efforts (36). Out of the 10,000 vectors, 3,600 were murmurs, out of which 730 were systolic murmurs. GMMs were used again to increase the number of audio vectors represented by systolic murmurs to a size of 5,000. This targeted augmentation ensures that the dataset is well-balanced, particularly for the systolic murmur class, which is crucial for training a robust and unbiased classification model (37).


Figure 2
Flowchart depicting the processing pipeline for heart sound classification, starting with audio files from Kaggle and clinical trials, followed by vector conversion, synthesis using Gaussian Mixture Models, feature extraction, and sequential classification using CNN-LSTM models into murmur types (systolic or diastolic) and further into ESM (ejection systolic murmur) or PSM (pan-systolic murmur).

Figure 2. Preliminary steps for data preprocessing on the original dataset of 480 audio files selected. Created using Canva, licensed under Free Content License.

Specifically, the distribution of murmurs and normal heartbeats in the synthetic samples closely mirrors that of the original dataset, which consisted of 129 murmurs and 351 normal heartbeats recorded from a limited number of subjects—approximately 480 in total. Hence, it can be concluded that the total number of subjects for whom the data has been generated artificially is 9,520. The synthetic data was generated in a way that maintains subject-level diversity while amplifying underrepresented classes, particularly systolic murmurs. This alignment indicates that the GMM-based augmentation has effectively preserved the statistical properties and variability of the original dataset.

The use of data augmentation in scenarios with limited real-world samples is well-supported in literature; for instance, Frid-Adar et al. (38) demonstrated in synthetic data can significantly improve model performance when real data is scarce. Similarly, the targeted augmentation used here ensures the generation of high-quality, representative data, thereby enhancing model generalizability while reducing potential class imbalance. The proportion of heartbeats and murmurs, as well as its granular classifications in the generated dataset are similar to their proportions in the original dataset, thus suggesting that the data synthesis is appropriate and can be used for further experiments.

5. Model Training: Various models were trained on the refined datasets obtained from each of the following classification tasks. The train test ratios for all tasks were kept constant at a 70–30 percent split:

a. Classification Task 1: Applied on the original dataset to separate the heartbeat sounds into normal heartbeats and murmurs

b. Classification Task 2: Applied on the Murmurs obtained from classification task 1 to obtain systolic and diastolic murmurs

c. Classification Task 3: Applied on the Systolic Murmurs obtained from classification task 2 to divide them into Pansystolic Murmurs (PSM) and Ejection Systolic Murmurs (ESM)


Various neural network models were applied on the dataset to obtain the best possible results:

1. CNN-LSTM:

The CNN-LSTM architecture was chosen due to its ability to effectively combine both spatial and temporal feature extraction, which is particularly important for the classification of heartbeat audio signals. CNNs can reduce noise by focusing on important features through convolutional filters, which makes the subsequent LSTM layers more effective in learning the temporal dependencies of the cleaned signal (39). Details of the models are as follows:

a. Input Layer: Processed numerical data representing the heartbeat audio signals.

b. Intermediate CNN and LSTM Layers, shown in Table 1.

c. Output Layer: SoftMax activation function to classify the audio signals into categories (e.g., normal heartbeat, murmur).


Table 1
www.frontiersin.org

Table 1. Model architectures employed for training and testing on the audio vector and feature dataset.

Furthermore, regularization methods were implemented in CNN-LSTM architecture to mitigate overfitting risks. A dropout of 20% was used in the LSTM layer, and the model was kept simple with a total of 5 layers.

2. BiLSTM: BiLSTMs have been successfully applied to various medical signal classification tasks, including ECG and phonocardiography (PCG) signals. Their effectiveness in capturing the temporal dynamics and dependencies in such data makes them a reliable choice for heartbeat classification (40). The details of the model are as follows:

a. Input Layer: Processed numerical data representing the heartbeat audio signals.

b. Intermediate BiLSTM and Dense Layers, shown in Table 1.

c. Output Layer of Size 2 Units: SoftMax activation to classify the audio signals into categories (e.g., systolic murmur, diastolic murmur).

3. CNN: Heartbeat signals can exhibit significant variability in both time and frequency domains. CNNs, with their ability to apply convolutional filters across the input signal, can robustly handle such variations and capture essential characteristics of the heartbeat patterns (41). The details of the model are as follows:

a. Input Layer: Processed numerical data representing the heartbeat audio signals.

b. Intermediate CNN Layers, shown in Table 1.

c. Output Layer: SoftMax activation function to provide classification probabilities. (e.g., Normal Heartbeat, Murmur)

4. BiGRU: Heartbeat signals are sequential in nature, and it is crucial to capture the temporal dependencies within the data. BiGRUs can process the input in both forward and backward directions, capturing dependencies from both past and future contexts, which is particularly beneficial for heartbeat classification (42). The details of the models are as follows:

a. Input Layer: Processed numerical data representing the heartbeat audio signals.

b. Intermediate BiLSTM and Dense Layers, shown in Table 1.

c. Output Layer of Size 2 Units: SoftMax activation to classify the audio signals into categories (e.g., systolic murmur, diastolic murmur).

The results obtained from the models for each of the classification are articulated in the Results section. As the CNN-LSTM model gave the best accuracy scores, it was employed in the MySteth Architecture. This was followed by a SHAP (SHapley Additive exPlanations) on 10 test cases.

2.2 Mysteth architecture

The MySteth Architecture was designed to handle the refined datasets and perform the classifications at each step. The following steps, shown in Figure 3, are part of the processing pipeline:

1. Recording: A person records their heartbeat using a smartphone in a silent environment, capturing a 3-second audio clip.

2. Down sampling and Compression: Linear Predictive Coding (LPC) facilitates a two-step procedure of down sampling and compression of the recorded audio before training the models. By lowering the audio's sampling rate, down sampling effectively minimizes the amount of data and computing load while preserving crucial information. The spectral envelope of the digital voice signals is then compressed using LPC compression. By preserving important components of the heart sounds, this approach improves the efficacy of feature extraction (43). LPC minimizes the amount of data while preserving important information, which makes processing and analyzing the cardiac sounds simpler and quicker. In order to ensure that deep learning models can effectively capture and learn from the key elements of the heart sounds throughout the training phase, this preliminary step optimizes the data for the models (44). The integration of LPC for down sampling and compression in the preprocessing pipeline optimizes the data for deep learning models (45).

3. Heartbeat Signal Verification: The extracted features are compared to a reference model of a normal “lub-dub” heartbeat. The patient is asked to re-record the heartbeat if the signal does not match the classic “lub-dub” pattern.

4. Model Training: The audio vector and the extracted features passes through a pre-trained CNN-LSTM model that has shown the best results (as tabulated in Tables 13). This model had been trained previously on huge amounts of data (as explained in Part A). CNN and LSTM were applied serially due to the complementary nature of their roles in feature extraction and sequence modelling. The CNNs can preprocess and distil the essential features, which the LSTMs can then analyze in a temporal context (46).

a. Static feature extraction is done by Convolutional Neural Networks (CNN).

b. Temporal characteristics are extracted using Long Short-Term Memory (LSTM).

c. Three convolution layers with kernel sizes of 9, 64, and 32 are applied, followed by batch normalization after each convolution layer. The details of the layers are given in Part A.

d. For LSTM-based models, two layers of sizes 8 and 4 are added, followed by a dense layer. The details of the layers are given in Part A.

5. Classification Task 1: Two categories are created from the processed audio: murmur and normal heartbeat.

6. Classification Task 2: Murmurs are further divided into diastolic and systolic forms.

7. Classification Task 3: Systolic Murmurs are classified into Pansystolic Murmur (PSM) and Ejection Systolic Murmur (ESM).

Figure 3
Flowchart illustrating a mobile-based heart sound classification system. A three-second audio recording undergoes vector conversion, downsampling, and linear predictive coding, followed by comparison with the standard

Figure 3. Mysteth architecture combining CNNs and LSTMs to classify heartbeats into its granular levels. Created using Canva, licensed under Free Content License.

Table 2
www.frontiersin.org

Table 2. Evaluation metrics for different models for classifying heart sounds into normal, and murmur.

Table 3
www.frontiersin.org

Table 3. Evaluation metrics for different models for classifying murmurs into systolic murmur and diastolic murmur.

3 Results

This section presents the results obtained from classifying heartbeat sounds using different models (as explained in Part A of the Methods section) into the following categories: Normal heartbeat, Murmurs, Systolic Murmur, Diastolic Murmur, PSM (Pansystolic Murmur), and ESM (Ejection Systolic Murmur).

1. Classification results for Normal Heartbeat, and Murmur: The classification accuracy for Normal Heartbeat and Murmurs was evaluated using different models on the compressed audio representations obtained through Linear Predictive Coding (LPC). An overview of the outcomes, shown in Table 2, the Receiver Operating Characteristic Curve for the classification using the CNN-LSTM Hybrid Model, shown in Figure 4 justify the selection of a CNN-LSTM model.


Figure 4
Receiver Operating Characteristic (ROC) curve showing model performance with true positive rate plotted against false positive rate. The orange curve represents the model with an Area Under the Curve (AUC) of 0.93, indicating high classification accuracy. A diagonal dashed line represents random guessing.

Figure 4. ROC curve for classification of heart sounds into normal and murmur using the CNN-LSTM model.

The classification of heartbeat sounds is an essential task in the medical field, as it helps healthcare professionals diagnose various cardiovascular conditions. Although numerous research activities have been carried out to enhance the precision of heartbeat sound categorization, the majority of them have concentrated on refining data pre-processing methods or employing a single primary method such as neural networks, support vector machines, or hidden Markov models (47).

The impact of each feature on the model, shown in Figure 5, to classify the heartbeat sounds into normal heartbeats or murmurs depicts that the MFCCs, specifically mfcc6, mfcc11, mfcc3 and mfcc8 have a considerable amount of weightage while making the predictions. It can also be concluded that the values of these features directly affect their SHAP values. For example, a high mfcc6 gives a highly positive SHAP value when predicting normal heartbeats, and a highly negative SHAP value when predicting murmurs. A similar trend is observed for almost all features. It is also noted that the features like chroma STFT, spectral bandwidth, RMSE and zero crossing rate have least impact on the model outputs.

2. Classification results for systolic murmur, and diastolic murmur: The classification accuracy for distinguishing between Systolic and Diastolic Murmurs was assessed using various models on the compressed representations of the murmur audio. The results, shown in Table 3, Receiver Operating Characteristic Curve for the classification using the CNN-LSTM Hybrid Model, shown in Figure 6 depict the efficiency of each model.


Figure 5
Two SHAP summary plots comparing feature importance for two different model outputs. Both plots rank features like mfcc6, mfcc11, and mfcc3 as highly influential, with color gradients representing feature values from low (blue) to high (red) and horizontal spread indicating SHAP value impact on model predictions. Left plot shows impact up to 0.15, while right plot extends to 0.2.

Figure 5. SHAP explanations for classifying heartbeat sounds into (a) normal heartbeat and (b) murmur.

Figure 6
Receiver Operating Characteristic (ROC) curve displaying model performance, with the orange curve indicating an AUC of 0.83. The true positive rate is plotted against the false positive rate, and the diagonal dashed line represents random classification.

Figure 6. ROC curve for classification of murmurs into systolic murmurs and diastolic murmurs using the CNN-LSTM hybrid model.

A hybrid classifier can significantly enhance classification accuracy. When combined in a hybrid CNN-LSTM model, these models can effectively extract deep features and contextual time data from Phonocardiogram (PCG) signals. The CNN component handles feature extraction, while the LSTM module extracts time-dependent features (48).

The impact of each feature on the model to classify the murmurs into systolic or diastolic murmurs, shown in Figure 7 depicts that the audio features like Rolloff, spectral centroid and spectral bandwidth have a considerable amount of weightage while making the predictions, which is contrary to the observation in the previous classification step. Among the MFCCs, mfcc2 seems to affect the model output the most. The values of these features directly affect their SHAP values. It is also noted that the features like chroma STFT, RMSE and zero crossing rate have least impact on the model outputs, just as in the previous classification step.

3. Classification results for PSM, and ESM: The accuracy for further classifying systolic murmurs into PSM and ESM was evaluated using various models on the compressed feature representations obtained from LPC feature extraction method. The outcomes, shown in Table 4, the Receiver Operating Characteristic Curve for the classification using the CNN-LSTM Hybrid Model, shown in Figure 8 justify the utility of the CNN-LSTM model.


Figure 7
Two SHAP summary plots displaying the impact of audio features on model outputs. Both plots show rolloff, spectral centroid, and spectral bandwidth as the top influential features. The color gradient from blue to red represents low to high feature values, and the horizontal axis indicates SHAP value impact. The left plot shows a wider positive SHAP range, while the right plot emphasizes negative contributions.

Figure 7. SHAP explanations for classifying murmurs into (a) systolic murmur and (b) diastolic murmur.

Table 4
www.frontiersin.org

Table 4. Evaluation metrics for different models for classifying systolic murmurs further into PSM, ESM.

Figure 8
Receiver Operating Characteristic (ROC) curve showing a model's classification performance, with true positive rate plotted against false positive rate. The orange line represents the model's ROC curve with an Area Under the Curve (AUC) of 0.70, indicating moderate predictive accuracy, while the dashed diagonal line represents random performance.

Figure 8. ROC curve for classification of diastolic murmurs into ESM and PSM using the CNN-LSTM hybrid model.

This hybrid approach has been shown to outperform single CNN or LSTM-based methods, producing richer and more concentrated models with higher performance and fewer parameters. These findings demonstrate how well different recurrent neural networks function in conjunction with convolutional neural networks to tackle challenging audio categorization problems. The utilization of LPC for feature extraction significantly contributes to the models' performance, especially in distinguishing subtle differences in heart sounds (49).

The impact of each feature on the model to classify the systolic murmurs into ESM or PSM, shown in Figure 9, depicts that the audio features like Rolloff, spectral centroid and spectral bandwidth have a considerable amount of weightage while making the predictions, which is similar to the observation in the previous classification step. Among the MFCCs, mfcc3 seems to affect the model output the most. At this classification step, the values of these features do not affect their SHAP values, which is contrary to the previous two classification steps. It is also noted that the features like chroma STFT, RMSE and zero crossing rate have least impact on the model outputs, just as in the previous classification steps.

Figure 9
Two SHAP summary plots showing the impact of audio features on model outputs. In both plots, spectral centroid, spectral bandwidth, and rolloff appear as top contributors, followed by various MFCC coefficients. Each dot represents a SHAP value colored by feature value from low (blue) to high (red), with horizontal spread indicating influence on model predictions. The left plot shows moderate positive and negative impacts, while the right plot displays a wider SHAP value range, indicating stronger feature effects.

Figure 9. SHAP explanations for classifying diastolic murmurs into (a) ESM and (b) PSM.

The statistical analysis, shown in Table 5, depicts the probable accuracy, precision and recall values obtained in a 95% confidence interval. These values are similar to the values obtained on the testing dataset, thus strengthening our results. Furthermore, statistical significance testing has been done against a random classifier. The null hypothesis states that the random classifier would be just as good as the model presented in the paper. The accuracy statistic is not very close, but the p-values are negligible for all classification tasks, signifying that our model performs much better than a random classifier baseline, hence disproving the null hypothesis. The confusion matrices for each classification task, shown in Figure 10, are representative of the metrics obtained. These have been reported for the testing dataset (whose size was around 3,000 for the initial, 1,080 for the intermediate, and 1,500 for the final classification tasks). A very important observation is that the number of heartbeats that are actually murmurs but classified as normal are very low [76 out of 3,000, as can be seen in Figure 10(a)]. This reiterated the fact that MySteth presents a reliable algorithm to be used for home-based heart health monitoring.

Table 5
www.frontiersin.org

Table 5. Confidence intervals for evaluation metrics, and statistical significance testing for the model.

Figure 10
Three confusion matrices illustrating performance at different stages of a heart sound classification model. (a) Initial classification shows high accuracy in distinguishing normal (1549) from murmur (1004) sounds, with some misclassifications. (b) Intermediate classification distinguishes diastolic (823) from systolic (153) murmurs with moderate accuracy. (c) Final classification differentiates ESM (666) from PSM (287) with a notable number of classification errors.

Figure 10. Confusion matrices obtained for the (a) initial, (b) intermediate, and (c) final classification tasks.

In conclusion, it has been established that the CNN-LSTM model out-performs the rest of the models when it is desired to classify heartbeats into higher granularity levels, while also giving insights into an interpretable decision-making process in order the predict the outcome at each classification step. While lightweight models such as shallow CNNs or simple feedforward networks may offer faster inference times, they lack the capacity to capture complex temporal patterns. Hence, the proposed CNN-LSTM model seems to be a good-fit in order to perform accurate classification while gaining the trust of medical professionals. This makes it a useful aspect of a home monitoring device, while decreasing the load on doctors.

4 Discussion

The introduction of MySteth as an innovative at-home heart diagnostic tool represents an advancement in the field of cardiac care, addressing critical gaps in the accessibility and convenience of preliminary heart investigations. This discussion focuses on the unique aspects and justifications for our approach, emphasizing the integration of CNN-LSTM architectures with Linear Predictive Coding (LPC) preprocessing, and the impact of these choices on the efficacy and practicality of MySteth.

The primary motivation for employing a hybrid CNN-LSTM model stems from the complementary strengths of Convolutional Neural Networks (CNNs) and Long Short-Term Memory networks (LSTMs) in handling the complexities of heartbeat sound classification (50). CNNs are adept at extracting spatial features from the input data, capturing local patterns and significant characteristics of the heart sounds. This capability is crucial for identifying the nuanced features present in heartbeat signals, such as murmurs and other anomalies. LSTMs, on the other hand, excel at modelling temporal dependencies and sequential patterns within the data. By integrating LSTMs with CNNs, we ensure that the model not only recognizes spatial features but also understands how these features evolve over time (51). This combination is particularly effective for analyzing heartbeat sounds, which inherently possess both spatial and temporal dimensions.

Linear Predictive Coding (LPC) plays a pivotal role in our approach by facilitating data compression and enhancing feature extraction. LPC reduces the complexity of the raw audio data while preserving essential information, making the subsequent processing by CNN and LSTM layers more efficient. This preprocessing step is crucial for improving the model's ability to detect subtle patterns and anomalies in the cardiac sounds, thereby enhancing classification accuracy (52). By incorporating LPC, we address the challenge of high data volume and computational load, enabling the use of advanced deep learning models even in resource-constrained environments. This efficiency is particularly beneficial for at-home diagnostic tools like MySteth, where minimal hardware requirements and quick processing are critical for user adoption and practicality (53).

Compared to conventional methods of heartbeat classification, which often rely on manual feature extraction and traditional machine learning algorithms, our CNN-LSTM approach offers several distinct advantages. Traditional methods can be limited by their dependency on handcrafted features and their inability to fully capture the complexity of the heartbeat signals. In contrast, deep learning models, particularly the CNN-LSTM combination, automatically learn relevant features from the data, leading to more accurate and robust classifications (54). Moreover, the ability to handle large and complex datasets without significant manual intervention makes our approach more scalable and adaptable to different healthcare settings. High accuracy rates have been achieved by our models, including the exceptional performance of the CNN-LSTM model with a 92% accuracy for classifying the heartbeats into normal and murmurs, and 91% for classifying the murmurs into systolic and diastolic murmurs. This underscores the effectiveness of our method in differentiating between normal and pathological heart sounds as well as finer distinctions such as various types of murmurs.

Unlike the recently developed hardware-centric designs (3032), MySteth presents an end-to-end signal-processing using a deep learning model for heart sounds to perform segmentation and classification. This enables deployment on commodity devices and potentially broadening access where specialized sensors like PVDF triboelectric (30, 32) or NIV turbines (31) are impractical. Patient-cohort validations like B-/C-point slopes vs. stenosis (30, 32), and device-level performance envelopes (31) demonstrate physiologically interpretable trends. MySteth currently depends on SHAP explanations to justify the results of its analysis. Triboelectric ICG features (30, 32) have direct hemodynamic interpretations. However, MySteth uses algorithmic acoustic features, with less explicit references to mechanistic markers, such as the timing or the severity correlation, which may be a matter of concern for improving clinician trust to take triage decisions. These systems push the frontier in hemodialysis vascular monitoring (30, 32) and home respiratory therapy (31). MySteth advances acoustic cardiac screening on accessible hardware, that is a simple smartphone, thus filling a different but clinically adjacent niche that emphasizes software-driven auscultation and potential population-scale reach.

This work represents a pilot study aimed at evaluating the feasibility and effectiveness of deep learning–based heartbeat classification, and the results so far have been encouraging. The CNN-LSTM model demonstrated high accuracy in classifying normal and pathological heart sounds under controlled conditions, indicating strong potential for clinical relevance. While clinical validation has not yet been initiated, it is a key focus of our future work. While the manuscript does not explicitly analyze the impact of domestic ambient noise on system performance, it is important to note that domestic sounds—such as speech, appliances, or external traffic, have acoustic characteristics that are significantly different from those of heart sounds. These can be integrated as a preprocessing layer into the overall classification pipeline to improve robustness in real-world environments. Prior studies have demonstrated that heart sounds possess distinct acoustic characteristics, such as temporal regularity and specific frequency ranges, that differentiate them from other bodily sounds, making them suitable candidates for signal separation and denoising. Springer et al. (55) emphasized the effectiveness of frequency and envelope-based methods for isolating heart sounds from noise.

The current evaluation has been performed entirely on controlled and expert-annotated data. While this setup ensures high-quality ground truth for model development and benchmarking, it does not fully capture the variability, and recording artifacts present in real-world environments. Translation to home and clinical utility requires systematic prospective validation and robust noise-adaptation strategies. A prospective validation study can be conducted on diverse patient populations across different age groups, cardiac conditions, and recording environments. This study could include recordings from multiple devices (smartphones, digital stethoscopes) to assess hardware variability. Further, annotation of new data can be done correctly by multiple expert cardiologists to ensure reliable ground truth. Stratified sampling can also be done to capture a balanced representation of normal heart sounds and pathological murmurs, including rare subtypes.

The minimal hardware requirements and straightforward implementation of MySteth mean it can be readily adopted in various healthcare environments, including remote or under-resourced areas. This accessibility addresses a critical need in global healthcare, providing reliable and early detection tools for heart disease, which remains a leading cause of mortality worldwide (56). This paper distinguishes itself from previous research by introducing MySteth, a novel home-based heart monitoring tool that utilizes deep learning techniques to classify heart sounds with enhanced accuracy and detail. Unlike earlier studies that primarily concentrated on phonocardiography or ECG data—methods not easily accessible for home use—MySteth employs commonly available devices like smartphones and digital stethoscopes. It extends beyond basic heart sound classification by differentiating between normal heartbeats, murmurs, and specific subtypes such as Ejection Systolic Murmurs (ESM) and Pansystolic Murmurs (PSM). This level of granularity, particularly in home settings, has not been achieved by previous studies. The use of deep learning models for more precise and real-time analysis, combined with its accessibility and cost-effectiveness, marks this paper as a significant advancement in cardiac diagnostics over prior research.

This study focuses primarily on the development and evaluation of a pre-trained deep learning model for heartbeat classification, which can be used directly for real-time heartbeat classification. As the model is designed to be used as a pre-trained solution, the end user is not required to perform on-device training or intensive computation locally, hence nullifying the requirement of strong hardware for the same. Therefore, issues related to power consumption, battery life, and energy optimization strategies were not within the scope of this work, and need not be a cause of concern for the end-user. Finally, it is also noteworthy that the model predictions provide a good amount of interpretability using the SHAP values and help medical professionals gauge a better insight into the decision-making process, thus functioning as a helpful home heart-screening device for patients as well as a useful understanding of the prediction to the medical practitioners.

5 Future work

While our work presents significant advancements, there are limitations and areas for improvement. As already stated, not much study has been done in the area of categorizing mobile phone heartbeat sounds (57). To improve research in this area, a larger and more realistic dataset must be created (58). The models are trained and validated on specific datasets which may not encompass the full variability seen in global populations. Future work should focus on incorporating more diverse datasets to enhance generalizability. In this work, the audio was encoded using linear predictive coding. To compress audio, more encoding methods can be employed, such as auto-encoders. While our method is efficient, optimizing it further for real-time processing and deployment on portable devices could enhance its practical application. Future research should explore seamless integration with existing clinical workflows, ensuring that the technology is user-friendly for healthcare professionals. To confirm the models' long-term dependability and efficacy in practical situations, longitudinal research and comprehensive clinical trials are required (59).

The current study utilized uncleaned data recorded from smartphones, which included a significant amount of noise due to breathing artifacts. These artifacts can adversely affect the accuracy of classification. Future research could focus on removing these artifacts to enhance the signal-to-noise ratio, thereby improving the accuracy of murmur classification. Despite the presence of breathing artifacts in the current uncleaned data, the preliminary classification accuracies achieved were promising for screening purposes. This suggests that even without artifact removal, the current results may still offer valuable diagnostic information in a clinical setting. Further investigation is required to explore the potential of this technology as a tool for non-invasive heart screening, particularly in resource-limited settings where access to traditional cardiac diagnostics is limited.

Although domestic and breathing noise are recognized as important factors in real-world deployment, the present study did not include quantitative noise robustness experiments. The preprocessing pipeline can be adapted to better handle real-world acoustic conditions by incorporating adaptive band-pass filters, wavelet-based denoising, or spectral subtraction techniques to remove ambient and breathing artifacts without compromising clinically relevant heart sounds. Training can be made more robust by training with synthetically augmented datasets that include controlled levels of domestic noise (speech, TV, traffic, appliance hums, etc.), enabling the model to generalize better to variable acoustic environments. The utility can be improved by adding a pre-classification module that evaluates signal-to-noise ratio (SNR) and prompts re-recording if background noise is too high.

The deep learning model presented is relatively simple and could be enhanced by incorporating more complex features and deeper neural network architectures. By leveraging recent advances in deep learning, such as attention mechanisms and unsupervised feature learning, the classification accuracy could be further improved.

This work represents a preliminary investigation aimed at evaluating the feasibility and effectiveness of deep learning–based heartbeat classification, and the results so far have been encouraging. The CNN-LSTM model demonstrated high accuracy in classifying normal and pathological heart sounds under controlled conditions, indicating strong potential for clinical relevance. The next step involves testing the system in real-world settings, including diverse patient populations and varied acoustic environments, such as outpatient clinics and home-based monitoring.

The current study primarily utilizes synthetic data generated through a Gaussian Mixture Model (GMM)-based augmentation approach, which was designed to preserve subject-level diversity while amplifying underrepresented classes, particularly systolic murmurs. This method ensures that the augmented dataset maintains the statistical properties and variability of the original recordings, thereby supporting robust model training. To address generalizability beyond the training data, future work will focus on collecting and evaluating data from real-world use of the MySteth prototype in diverse clinical and home environments.

In conclusion, this preliminary work lays the groundwork for future efforts to enhance heartbeat classification accuracy using smartphone recordings. Proposed future directions include removing noise artifacts from the data, employing more sophisticated deep learning models, expanding the classification scope, and collecting larger datasets. Ultimately, these advancements will refine murmur detection and improve the quality of cardiac diagnostics provided by smartphone recordings.

6 Conclusions

MySteth is a tool that our study introduces. Using deep learning algorithms and just the sound of a heartbeat recorded using a phone or digital stethoscope, the authors investigated the screening of a broad class of heart disorders. Heartbeats can be categorized by MySteth into three categories: normal, systolic, and diastolic murmurs. Systolic murmurs can also be further classified into two categories: Ejection Systolic Murmur (ESM) and Pansystolic Murmur (PSM).

In order to keep the condition from getting worse to the point where it becomes fatal or irreversible, this effort can be very helpful in identifying the onset of a wide class of cardiovascular heart diseases (60). MySteth, a tool in the field of heart sound classification, can significantly contribute to preventive healthcare and lower the total burden of cardiovascular illnesses by enabling at-home early screening and precise diagnosis of heart murmurs and other irregularities. Because of its architecture, the instrument can be deployed in a variety of contexts, such as remote and rural locations, allowing disadvantaged groups to benefit from modern diagnostics. The scalable and adaptable nature of MySteth ensures it can be integrated into different healthcare environments, from large urban hospitals to small rural clinics and even home-based care. By addressing both the technological and practical challenges in this domain, MySteth stands out as a viable solution with significant potential for improving global health outcomes.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/kinguistics/heartbeat-sounds.

Ethics statement

This study uses publicly available Kaggle dataset (https://www.kaggle.com/kinguistics/heartbeat-sounds) to identify murmurs in heartbeat sound audios. As the dataset is in the public domain, so no ethical approval was required. This dataset was de-identified, hence did not need ethical approval.

Author contributions

KJ: Methodology, Writing – review & editing, Investigation, Data curation, Formal analysis. RJ: Investigation, Validation, Data curation, Conceptualization, Writing – original draft. SM: Conceptualization, Investigation, Validation, Data curation, Writing – original draft. SA: Validation, Writing – review & editing, Supervision, Formal analysis, Conceptualization, Methodology, Investigation.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. APC will be supported by Molde University College, Norway.

Acknowledgments

The authors would like to thank Vishal Kumar Lal, Department of Biosciences and Biotechnology, Indian Institute of Technology, Kharagpur, for his assistance in creating a synthetic dataset using the Librosa library, extracting advanced audio features, data preprocessing, and applying various classical ML models.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Roy JK, Roy TS, Mukhopadhyay SC. Heart sound: detection and analytical approach towards diseases. In: Jayasundera KP, Postolache OA, editors. Modern Sensing Technologies. Cham: Springer International Publishing (2018). p. 103–45.

Google Scholar

2. Kumar D, Carvalho P, Couceiro R, Antunes M, Paiva RP, Henriques J. Heart murmur classification using complexity signatures. 2010 20th International Conference on Pattern Recognition; IEEE (2010). p. 2564–7.

Google Scholar

3. Reimer-Kent J. Heart sounds: are you listening? Part 2. Can J Cardiovasc Nurs. (2013) 23(3):3. 23984480

PubMed Abstract | Google Scholar

4. Walker HK, Hall WD, Hurst JW. Clinical Methods: The History, Physical, and Laboratory Examinations, Vol. 2. Boston: Butterworths (1976).

Google Scholar

5. Deng L, O’Shaughnessy D. Speech Processing: A Dynamic and Optimization-Oriented Approach. Lausanne: Taylor & Francis (2003).

Google Scholar

6. Patnaik AN. First heart sound. Indian J Cardiovasc Dis Women WINCARS. (2019) 4(02):107–9. doi: 10.1055/s-0039-1694856

Crossref Full Text | Google Scholar

7. Nishimura RA, Otto CM, Bonow RO, Carabello BA, Erwin JP III, Guyton RA, et al. 2014 AHA/ACC guideline for the management of patients with valvular heart disease: executive summary: a report of the American College of Cardiology/American Heart Association task force on practice guidelines. Circulation. (2014) 129(23):2440–92. doi: 10.1161/CIR.0000000000000029

PubMed Abstract | Crossref Full Text | Google Scholar

8. Benjamin EJ, Muntner P, Alonso A, Bittencourt MS, Callaway CW, Carson AP, et al. Heart disease and stroke statistics—2019 update: a report from the American Heart Association. Circulation. (2019) 139(10):e56–528. doi: 10.1161/CIR.0000000000000659

PubMed Abstract | Crossref Full Text | Google Scholar

9. Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. (2019) 25(1):44–56. doi: 10.1038/s41591-018-0300-7

PubMed Abstract | Crossref Full Text | Google Scholar

10. Hussain MM, Rafi U, Imran A, Rehman MU, Abbas SK. Risk factors associated with cardiovascular disorders: risk factors associated with cardiovascular disorders. Pak BioMedical J. (2024) 7(2):3–10. doi: 10.54393/pbmj.v7i02.1034

Crossref Full Text | Google Scholar

11. Satler LF, Green CE, Wallace RB, Rackley CE. Coronary artery disease in the elderly. Am J Cardiol. (1989) 63(3):245–8. doi: 10.1016/0002-9149(89)90294-4

PubMed Abstract | Crossref Full Text | Google Scholar

12. Vasan RS. Biomarkers of cardiovascular disease: molecular basis and practical considerations. Circulation. (2006) 113(19):2335–62. doi: 10.1161/CIRCULATIONAHA.104.482570

PubMed Abstract | Crossref Full Text | Google Scholar

13. Guasti L, Dilaveris P, Mamas MA, Richter D, Christodorescu R, Lumens J, et al. Digital health in older adults for the prevention and management of cardiovascular diseases and frailty. A clinical consensus statement from the ESC Council for Cardiology Practice/Taskforce on Geriatric Cardiology, the ESC Digital Health Committee and the ESC Working Group on e-Cardiology. ESC Heart Fail. (2022) 9(5):2808–22. doi: 10.1002/ehf2.14022

PubMed Abstract | Crossref Full Text | Google Scholar

14. Latif S, Usman M, Rana R, Qadir J. Phonocardiographic sensing using deep learning for abnormal heartbeat detection. IEEE Sens J. (2018) 18(22):9393–400. doi: 10.1109/JSEN.2018.2870759

Crossref Full Text | Google Scholar

15. Altaf A, Mahdin H, Mahmood A, Ninggal MIH, Altaf A, Javid I. Systematic review for phonocardiography classification based on machine learning. Int J Adv Comput Sci Appl. (2023) 14(8). doi: 10.14569/IJACSA.2023.0140889

Crossref Full Text | Google Scholar

16. Brites ISG, da Silva LM, Barbosa JLV, Rigo SJ, Correia SD, Leithardt VRQ. Machine learning and iot applied to cardiovascular diseases identification through heart sounds: a literature review. Informatics. (2021) 8(4):73. doi: 10.3390/informatics8040073

Crossref Full Text | Google Scholar

17. Abdelmageed S. Wrist-based Phonocardiogram Diagnosis Leveraging Machine Learning. Vaasa: Vaasan yliopisto (2019).

Google Scholar

18. Alkahtani HK, Haq IU, Ghadi YY, Innab N, Alajmi M, Nurbapa M. Precision diagnosis: an automated method for detecting congenital heart diseases in children from Phonocardiogram Signals Employing Deep Neural Network. IEEE Access. (2024). doi: 10.1109/ACCESS.2024.3395389

Crossref Full Text | Google Scholar

19. Li H, Wang X, Liu C, Zeng Q, Zheng Y, Chu X, et al. A fusion framework based on multi-domain features and deep learning features of phonocardiogram for coronary artery disease detection. Comput Biol Med. (2020) 120:103733. doi: 10.1016/j.compbiomed.2020.103733

PubMed Abstract | Crossref Full Text | Google Scholar

20. Azmeen A, Vakilzadian H, Haider H, Mathers DH, Zimmerman R, Bedi S, et al. Heart sounds: past, present, and future from a technological and clinical perspective–a systematic review. Proc Inst Mech Eng H. (2023) 237(6):669–82. doi: 10.1177/095441192311728

PubMed Abstract | Crossref Full Text | Google Scholar

21. Chamberlain D, Mofor J, Fletcher R, Kodgule R. Mobile stethoscope and signal processing algorithms for pulmonary screening and diagnostics. 2015 IEEE Global Humanitarian Technology Conference (GHTC); IEEE (2015). p. 385–92

Google Scholar

22. Luo H, Lamata P, Bazin S, Bautista T, Barclay N, Shahmohammadi M, et al. Smartphone as an electronic stethoscope: factors influencing heart sound quality. Eur Heart J Digit Health. (2022) 3(3):473–80. doi: 10.1093/ehjdh/ztac044

PubMed Abstract | Crossref Full Text | Google Scholar

23. Wu YC, Han CC, Chang CS, Chang FL, Chen SF, Shieh TY, et al. Development of an electronic stethoscope and a classification algorithm for cardiopulmonary sounds. Sensors. (2022) 22(11):4263. doi: 10.3390/s22114263

PubMed Abstract | Crossref Full Text | Google Scholar

24. Dwivedi S, Mohammad A. Heartbeat pattern and arrhythmia classification: a review. Front Biomed Technol. (2024) 11(1):130–48. doi: 10.18502/fbt.v11i1.14520

Crossref Full Text | Google Scholar

25. Stillman AE, Oudkerk M, Ackerman M, Becker CR, Buszman PE, De Feyter PJ, et al. Use of multidetector computed tomography for the assessment of acute chest pain: a consensus statement of the North American Society of Cardiac Imaging and the European Society of Cardiac Radiology. Int J Cardiovasc Imaging. (2007) 23:415–27. doi: 10.1007/s10554-007-9226-8

PubMed Abstract | Crossref Full Text | Google Scholar

26. Eyre H, Kahn R, Robertson RM, ACS/ADA/AHA Collaborative Writing Committee, ACS/ADA/AHA Collaborative Writing Committee Members, Clark NG, et al. Preventing cancer, cardiovascular disease, and diabetes: a common agenda for the American Cancer Society, the American Diabetes Association, and the American Heart Association. Circulation. (2004) 109(25):3244–55. doi: 10.1161/01.CIR.0000133321.00456.00

PubMed Abstract | Crossref Full Text | Google Scholar

27. Moses JC, Adibi S, Angelova M, Islam SMS. Smart home technology solutions for cardiovascular diseases: a systematic review. Applied System Innovation. (2022) 5(3):51. doi: 10.3390/asi5030051

Crossref Full Text | Google Scholar

28. Reddy KS, Mathur MR. Global burden of CVD: prevalence, pattern, and trends. In: Haring R, editor. Handbook of Global Health. Cham: Springer International Publishing (2021). p. 423–37.

Google Scholar

29. Miller JC, Skoll D, Saxon LA. Home monitoring of cardiac devices in the era of COVID-19. Curr Cardiol Rep. (2021) 23:1–9. doi: 10.1007/s11886-020-01431-w

PubMed Abstract | Crossref Full Text | Google Scholar

30. Pullano SA, Oliva G, Presta P, Carullo N, Musolino M, Andreucci M, et al. A portable easy-to-use triboelectric sensor for arteriovenous fistula monitoring in dialysis patients. Sensors Int. (2025) 6:100309. doi: 10.1016/j.sintl.2024.100309

Crossref Full Text | Google Scholar

31. Menniti M, Laganà F, Oliva G, Bianco M, Fiorillo AS, Pullano SA. Development of non-invasive ventilator for homecare and patient monitoring system. Electronics. (2024) 13(4):790. doi: 10.3390/electronics1304079

Crossref Full Text | Google Scholar

32. Pullano SA, Oliva G, Lagana F, Fiorillo AS, Prattico D, Presta P, et al. Development of a wearable device for arteriovenous Fistula monitoring. 2024 IEEE Sensors Applications Symposium (SAS); IEEE (2024). p. 1–4

Google Scholar

33. Ahlstrom C, Hult P, Rask P, Karlsson JE, Nylander E, Dahlström U, et al. Feature extraction for systolic heart murmur classification. Ann Biomed Eng. (2006) 34:1666–77. doi: 10.1007/s10439-006-9187-4

PubMed Abstract | Crossref Full Text | Google Scholar

34. Mesin L. Heartbeat monitoring from adaptively down-sampled electrocardiogram. Comput Biol Med. (2017) 84:217–25. doi: 10.1016/j.compbiomed.2017.03.023

PubMed Abstract | Crossref Full Text | Google Scholar

35. Verma P, Yadav A, Sharma SD, Hooda D. Signal processing and feature engineering of respiratory disease (2023).

36. Chen T, Morris J, Martin E. Probability density estimation via an infinite Gaussian mixture model: application to statistical process monitoring. J R Stat Soc Ser C Appl Stat. (2006) 55(5):699–715. doi: 10.1111/j.1467-9876.2006.00560.x

Crossref Full Text | Google Scholar

37. Moreno-Barea FJ, Franco L, Elizondo D, Grootveld M. Application of data augmentation techniques towards metabolomics. Comput Biol Med. (2022) 148:105916. doi: 10.1016/j.compbiomed.2022.105916

PubMed Abstract | Crossref Full Text | Google Scholar

38. Frid-Adar M, Diamant I, Klang E, Amitai M, Goldberger J, Greenspan H. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing. (2018) 321:321–31. doi: 10.1016/j.neucom.2018.09.013

Crossref Full Text | Google Scholar

39. Borré A, Seman LO, Camponogara E, Stefenon SF, Mariani VC, Coelho LDS. Machine fault detection using a hybrid CNN-LSTM attention-based model. Sensors. (2023) 23(9):4512. doi: 10.3390/s23094512

Crossref Full Text | Google Scholar

40. Aldughayfiq B, Ashfaq F, Jhanjhi NZ, Humayun M. A deep learning approach for atrial fibrillation classification using multi-feature time series data from ecg and ppg. Diagnostics. (2023) 13(14):2442. doi: 10.3390/diagnostics13142442

PubMed Abstract | Crossref Full Text | Google Scholar

41. Lee J, Park J, Kim KL, Nam J. SampleCNN: end-to-end deep convolutional neural networks using very small filters for music classification. Appl Sci. (2018) 8(1):150. doi: 10.3390/app8010150

Crossref Full Text | Google Scholar

42. Chen J, Shen M, Ma W, Zheng W. A spatio-temporal learning-based model for sleep apnea detection using single-lead ECG signals. Front Neurosci. (2022) 16:972581. doi: 10.3389/fnins.2022.972581

PubMed Abstract | Crossref Full Text | Google Scholar

43. Ibarra-Hernández RF, Alonso-Arévalo MA, Cruz-Gutiérrez A, Licona-Chávez AL, Villarreal-Reyes S. Design and evaluation of a parametric model for cardiac sounds. Comput Biol Med. (2017) 89:170–80. doi: 10.1016/j.compbiomed.2017.08.007

PubMed Abstract | Crossref Full Text | Google Scholar

44. Shipp S. Neural elements for predictive coding. Front Psychol. (2016) 7:1792. doi: 10.3389/fpsyg.2016.01792

PubMed Abstract | Crossref Full Text | Google Scholar

45. Singh AK, Krishnan S. ECG signal feature extraction trends in methods and applications. Biomed Eng Online. (2023) 22(1):22. doi: 10.1186/s12938-023-01075-1

PubMed Abstract | Crossref Full Text | Google Scholar

46. Ahmad T, Wu J, Alwageed HS, Khan F, Khan J, Lee Y. Human activity recognition based on deep-temporal learning using convolution neural networks features and bidirectional gated recurrent unit with features selection. IEEE Access. (2023) 11:33148–33159. doi: 10.1109/ACCESS.2023.3263155

Crossref Full Text | Google Scholar

47. Chen J, Guo Z, Xu X, Jeon G, Camacho D. Artificial intelligence for heart sound classification: a review. Expert Syst. (2024) 41(4):e13535. doi: 10.1111/exsy.13535

Crossref Full Text | Google Scholar

48. Koşar E, Barshan B. A new CNN-LSTM architecture for activity recognition employing wearable motion sensor data: Enabling diverse feature extraction. Eng Appl Artif Intell. (2023) 124:106529. doi: 10.1016/j.engappai.2023.106529

Crossref Full Text | Google Scholar

49. Khan FA, Abid A, Khan MS. Automatic heart sound classification from segmented/unsegmented phonocardiogram signals using time and frequency features. Physiol Meas. (2020) 41(5):055006. doi: 10.1088/1361-6579/ab8770

PubMed Abstract | Crossref Full Text | Google Scholar

50. Lyu X, Rani S, Manimurugan S, Maple C, Feng Y. A deep neuro-fuzzy method for ECG big data analysis via exploring multimodal feature fusion. IEEE Trans Fuzzy Syst. (2024). doi: 10.1109/TFUZZ.2024.3416217

Crossref Full Text | Google Scholar

51. Barros B, Lacerda P, Albuquerque C, Conci A. Pulmonary COVID-19: learning spatiotemporal features combining CNN and LSTM networks for lung ultrasound video classification. Sensors. (2021) 21(16):5486. doi: 10.3390/s21165486

PubMed Abstract | Crossref Full Text | Google Scholar

52. Paschalis SM, Hutapea DKY, Bachri KO. Heart sound processing for early diagnostic of heart abnormalities using support vector machine. J ELTIKOM. (2024) 8(1):57–65. doi: 10.31961/eltikom.v8i1.1031

Crossref Full Text | Google Scholar

53. Benjamin EJ, Virani SS, Callaway CW, Chamberlain AM, Chang AR, Cheng S, et al. Heart disease and stroke statistics—2018 update: a report from the American Heart Association. Circulation. (2018) 137(12):e67–492. doi: 10.1161/CIR.000000000000055

PubMed Abstract | Crossref Full Text | Google Scholar

54. Chen C, Hua Z, Zhang R, Liu G, Wen W. Automated arrhythmia classification based on a combination network of CNN and LSTM. Biomed Signal Process Control. (2020) 57:101819. doi: 10.1016/j.bspc.2019.101819

Crossref Full Text | Google Scholar

55. Springer DB, Tarassenko L, Clifford GD. Logistic regression-HSMM-based heart sound segmentation. IEEE Trans Biomed Eng. (2015) 63(4):822–32. doi: 10.1109/TBME.2015.2475278

PubMed Abstract | Crossref Full Text | Google Scholar

56. Nowbar AN, Gitto M, Howard JP, Francis DP, Al-Lamee R. Mortality from ischemic heart disease: analysis of data from the World Health Organization and coronary artery disease risk factors from NCD Risk Factor Collaboration. Circulation. (2019) 12(6):e005375. doi: 10.1161/CIRCOUTCOMES.118.00537

PubMed Abstract | Crossref Full Text | Google Scholar

57. Li KHC, White FA, Tipoe T, Liu T, Wong MC, Jesuthasan A, et al. The current state of mobile phone apps for monitoring heart rate, heart rate variability, and atrial fibrillation: narrative review. JMIR mHealth and uHealth. (2019) 7(2):e11606. doi: 10.2196/11606

PubMed Abstract | Crossref Full Text | Google Scholar

58. De Chazal P, O'Dwyer M, Reilly RB. Automatic classification of heartbeats using ECG morphology and heartbeat interval features. IEEE Trans Biomed Eng. (2004) 51(7):1196–206. doi: 10.1109/TBME.2004.827359

PubMed Abstract | Crossref Full Text | Google Scholar

59. Mallinckrod CH, Lane PW, Schnell D, Peng Y, Mancuso JP. Recommendations for the primary analysis of continuous endpoints in longitudinal clinical trials. Drug Inf J. (2008) 42(4):303–19. doi: 10.1177/009286150804200402

Crossref Full Text | Google Scholar

60. Nashif S, Raihan MR, Islam MR, Imam MH. Heart disease detection by using machine learning algorithms and a real-time cardiovascular health monitoring system. World J Eng Technol. (2018) 6(4):854–73. doi: 10.4236/wjet.2018.64057w

Crossref Full Text | Google Scholar

Keywords: cardiac disease screening, heart sounds, murmurs, systolic murmurs, diastolic murmurs

Citation: Jain K, Jain R, Mohammad SK and Aggarwal S (2025) MYSTETH—home-based heart monitoring. Front. Digit. Health 7:1616334. doi: 10.3389/fdgth.2025.1616334

Received: 22 April 2025; Accepted: 25 August 2025;
Published: 30 September 2025.

Edited by:

Fried Michael Dahlweid, Dedalus S.p.A., Italy

Reviewed by:

Salvatore Andrea Pullano, University Magna Graecia of Catanzaro, Italy
Kazi A. Kalpoma, Ahsanullah University of Science and Technology, Bangladesh

Copyright: © 2025 Jain, Jain, Mohammad and Aggarwal. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Swati Aggarwal, c3dhdGkuYWdnYXJ3YWxAaGltb2xkZS5ubw==

Present addresses: Rohit Jain,Intuit, Inc., Mountain View, CA, United States
Salik Khwaja Mohammad,Google, Inc., Boulder, CO, United States

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.