<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Comput. Neurosci.</journal-id>
<journal-title>Frontiers in Computational Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Comput. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5188</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fncom.2020.564015</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>A Study on Arrhythmia via ECG Signal Classification Using the Convolutional Neural Network</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Wu</surname> <given-names>Mengze</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Lu</surname> <given-names>Yongdi</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Yang</surname> <given-names>Wenli</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Wong</surname> <given-names>Shen Yuong</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/967897/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Information Engineering, Wuhan University of Technology</institution>, <addr-line>Wuhan</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Electrical and Electronics Engineering, Xiamen University Malaysia</institution>, <addr-line>Sepang</addr-line>, <country>Malaysia</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Electrical and Automation Engineering, Nanjing Normal University</institution>, <addr-line>Nanjing</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Germ&#x000E1;n Mato, Bariloche Atomic Centre (CNEA), Argentina</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Philip Warrick, PeriGen Inc., Montr&#x000E9;al, Canada; Pawe&#x00142; P&#x00142;awiak, Cracow University of Technology, Poland</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Shen Yuong Wong <email>shenyuong.wong&#x00040;xmu.edu.my</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>05</day>
<month>01</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2020</year>
</pub-date>
<volume>14</volume>
<elocation-id>564015</elocation-id>
<history>
<date date-type="received">
<day>21</day>
<month>05</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>02</day>
<month>11</month>
<year>2020</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2021 Wu, Lu, Yang and Wong.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Wu, Lu, Yang and Wong</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>Cardiovascular diseases (CVDs) are the leading cause of death today. The current identification method of the diseases is analyzing the Electrocardiogram (ECG), which is a medical monitoring technology recording cardiac activity. Unfortunately, looking for experts to analyze a large amount of ECG data consumes too many medical resources. Therefore, the method of identifying ECG characteristics based on machine learning has gradually become prevalent. However, there are some drawbacks to these typical methods, requiring manual feature recognition, complex models, and long training time. This paper proposes a robust and efficient 12-layer deep one-dimensional convolutional neural network on classifying the five micro-classes of heartbeat types in the MIT- BIH Arrhythmia database. The five types of heartbeat features are classified, and wavelet self-adaptive threshold denoising method is used in the experiments. Compared with BP neural network, random forest, and other CNN networks, the results show that the model proposed in this paper has better performance in accuracy, sensitivity, robustness, and anti-noise capability. Its accurate classification effectively saves medical resources, which has a positive effect on clinical practice.</p></abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>ECG</kwd>
<kwd>anti-noise performance</kwd>
<kwd>feature classification</kwd>
<kwd>convolutional neural network</kwd>
</kwd-group>
<contract-sponsor id="cn001">Xiamen University<named-content content-type="fundref-id">10.13039/501100008865</named-content></contract-sponsor>
<counts>
<fig-count count="3"/>
<table-count count="13"/>
<equation-count count="7"/>
<ref-count count="38"/>
<page-count count="10"/>
<word-count count="7348"/>
</counts>
</article-meta>
</front>
<body>
<fig position="float">
<label>Graphical Abstract</label>
<p>The graphical abstract of the methodology.</p>
<graphic xlink:href="fncom-14-564015-g0003.tif"/>
</fig>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>Cardiovascular disease is a common disease that seriously threatens human health, especially the health of middle-aged and older people. It is characterized by high prevalence, high disability, and high mortality. Nowadays, the world is facing with the aging population. The increasing aggravation of cardiovascular disease has become a major public health problem (Mc Namara et al., <xref ref-type="bibr" rid="B19">2019</xref>). ECG analysis is an effective way of evaluating heart health. Therefore, the identification and classification of ECG signals are essential to cardiovascular diseases. Not only for early prevention but also necessary for timely detection and proper treatment. It is of considerable significance to study the classification of related ECG signals (Guo et al., <xref ref-type="bibr" rid="B9">2016</xref>; Yin et al., <xref ref-type="bibr" rid="B36">2016</xref>).</p>
<p>The electrocardiogram is a visual time series that records the electrical activity generated by each cardiac cycle of the heart in real-time and is now widely used in heart rate detection (Homaeinezhad et al., <xref ref-type="bibr" rid="B10">2012</xref>). This non-invasive detection method is easy to operate and has become an essential tool for assisting doctors in analyzing pathology. At this stage, the judgment of cardiovascular disease mainly depends on the doctor&#x00027;s experience. However, there are many types of heart diseases, and long-term manual detection makes it easy to cause false detection. How to quickly and accurately analyze specific diseases has become a new problem (Song et al., <xref ref-type="bibr" rid="B28">2014</xref>). In addition, the traits of ECG signals include random, low-frequency, and susceptible, resulting in the diagnosis results are unstable. Intelligent automatic recognition and classification of ECG signals have become an inevitable choice to improve the efficiency and accuracy of ECG recognition (Yu and Chen, <xref ref-type="bibr" rid="B37">2007</xref>).</p>
<p>With the maturity of Artificial Intelligence (AI) technology, many machine learning methods are used in the ECG signal feature detection, aiming at solving the problems related to large amounts of ECG signal feature data and a heavy load of manual detection. The typical methods are neural networks (NN) (Jiang and Kong, <xref ref-type="bibr" rid="B11">2007</xref>), support vector machine (SVM) (Osowski et al., <xref ref-type="bibr" rid="B21">2004</xref>), path forest (Luz et al., <xref ref-type="bibr" rid="B16">2013</xref>), Independent Component Correlation Algorithm (ICA) (Sarfraz et al., <xref ref-type="bibr" rid="B26">2014</xref>). Regarding neural networks, Jiang and Kong (<xref ref-type="bibr" rid="B11">2007</xref>) propose an evolutionary block-based neural network (BbNNs) for the classification of ECG signals. The BbNN is composed of a group of two-dimensional modular networks with flexible structure and internal configuration. The Hermite transform coefficient and the time interval between adjacent two R peaks are used as the input of BbNN. Osowski et al. (<xref ref-type="bibr" rid="B21">2004</xref>) introduce Hermite function as a feature extraction method in the SVM classifier, and also use higher-order statistics (HOS) to better extract features. Luz et al. (<xref ref-type="bibr" rid="B16">2013</xref>) introduced an optimal path forest classifier (OPF) to compare the performance of 6 distance metrics, six feature extraction algorithms, and three classifiers in two variants of the same data set. Although the accuracy of OPF is not as good as that of SVM, OPF is more efficient than SVM in terms of calculation time during training and testing phases. Sarfraz et al. (<xref ref-type="bibr" rid="B26">2014</xref>) use the basic function of a typical ECG signal obtained by ICA for pattern recognition. The features obtained by ICA are used as the training set. Although these methods have good performance and have achieved certain results, it is difficult to put into practice due to the severe deficiency of requiring manual design features.</p>
<p>In recent years, the machine learning and deep learning network has not only made remarkable achievements in the fields of image processing, audio recognition and many other fields (Wong et al., <xref ref-type="bibr" rid="B32">2015a</xref>,<xref ref-type="bibr" rid="B33">b</xref>, <xref ref-type="bibr" rid="B31">2016</xref>; Kandala et al., <xref ref-type="bibr" rid="B12">2019</xref>; P&#x00142;awiak et al., <xref ref-type="bibr" rid="B23">2019</xref>, <xref ref-type="bibr" rid="B24">2020</xref>), it has also been commonly used in the assisted diagnosis of heart disease based on ECG signals(Zubair et al., <xref ref-type="bibr" rid="B38">2016</xref>; Acharya et al., <xref ref-type="bibr" rid="B1">2017a</xref>,<xref ref-type="bibr" rid="B2">b</xref>; Yildirim et al., <xref ref-type="bibr" rid="B35">2018</xref>; Gao et al., <xref ref-type="bibr" rid="B8">2019</xref>; Atal and Singh, <xref ref-type="bibr" rid="B5">2020</xref>; P&#x00142;awiak and Acharya, <xref ref-type="bibr" rid="B25">2020</xref>). P&#x00142;awiak and Acharya (<xref ref-type="bibr" rid="B25">2020</xref>) use a deep genetic ensemble of classifiers to classify long-duration ECG signal (10 s). Gao et al. (<xref ref-type="bibr" rid="B8">2019</xref>) implemented an effective long short-term memory (LSTM) recurrence network model to classify 8 types of heartbeats. Atal and Singh (<xref ref-type="bibr" rid="B5">2020</xref>) proposed an optimization-based deep convolutional neural network to classify five different heartbeats. Compared with traditional neural networks, deep learning network can automatically extract features, recognize intricate data patterns, and eliminate complex signal preprocessing. Deep learning network also has a stronger nonlinear fitting ability, which has a better effect in identifying single-lead, multi-class, and unbalanced ECG datasets (Acharya et al., <xref ref-type="bibr" rid="B1">2017a</xref>). Convolutional neural network (CNN) is a feedforward neural network that has been widely researched and used in deep learning, which has been applied successfully for the classification of arrhythmia (Zubair et al., <xref ref-type="bibr" rid="B38">2016</xref>; Acharya et al., <xref ref-type="bibr" rid="B1">2017a</xref>,<xref ref-type="bibr" rid="B2">b</xref>; Yildirim et al., <xref ref-type="bibr" rid="B35">2018</xref>; Gao et al., <xref ref-type="bibr" rid="B8">2019</xref>; Atal and Singh, <xref ref-type="bibr" rid="B5">2020</xref>; P&#x00142;awiak and Acharya, <xref ref-type="bibr" rid="B25">2020</xref>) ECG signals. In the previous literature (Zubair et al., <xref ref-type="bibr" rid="B38">2016</xref>; Acharya et al., <xref ref-type="bibr" rid="B1">2017a</xref>,<xref ref-type="bibr" rid="B2">b</xref>; Yildirim et al., <xref ref-type="bibr" rid="B35">2018</xref>; Atal and Singh, <xref ref-type="bibr" rid="B5">2020</xref>), most of the works focus on the recognition of five main macro classes, namely Non-ectopic (N); Supraventricular ectopic (S); Ventricular ectopic (V); Fusion (F); Unknown (Q).</p>
<p>There is very little effort devoted to classify the micro-classes of the ECG signal, hence it serves as our main motivation to study the micro-classification heartbeats, of five types, i.e., Normal (NOR), Left Bundle Branch Block (LBBB), Right Bundle Branch Block (RBBB), Atrial Premature (AP), Premature Ventricular Contraction (PVC). The contribution of this paper is two-fold, in which the proposed algorithm is endowed with an ability to effectively process the non-filtered dataset with its potential anti-noise features, and secondly this paper presents an analysis of micro-classes of the ECG signal that compares some techniques of machine learning such as BP and Random Forest. The results can be served as a good source of benchmark literature to other researchers in the same field for future research work.</p>
<p>Section ECG Data Processing of the paper mainly introduces the ECG dataset used in this study and provides a detailed description of the data segmentation and preprocessing. In section Methodology, the architecture of the proposed algorithm and the specific experiment design is outlined. In section Result and Discussion, the performance and robustness of the proposed network is evaluated on the MIT-BIH Arrhythmia database and compared with BP, Random Forest, and several benchmarked CNN networks. Finally, section Conclusion summarizes the paper.</p>
</sec>
<sec id="s2">
<title>ECG Data Processing</title>
<sec>
<title>ECG Dataset</title>
<p>The MIT-BIH database, an ECG database provided by the Massachusetts Institute of Technology and based on international standards and annotated information by multiple experts (Moody and Mark, <xref ref-type="bibr" rid="B20">2001</xref>) is used in this study. The MIT-BIH database has been frequently used by the academic community in research for the detection and classification of arrhythmic heartbeats. The MIT-BIH database contains 48 ECG recordings, each recording time is 30 min, the sampling frequency is 360 Hz, and each ECG record is composed of two leads. MIT-BIH database can make adjustments and corrections based on the information annotated by experts and optimization algorithms. Furthermore, it learns from existing solutions for self-optimization.</p>
</sec>
<sec>
<title>Pre-processing</title>
<p>ECG signals collected in a clinical environment are usually mixed with different interference, such as power frequency interference, baseline drift, and EMG interference. The raw data needs to be de-noised to make the classification more accurate. The bandpass filters, low-pass filters, wavelet transforms are widely used in the field of ECG denoising (Ahlstrom and Tompkins, <xref ref-type="bibr" rid="B3">1985</xref>; Bazi et al., <xref ref-type="bibr" rid="B7">2013</xref>; Wang et al., <xref ref-type="bibr" rid="B29">2015</xref>; Yadav et al., <xref ref-type="bibr" rid="B34">2015</xref>). In this paper, the wavelet transform method is used to preprocess the ECG signal. Wavelet transform is an algorithm that decomposes non-stationary signals into scale signals of different frequency bands. The filter uses an adaptive threshold filtering algorithm (Alfaouri and Daqrouq, <xref ref-type="bibr" rid="B4">2008</xref>; Awal et al., <xref ref-type="bibr" rid="B6">2014</xref>), and selects Sym4 in the Symlet wavelet function family as the wavelet function (Singh and Tiwari, <xref ref-type="bibr" rid="B27">2006</xref>). Because the convolutional neural network has the feature of automatically extracting features from the inside of the signal, this paper only performs simple filtering on the signal, which can enhance the generalization of the network and reduce signal distortion. <xref ref-type="fig" rid="F1">Figure 1</xref> shows the ECG signal before and after filtering.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>ECG signal before and after filtering.</p></caption>
<graphic xlink:href="fncom-14-564015-g0001.tif"/>
</fig>
</sec>
<sec>
<title>Data Segmentation</title>
<p>In the MIT-BIH dataset, each heartbeat is marked with a disease annotation. This paper selects five heartbeats for classification, normal (NOR), left bundle branch block (LBBB), and right bundle branch block (RBBB), Atrial premature beats (AP), premature ventricular beats (PVC). The process starts by using the Pan-Tompkins algorithm to detected R-peak (Pan and Tompkins, <xref ref-type="bibr" rid="B22">1985</xref>). Dataset is segmented into 360 samples and centered around the detected R-peaks. The process selects a single lead in the dataset, and all segments use the Z-score normalizing method (Acharya et al., <xref ref-type="bibr" rid="B2">2017b</xref>).</p>
</sec>
<sec>
<title>Data Enhancement</title>
<p>The unbalanced training set affects the feature learning of the convolutional neural network (Masko and Hensman, <xref ref-type="bibr" rid="B18">2015</xref>), thereby reducing the recognition accuracy. This paper selects 16 recordings that contain most of these five heartbeats from the MIT-BIH dataset. After denoising and segmenting, oversampling is performed on the under-represented classes (Masko and Hensman, <xref ref-type="bibr" rid="B18">2015</xref>). It randomly duplicates the under-represented classes, and discards the over-represented classes, which ultimately reduces the data imbalance of the training set. According to <xref ref-type="table" rid="T1">Table 1</xref>, only the C1 and C4 are enhanced because they greatly deviated from the average. The C2-C4 is not enhanced since they are only slightly imbalanced.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Unbalanced and balanced dataset.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Class</bold></th>
<th valign="top" align="center" colspan="3" style="border-bottom: thin solid #000000;"><bold>Classification dataset</bold></th>
</tr>
<tr>
<th/>
<th valign="top" align="left"><bold>Type</bold></th>
<th valign="top" align="center"><bold>Unbalanced</bold></th>
<th valign="top" align="center"><bold>Balanced</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">C1</td>
<td valign="top" align="left">NOR</td>
<td valign="top" align="center">18,653</td>
<td valign="top" align="center">7,563</td>
</tr>
<tr>
<td valign="top" align="left">C2</td>
<td valign="top" align="left">LBBB</td>
<td valign="top" align="center">6,612</td>
<td valign="top" align="center">6,612</td>
</tr>
<tr>
<td valign="top" align="left">C3</td>
<td valign="top" align="left">RBBB</td>
<td valign="top" align="center">7,165</td>
<td valign="top" align="center">7,165</td>
</tr>
<tr>
<td valign="top" align="left">C4</td>
<td valign="top" align="left">AP</td>
<td valign="top" align="center">4,758</td>
<td valign="top" align="center">4,758</td>
</tr>
<tr>
<td valign="top" align="left">C5</td>
<td valign="top" align="left">PVC</td>
<td valign="top" align="center">1,208</td>
<td valign="top" align="center">6,324</td>
</tr>
<tr>
<td valign="top" align="center" colspan="2"><bold>Total data</bold></td>
<td valign="top" align="center">38,396</td>
<td valign="top" align="center">32,422</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec>
<title>Ten-Fold Cross Validation</title>
<p>The original and de-noised data samples are used for experiments and 10-fold cross-validation is employed. This paper selects 16 recordings among the 48 recordings in the database, which contains nearly all of the five heartbeats that need to classify. A total of 32,422 heartbeats are extracted from 16 recordings, which are separated equally into 10 groups. 9 out of 10 groups are used in the training process while the remaining one of the remaining 10 groups is used for validation to get the optimal parameters. This process is iterated by 10 times by shifting test data. The performance is recorded after every iteration and integrated into one final confusion matrix at the end of the 10th iteration. The overall performance is calculated from the integrated confusion matrix.</p>
</sec>
</sec>
<sec sec-type="methods" id="s3">
<title>Methodology</title>
<sec>
<title>The Architecture</title>
<p>This paper proposes a one-dimensional 12-layer convolution neural network (CNN) network structure to classify the five sub-classes of cardiac arrhythmia. CNN is a network consists of the input layer, convolution layer, pooling layer, fully-connected layer, and output layer. In contrast to traditional neural networks, CNN has convolution and pooling layers, which can extract and map features from input data to speed up learning and reduce over-fitting. Because the CNN has the feature of the multilayer perception, the two-dimensional convolution neural network has been widely used in image processing (Li et al., <xref ref-type="bibr" rid="B14">2014</xref>; Wei et al., <xref ref-type="bibr" rid="B30">2015</xref>). In this paper, we propose a one-dimensional 12 layer CNN to process a one-dimensional time series with uniform interval sampling (Kiranyaz et al., <xref ref-type="bibr" rid="B13">2015</xref>). Several modifications are made in the network structure, i.e., the proposed CNN network uses the average-pooling layer instead of the max-pooling layer of the compared CNN network. The average-pooling layer can preserve the overall feature of the input data, which will be beneficial to classify the heartbeats. Also, the proposed CNN network has one more alternating convolution and pooling layer as compared to the benchmark CNN network. <xref ref-type="table" rid="T2">Table 2</xref> summarizes the proposed CNN network architecture, including 8 alternating convolutions and average-pooling layers. They are followed by a dropout layer and two fully-connected layers, as seen in <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
<list list-type="roman-lower">
<list-item><p><bold>Convolution layer</bold></p>
<p>To process one-dimensional ECG signal, this paper uses a one-dimensional convolution kernel, which convolutes independently of the feature map of the previous layer. The output of the convolution layer is obtained by offsetting the convolution kernel and transferring it to the nonlinear activation function. The output expression is shown in Equation (1).
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x0002B;</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msubsup><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>.</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mtext>&#x000A0;</mml:mtext><mml:mo>*</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
Where, <inline-formula><mml:math id="M2"><mml:msubsup><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the output of the ith neuron in layer <italic>l</italic>, <italic>f</italic>() is the activation function and <inline-formula><mml:math id="M3"><mml:msubsup><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the offset of the neuron in layer <italic>l</italic>. <inline-formula><mml:math id="M4"><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the output of neuron in layer <italic>l-1</italic>, <inline-formula><mml:math id="M5"><mml:msubsup><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>.</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the <italic>k</italic><sup><italic>th</italic></sup> convolution kernels in <italic>l</italic><sup><italic>th</italic></sup> layer.</p></list-item>
<list-item><p><bold>Pooling layer</bold></p>
<p>Convolution of the next layer is commonly the pooling layer. By reducing the dimension of convolution layer output data, network complexity is reduced, as well as overfitting phenomenon. Robustness of the network is enhanced in this process. The pooling layer averages or maximizes the output features of the convolutional layer, and the corresponding methods are respectively, average pooling or maximum pooling. The output expression by Equation (2)
<disp-formula id="E2"><label>(2)</label><mml:math id="M6"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>o</mml:mi><mml:mi>l</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p><inline-formula><mml:math id="M7"><mml:msubsup><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the output of the <italic>i</italic><sup><italic>th</italic></sup> neuron in the <italic>l</italic> layer, <italic>f</italic>() is the activation function, <inline-formula><mml:math id="M8"><mml:msubsup><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the offset of the neurons in <italic>l</italic> layer,<inline-formula><mml:math id="M9"><mml:msubsup><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the sampling weight coefficient, <inline-formula><mml:math id="M10"><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the output of the neuron in <italic>l-1</italic> layer, <italic>pool</italic>() is the pooling function.</p></list-item>
<list-item><p><bold>Fully-connected layer</bold></p>
<p>After extracting features from multiple convolution layers and pooling layers, the fully-connected layer is used to expand the connection of all features. Finally, the SoftMax layer makes a logistic regression classification. Fully-connected layer transfers the weighted sum of the output of the previous layer to the activation function. The expression of the output is shown in Equation (3)
<disp-formula id="E3"><label>(3)</label><mml:math id="M11"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p><inline-formula><mml:math id="M12"><mml:msubsup><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the output of the <italic>l</italic> layer of the <italic>i</italic><sup><italic>th</italic></sup> neuron, <italic>f</italic>() is the activation function, <inline-formula><mml:math id="M13"><mml:msubsup><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the offset of the <italic>l</italic><sup><italic>th</italic></sup> layer of the neuron, <inline-formula><mml:math id="M14"><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the output of the layer <italic>l-1</italic> of the neuron, <inline-formula><mml:math id="M15"><mml:msubsup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the network weight.</p></list-item>
<list-item><p><bold>Dropout layer</bold></p>
<p>There is usually a dropout layer before the fully-connected layer. The dropout layer will temporarily disconnect some neurons from the network according to the certain probability during the training of the convolution neural network, which reduces the joint adaptability between neuron nodes, reduces overfitting, and enhances the generalization ability of the network.</p></list-item>
<list-item><p><bold>Training algorithm</bold> The training algorithm of the convolution neural network is a backward propagation algorithm based on gradient descent. The network hyperparameters are estimated by the loss function, which is the deviations of the output vector and the expected output vector. Hyperparameters include the convolution kernel parameter <italic>W</italic> of the convolution layer, the sampling weight coefficient &#x003B1; of the pooling layer, the network weight <italic>w</italic> of the fully-connected layer and the offset <italic>b</italic> of each layer. The training of a convolution neural network consists of two phases, forward propagation and reverse propagation. In the forward propagation stage, the training data is input into the neural network, and the output vectors of the middle and output layers are calculated. In the reverse propagation stage, the output vectors of the output layer are compared with the expected output vectors and calculated the loss function with respect to the weights of the network. The loss is propagated back to the initial layers (in reverse direction) using the gradient descent method to update the weights for each neuron in every layer. Gradient descent comprises two steps: calculating gradients of the loss function, which is calculated by chain rules, then updating weight in the opposite or reverse direction of the gradient of the loss function, which is distinct from the forward calculation of loss function. A cost function is also calculated for the neuron output in each hidden layer to optimize the network hyperparameters continuously. The network ends training when it reaches the set error after multiple iterations.</p></list-item>
</list>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>A summary table of the proposed CNN model for this work.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Layers</bold></th>
<th valign="top" align="left"><bold>Type</bold></th>
<th valign="top" align="center"><bold>Output</bold></th>
<th valign="top" align="center"><bold>Kernel size</bold></th>
<th valign="top" align="center"><bold>Stride</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Layer 1</td>
<td valign="top" align="left">Convolution</td>
<td valign="top" align="center">360<sup>&#x0002A;</sup>16</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>13</td>
<td valign="top" align="center">1</td>
</tr>
<tr>
<td valign="top" align="left">Layer 2</td>
<td valign="top" align="left">Average-Pooling</td>
<td valign="top" align="center">179<sup>&#x0002A;</sup>16</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>3</td>
<td valign="top" align="center">2</td>
</tr>
<tr>
<td valign="top" align="left">Layer 3</td>
<td valign="top" align="left">Convolution</td>
<td valign="top" align="center">179<sup>&#x0002A;</sup>32</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>15</td>
<td valign="top" align="center">1</td>
</tr>
<tr>
<td valign="top" align="left">Layer 4</td>
<td valign="top" align="left">Average-Pooling</td>
<td valign="top" align="center">89<sup>&#x0002A;</sup>32</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>3</td>
<td valign="top" align="center">2</td>
</tr>
<tr>
<td valign="top" align="left">Layer 5</td>
<td valign="top" align="left">Convolution</td>
<td valign="top" align="center">89<sup>&#x0002A;</sup>64</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>17</td>
<td valign="top" align="center">1</td>
</tr>
<tr>
<td valign="top" align="left">Layer 6</td>
<td valign="top" align="left">Average-Pooling</td>
<td valign="top" align="center">44<sup>&#x0002A;</sup>64</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>3</td>
<td valign="top" align="center">2</td>
</tr>
<tr>
<td valign="top" align="left">Layer 7</td>
<td valign="top" align="left">Convolution</td>
<td valign="top" align="center">44<sup>&#x0002A;</sup>128</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>19</td>
<td valign="top" align="center">1</td>
</tr>
<tr>
<td valign="top" align="left">Layer 8</td>
<td valign="top" align="left">Average-Pooling</td>
<td valign="top" align="center">21<sup>&#x0002A;</sup>128</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>3</td>
<td valign="top" align="center">2</td>
</tr>
<tr>
<td valign="top" align="left">Layer 9</td>
<td valign="top" align="left">Dropout</td>
<td valign="top" align="center">21<sup>&#x0002A;</sup>128</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">Layer 10</td>
<td valign="top" align="left">Fully-connected</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>35</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">Layer 11</td>
<td valign="top" align="left">Fully-connected</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>5</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">Layer 12</td>
<td valign="top" align="left">SoftMax</td>
<td valign="top" align="center">1<sup>&#x0002A;</sup>5</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>The architecture for the proposed CNN model.</p></caption>
<graphic xlink:href="fncom-14-564015-g0002.tif"/>
</fig>
<p>Layer 1 is convoluted with the kernel size of 13 and the number of the filters of 16. An average-pooling layer with the size of 3 is applied, hence the output of layer 2 is reduced to 179<sup>&#x0002A;</sup>16. Then the feature map of layer 2 is convoluted with the kernel size of 15, and the number of the filter is 32 in layer 3. Again, an average-pooling layer is applied with the size of 3, reducing the neurons from 176<sup>&#x0002A;</sup>16 to 89<sup>&#x0002A;</sup>32 (layer 4). The convolution kernel size is 17, and the number of the filter is 64 in layer 5. An average-pooling layer with the size of 3 is applied after layer 5, reducing output to 44<sup>&#x0002A;</sup>64 (layer 6). The output of the layer 6 is convoluted with the kernel size of 19 and the filter number of 128 in layer 7. An average-pooling layer is applied afterward with a size of 3 (layer 8). Layer 9 is the dropout layer, which is set to 50%. Layer 10 is connected 35 neurons to layer 11. Finally, layer 11 connected 5 neurons to the SoftMax layer. A rectifier linear unit (ReLU) is used as an activation function before every average pooling layer. All fully-connected layer is applied with the L2 factor of 5 to reduce overfitting.</p>
</sec>
<sec>
<title>Experiment Setup</title>
<p>In this paper, we ran a total of 60 epochs with a batch size of 36. The learn rate drop factor, learn rate drop period, and learning rate parameters are set to 0.1, 20, 10<sup>&#x02212;3</sup>. The parameters of all proposed classifiers in <xref ref-type="table" rid="T3">Table 3</xref> are selected for use based on the best results of ten-fold cross validation.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>All parameters of all proposed classifiers for this work.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Network</bold></th>
<th valign="top" align="left"><bold>Parameters</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">1D-CNN</td>
<td valign="top" align="left">Epoch 60, MiniBatchSize 36, Optimizer adam, LearnRateDropFactor 0.1, LearnRateDropPeriod 20, InitialLearnRate 10e-3</td>
</tr>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="left">NumberOfTree 500, MTry 18</td>
</tr>
<tr>
<td valign="top" align="left">BP</td>
<td valign="top" align="left">MaxEpoch 100, LearnRate 0.1, TrainingFunction trainrp, TrainingGoal 10e-3, Hidden node 20</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Referring to <xref ref-type="table" rid="T3">Table 3</xref>, &#x0201C;MaxEpoch&#x0201D; is the maximum number of epochs to use for training. &#x0201C;MiniBatchSize&#x0201D; is the size of the mini-batch to use for each training iteration. A mini-batch is a subset of the training set that is used to evaluate the gradient of the loss function and update the weights. &#x0201C;InitialLearnRate&#x0201D; is the initial learning rate used for training. &#x0201C;LearnRateDropPeriod&#x0201D; is the number of epochs that passes between adjustments to the learning rate during training. &#x0201C;LearnRateDropFactor&#x0201D; is the multiplicative factor by which the learning rate drops during training. &#x0201C;Optimizer&#x0201D; is Adam (adaptive moment estimation) optimizer that we used. &#x0201C;NumberOfTree&#x0201D; is the number of trees that were used in random forest. &#x0201C;MTry&#x0201D; of the random forest is the number of features that were randomly selected in each split. The training function of the BP network is &#x0201C;trainrp,&#x0201D; which updates weight and bias values according to the resilient backpropagation algorithm. &#x0201C;TrainingGoal&#x0201D; is the mean squared error needing to reach at the end of the training. &#x0201C;Hidden node&#x0201D; is the number of neurons in the hidden layer.</p>
<p>In this paper, we construct BP neural network and Random Forest network for better comparison analysis, using the identical datasets. The BP network has 360 nodes in the input layer, correlated with 360 features of each sample has 20 neurons in the hidden layer, and 5 nodes in the output layer that represents five sub-classes; the largest value among the five neurons will be deemed as the classification output. On the other hand, the Random Forest network builds many decision trees and selects features randomly from random samples with bagging strategy, and uses the tress to vote for the input vector to get a class label. The input of each sample has 360 features, the number of the tree is set to 500, and 18 features are randomly selected to consider each split. The output is a five-class voter; the largest value will be taken as the classification output.</p>
<p>Furthermore, for fair comparison, we reconstruct a CNN network identical to Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>) to benchmark with our proposed algorithm (hereinafter referred to as the compared CNN network). To the best of our knowledge, there is no evidence in literature to study the micro-classification of heartbeats. Hence it is our biggest motivation to prove the viability of one-dimensional 12-layer CNN on the classifying the sub-classes of Arrhythmia by reconstructing other CNN network for better and fairer comparison.</p>
</sec>
<sec>
<title>Evaluation Index</title>
<p>In order to evaluate and compare the classification effects of each model more accurately, this paper uses confusion matrix, accuracy (Acc), sensitivity (Sen), specificity (Spe), and positive prediction rate (Ppr) (Kiranyaz et al., <xref ref-type="bibr" rid="B13">2015</xref>). Among them, the accuracy rate represents the ability to detect the real situation of the sample; the sensitivity represents the ability to distinguish various diseases; the specificity represents the ability to detect negatively for a certain disease; the positive predication represents the rate that proportion of positive identifications is actually correct. The corresponding expressions are formula (4&#x02013;7):</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M16"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>A</mml:mi><mml:mi>c</mml:mi><mml:mi>c</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E5"><label>(5)</label><mml:math id="M17"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>S</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(6)</label><mml:math id="M18"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>S</mml:mi><mml:mi>p</mml:mi><mml:mi>e</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E7"><label>(7)</label><mml:math id="M19"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mi>p</mml:mi><mml:mi>r</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where TP stands for True Positive, TN stands for True Negative, FP stands for False Positive, FN stands for False Negative, FP stands for False Positive (Zubair et al., <xref ref-type="bibr" rid="B38">2016</xref>).</p>
</sec>
</sec>
<sec id="s4">
<title>Result and Discussion</title>
<p>The proposed CNN algorithm is trained on a PC with Intel i5-7300HQ processors with 16GB of RAM and GTX1050 as GPU. It takes &#x0007E;4236 s to complete training, and &#x0007E;11 h to complete ten-fold cross validation. The average classification time of a single sample is 0.242 milliseconds. The implementation of the algorithm is using the MATLAB Deep Learning Toolbox and MATLAB Neural Network Toolbox.</p>
<p>The confusion matrix in <xref ref-type="table" rid="T4">Table 4</xref> shows the classification results of the proposed CNN network. The overall classification accuracy rate of the five micro-class classification of heartbeats reaches 97.41%, and the positive rate and specificity of each category are over 90%, which has fully illustrated the effectiveness of the model. However, despite the data enhancement, the total number of C4 samples in the data set is still relatively small compared to the other four categories. Hence, the model&#x00027;s sensitivity to the C4 type is quite low.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>The confusion matrix of the proposed CNN network.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="center" colspan="10"><bold>The proposed CNN network</bold></th>
</tr>
</thead>
<tbody>
<tr style="border-bottom: thin solid #000000;">
<td valign="top" align="left"><bold>Class</bold></td>
<td valign="top" align="center"><bold>C1</bold></td>
<td valign="top" align="center"><bold>C2</bold></td>
<td valign="top" align="center"><bold>C3</bold></td>
<td valign="top" align="center"><bold>C4</bold></td>
<td valign="top" align="center"><bold>C5</bold></td>
<td valign="top" align="center"><bold>Sen(%)</bold></td>
<td valign="top" align="center"><bold>Spe(%)</bold></td>
<td valign="top" align="center"><bold>Ppr(%)</bold></td>
<td valign="top" align="center"><bold>Acc(%)</bold></td>
</tr> <tr>
<td valign="top" align="left">C1</td>
<td valign="top" align="center">7,216</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">453</td>
<td valign="top" align="center">7</td>
<td valign="top" align="center">95.41</td>
<td valign="top" align="center">98.15</td>
<td valign="top" align="center">93.99</td>
<td valign="top" align="center">97.41</td>
</tr>
<tr>
<td valign="top" align="left">C2</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">6,609</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">99.96</td>
<td valign="top" align="center">99.99</td>
<td valign="top" align="center">99.97</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C3</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">7,149</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">99.78</td>
<td valign="top" align="center">99.98</td>
<td valign="top" align="center">99.93</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C4</td>
<td valign="top" align="center">333</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">4,297</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">90.31</td>
<td valign="top" align="center">98.77</td>
<td valign="top" align="center">92.67</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C5</td>
<td valign="top" align="center">13</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">14</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">6,312</td>
<td valign="top" align="center">99.81</td>
<td valign="top" align="center">99.88</td>
<td valign="top" align="center">99.51</td>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="T5">Tables 5</xref>&#x02013;<xref ref-type="table" rid="T7">7</xref> illustrate the confusion matrix of BP neural network, Random Forest network, and the compared CNN network, respectively. The result shows the accuracy rate of the proposed CNN algorithm is 97.41%, which is 10.16% higher than the BP neural network, 1.69% higher than the Random Forest, and 3.34% higher than the compared CNN network. Furthermore, the sensitivity and specificity of the proposed CNN network are higher than the other three networks. Compared with the traditional machine learning methods like BP network and the random forests, the CNN network has the weight-sharing feature, which significantly accelerates the optimization process. The proposed CNN network also shows better performance on extracting local features, which is essential to classify different heartbeat types. The proposed CNN network uses the average-pooling layer instead of the max-pooling layer of the compared CNN network. The average-pooling layer can preserve the overall feature of the input data, which will be beneficial to classify the ECG signal. Also, the proposed CNN network has one more alternating convolution and pooling layer than the compared CNN network. The filter size and number of each convolution layer are also larger than the compared CNN network.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>The confusion matrix of the BP network.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="center" colspan="10"><bold>BP network</bold></th>
</tr>
</thead>
<tbody>
<tr style="border-bottom: thin solid #000000;">
<td valign="top" align="left"><bold>Class</bold></td>
<td valign="top" align="center"><bold>C1</bold></td>
<td valign="top" align="center"><bold>C2</bold></td>
<td valign="top" align="center"><bold>C3</bold></td>
<td valign="top" align="center"><bold>C4</bold></td>
<td valign="top" align="center"><bold>C5</bold></td>
<td valign="top" align="center"><bold>Sen(%)</bold></td>
<td valign="top" align="center"><bold>Spe(%)</bold></td>
<td valign="top" align="center"><bold>Ppr(%)</bold></td>
<td valign="top" align="center"><bold>Acc(%)</bold></td>
</tr> <tr>
<td valign="top" align="left">C1</td>
<td valign="top" align="center">6,573</td>
<td valign="top" align="center">79</td>
<td valign="top" align="center">14</td>
<td valign="top" align="center">1,580</td>
<td valign="top" align="center">252</td>
<td valign="top" align="center">86.91</td>
<td valign="top" align="center">92.25</td>
<td valign="top" align="center">77.34</td>
<td valign="top" align="center">87.25</td>
</tr>
<tr>
<td valign="top" align="left">C2</td>
<td valign="top" align="center">354</td>
<td valign="top" align="center">6,430</td>
<td valign="top" align="center">13</td>
<td valign="top" align="center">242</td>
<td valign="top" align="center">33</td>
<td valign="top" align="center">97.24</td>
<td valign="top" align="center">97.51</td>
<td valign="top" align="center">90.92</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C3</td>
<td valign="top" align="center">33</td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">6,726</td>
<td valign="top" align="center">65</td>
<td valign="top" align="center">186</td>
<td valign="top" align="center">93.87</td>
<td valign="top" align="center">98.85</td>
<td valign="top" align="center">95.88</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C4</td>
<td valign="top" align="center">472</td>
<td valign="top" align="center">7</td>
<td valign="top" align="center">11</td>
<td valign="top" align="center">2,732</td>
<td valign="top" align="center">27</td>
<td valign="top" align="center">57.41</td>
<td valign="top" align="center">98.13</td>
<td valign="top" align="center">84.08</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C5</td>
<td valign="top" align="center">132</td>
<td valign="top" align="center">90</td>
<td valign="top" align="center">401</td>
<td valign="top" align="center">139</td>
<td valign="top" align="center">5,826</td>
<td valign="top" align="center">92.12</td>
<td valign="top" align="center">97.08</td>
<td valign="top" align="center">88.43</td>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>The confusion matrix of random forest.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="center" colspan="10" style="border-bottom: thin solid #000000;"><bold>Random forest network</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><bold>Class</bold></td>
<td valign="top" align="center"><bold>C1</bold></td>
<td valign="top" align="center"><bold>C2</bold></td>
<td valign="top" align="center"><bold>C3</bold></td>
<td valign="top" align="center"><bold>C4</bold></td>
<td valign="top" align="center"><bold>C5</bold></td>
<td valign="top" align="center"><bold>Sen(%)</bold></td>
<td valign="top" align="center"><bold>Spe(%)</bold></td>
<td valign="top" align="center"><bold>Ppr(%)</bold></td>
<td valign="top" align="center"><bold>Acc(%)</bold></td>
</tr> <tr>
<td valign="top" align="left">C1</td>
<td valign="top" align="center">7,162</td>
<td valign="top" align="center">7</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">766</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">94.69</td>
<td valign="top" align="center">96.87</td>
<td valign="top" align="center">90.22</td>
<td valign="top" align="center">95.72</td>
</tr>
<tr>
<td valign="top" align="left">C2</td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">6,577</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">24</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">99.47</td>
<td valign="top" align="center">99.86</td>
<td valign="top" align="center">99.47</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C3</td>
<td valign="top" align="center">11</td>
<td valign="top" align="center">12</td>
<td valign="top" align="center">7,044</td>
<td valign="top" align="center">24</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">98.31</td>
<td valign="top" align="center">99.81</td>
<td valign="top" align="center">99.33</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C4</td>
<td valign="top" align="center">364</td>
<td valign="top" align="center">10</td>
<td valign="top" align="center">8</td>
<td valign="top" align="center">3,934</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">82.68</td>
<td valign="top" align="center">98.60</td>
<td valign="top" align="center">91.08</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C5</td>
<td valign="top" align="center">20</td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">108</td>
<td valign="top" align="center">10</td>
<td valign="top" align="center">6,318</td>
<td valign="top" align="center">99.90</td>
<td valign="top" align="center">99.44</td>
<td valign="top" align="center">97.77</td>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>The confusion matrix of the compared CNN network (Acharya et al., <xref ref-type="bibr" rid="B2">2017b</xref>).</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="center" colspan="10"><bold>Acharya&#x00027;s CNN network</bold></th>
</tr>
</thead>
<tbody>
<tr style="border-bottom: thin solid #000000;">
<td valign="top" align="left"><bold>Class</bold></td>
<td valign="top" align="center"><bold>C1</bold></td>
<td valign="top" align="center"><bold>C2</bold></td>
<td valign="top" align="center"><bold>C3</bold></td>
<td valign="top" align="center"><bold>C4</bold></td>
<td valign="top" align="center"><bold>C5</bold></td>
<td valign="top" align="center"><bold>Sen(%)</bold></td>
<td valign="top" align="center"><bold>Spe(%)</bold></td>
<td valign="top" align="center"><bold>Ppr(%)</bold></td>
<td valign="top" align="center"><bold>Acc(%)</bold></td>
</tr> <tr>
<td valign="top" align="left">C1</td>
<td valign="top" align="center">6,873</td>
<td valign="top" align="center">11</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">1,043</td>
<td valign="top" align="center">9</td>
<td valign="top" align="center">90.87</td>
<td valign="top" align="center">95.71</td>
<td valign="top" align="center">86.73</td>
<td valign="top" align="center">94.07</td>
</tr>
<tr>
<td valign="top" align="left">C2</td>
<td valign="top" align="center">7</td>
<td valign="top" align="center">6,578</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">9</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">99.48</td>
<td valign="top" align="center">99.93</td>
<td valign="top" align="center">99.72</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C3</td>
<td valign="top" align="center">13</td>
<td valign="top" align="center">7</td>
<td valign="top" align="center">7,076</td>
<td valign="top" align="center">21</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">98.75</td>
<td valign="top" align="center">99.82</td>
<td valign="top" align="center">99.38</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C4</td>
<td valign="top" align="center">629</td>
<td valign="top" align="center">9</td>
<td valign="top" align="center">11</td>
<td valign="top" align="center">3,662</td>
<td valign="top" align="center">2</td>
<td valign="top" align="center">76.96</td>
<td valign="top" align="center">97.64</td>
<td valign="top" align="center">84.90</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">C5</td>
<td valign="top" align="center">41</td>
<td valign="top" align="center">7</td>
<td valign="top" align="center">73</td>
<td valign="top" align="center">23</td>
<td valign="top" align="center">6,310</td>
<td valign="top" align="center">99.77</td>
<td valign="top" align="center">99.44</td>
<td valign="top" align="center">97.76</td>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="T8">Table 8</xref> depicts the classification performance of the different networks applied to the original and denoising data sets. For the proposed CNN network, the accuracy rates of the original and de-noised data are 96.9 and 97.2%, respectively, and the accuracy rate of the classification of the original data is only decreased by 0.3%, which shows that the network proposed in this paper has a degree of noise resistance. The accuracy of BP neural network classification of raw data is dropped 3.3% compared to the de-noised data. The random forest is 0.5% lower, and the compared CNN network is also 0.5% lower, all of which are suffered more loss than the proposed CNN network.</p>
<table-wrap position="float" id="T8">
<label>Table 8</label>
<caption><p>The performance of four different approaches.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Network</bold></th>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="center"><bold>Acc(%)</bold></th>
<th valign="top" align="center"><bold>Sen(%)</bold></th>
<th valign="top" align="center"><bold>Spe(%)</bold></th>
<th valign="top" align="center"><bold>Ppr(%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">The proposed CNN</td>
<td valign="top" align="left">Denoising</td>
<td valign="top" align="center">97.41</td>
<td valign="top" align="center">97.05</td>
<td valign="top" align="center">99.35</td>
<td valign="top" align="center">97.21</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Raw</td>
<td valign="top" align="center">97.02</td>
<td valign="top" align="center">96.57</td>
<td valign="top" align="center">99.25</td>
<td valign="top" align="center">96.84</td>
</tr>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="left">Denoising</td>
<td valign="top" align="center">95.72</td>
<td valign="top" align="center">95.01</td>
<td valign="top" align="center">98.92</td>
<td valign="top" align="center">95.58</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Raw</td>
<td valign="top" align="center">95.08</td>
<td valign="top" align="center">94.50</td>
<td valign="top" align="center">98.81</td>
<td valign="top" align="center">95.09</td>
</tr>
<tr>
<td valign="top" align="left">BP</td>
<td valign="top" align="left">Denoising</td>
<td valign="top" align="center">87.25</td>
<td valign="top" align="center">85.51</td>
<td valign="top" align="center">96.77</td>
<td valign="top" align="center">87.33</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Raw</td>
<td valign="top" align="center">85.98</td>
<td valign="top" align="center">84.94</td>
<td valign="top" align="center">96.46</td>
<td valign="top" align="center">85.14</td>
</tr>
<tr>
<td valign="top" align="left">Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>)</td>
<td valign="top" align="left">Denoising</td>
<td valign="top" align="center">94.07</td>
<td valign="top" align="center">93.17</td>
<td valign="top" align="center">98.51</td>
<td valign="top" align="center">93.67</td>
</tr>
<tr>
<td/>
<td valign="top" align="left">Raw</td>
<td valign="top" align="center">93.22</td>
<td valign="top" align="center">92.23</td>
<td valign="top" align="center">98.30</td>
<td valign="top" align="center">92.79</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="T9">Table 9</xref> lists the standard deviations of the sensitivities, specificities, and positive rates of four different networks. According to <xref ref-type="table" rid="T9">Table 9</xref>, the standard deviations of the four metrics of the proposed CNN network are less than the BP neural network, random forest, and the compared CNN network. The result indicates that the model&#x00027;s performance against multiple classifications is relatively stable, and the recognition effect of each classification is consistent, which shows the robustness of the model.</p>
<table-wrap position="float" id="T9">
<label>Table 9</label>
<caption><p>The standard deviation of the four different networks.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Network</bold></th>
<th valign="top" align="center"><bold>Sen</bold></th>
<th valign="top" align="center"><bold>Spe</bold></th>
<th valign="top" align="center"><bold>Ppr</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">The proposed CNN</td>
<td valign="top" align="center">3.79</td>
<td valign="top" align="center">0.75</td>
<td valign="top" align="center">3.20</td>
</tr>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="center">6.43</td>
<td valign="top" align="center">1.18</td>
<td valign="top" align="center">4.07</td>
</tr>
<tr>
<td valign="top" align="left">BP</td>
<td valign="top" align="center">14.44</td>
<td valign="top" align="center">2.33</td>
<td valign="top" align="center">6.29</td>
</tr>
<tr>
<td valign="top" align="left">Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>)</td>
<td valign="top" align="center">8.75</td>
<td valign="top" align="center">1.62</td>
<td valign="top" align="center">6.49</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="T10">Table 10</xref> shows the existing literature of ECG classification. The dataset used in these literatures is not exactly the same, but the comparison is useful because classification is all on the same MIT-BIH database. Li and Zhou (<xref ref-type="bibr" rid="B15">2016</xref>) applied random forest classifier to recognize five main classes (N, Q, S, V, F), which achieved a 94.61% accuracy rate. Osowski et al. (<xref ref-type="bibr" rid="B21">2004</xref>) used an SVM classifier that achieved 98.18% accuracy on 13 classes of heartbeats. Martis et al. (<xref ref-type="bibr" rid="B17">2013</xref>) obtained 94.52% performance on five main classes (N, Q, S, V, F) in their studies. P&#x00142;awiak and Acharya (<xref ref-type="bibr" rid="B25">2020</xref>) used a deep genetic ensemble of classifiers to classify long-duration ECG signal, which achieved 94.6% of accuracy on 17 arrhythmia classes in the MIT-BIH database. Gao et al. (<xref ref-type="bibr" rid="B8">2019</xref>) implemented an effective long short-term memory (LSTM) recurrence network model to classify 8 types of heartbeats (N, LBBB, RBBB, APC,NESC, ABERR, NPC, AESC). Atal and Singh (<xref ref-type="bibr" rid="B5">2020</xref>) proposed an optimization-based deep convolutional neural network, achieving 93.19% accuracy on five main classes (N, Q, S, V, F). Acharya et al. (<xref ref-type="bibr" rid="B1">2017a</xref>) achieved 95.22% of accuracy on the classification of two types of heartbeats only (Normal and MI) with an 11-layer CNN network. The architecture of this CNN network was 4 alternating convolutions and max-pooling layers, followed by 3 fully-connected layers. In another work, Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>) used 9-layer CNN network to classify 5 main classes (N, Q, S, V, F) and they achieved 94.03% of accuracy. For fair comparison, CNN model of Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>) is reconstructed to benchmark with the proposed model. Zubair et al. (<xref ref-type="bibr" rid="B38">2016</xref>) obtained a 92.70% classification performance rate for five main classes (N, Q, S, V, F) using an 8-layer CNN-based network. Zubair et al. (<xref ref-type="bibr" rid="B38">2016</xref>) implemented 3 alternating convolution and max-pooling layer in the CNN, followed by one MLP layer. Generally speaking, the proposed CNN network achieved relatively high accuracy on the 5 micro-classes of heartbeats classification.</p>
<table-wrap position="float" id="T10">
<label>Table 10</label>
<caption><p>The performance comparison with other algorithms.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>References</bold></th>
<th valign="top" align="center"><bold>No. of classes</bold></th>
<th valign="top" align="left"><bold>Feature set</bold></th>
<th valign="top" align="left"><bold>Classifier</bold></th>
<th valign="top" align="center"><bold>Accuracy</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Li et al. (<xref ref-type="bibr" rid="B14">2014</xref>)</td>
<td valign="top" align="center">5</td>
<td valign="top" align="left">WPE &#x0002B; RR</td>
<td valign="top" align="left">RF</td>
<td valign="top" align="center">94.61%</td>
</tr>
<tr>
<td valign="top" align="left">Osowski et al. (<xref ref-type="bibr" rid="B21">2004</xref>)</td>
<td valign="top" align="center">13</td>
<td valign="top" align="left">HOS &#x0002B; Herminte</td>
<td valign="top" align="left">SVM</td>
<td valign="top" align="center">98.18%</td>
</tr>
<tr>
<td valign="top" align="left">Martis et al. (<xref ref-type="bibr" rid="B17">2013</xref>)</td>
<td valign="top" align="center">5</td>
<td valign="top" align="left">Cumulant &#x0002B; PCA</td>
<td valign="top" align="left">NN</td>
<td valign="top" align="center">94.52%</td>
</tr>
<tr>
<td valign="top" align="left">Acharya et al. (<xref ref-type="bibr" rid="B1">2017a</xref>)</td>
<td valign="top" align="center">2</td>
<td valign="top" align="left">End-to-end</td>
<td valign="top" align="left">CNN</td>
<td valign="top" align="center">95.22%</td>
</tr>
<tr>
<td valign="top" align="left">Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>)</td>
<td valign="top" align="center">5</td>
<td valign="top" align="left">End-to-end</td>
<td valign="top" align="left">CNN</td>
<td valign="top" align="center">94.03%</td>
</tr>
<tr>
<td valign="top" align="left">Zubair et al. (<xref ref-type="bibr" rid="B38">2016</xref>)</td>
<td valign="top" align="center">5</td>
<td valign="top" align="left">End-to-end</td>
<td valign="top" align="left">CNN</td>
<td valign="top" align="center">92.70%</td>
</tr>
<tr>
<td valign="top" align="left">P&#x00142;awiak and Acharya (<xref ref-type="bibr" rid="B25">2020</xref>)</td>
<td valign="top" align="center">17</td>
<td valign="top" align="left">Frequency components</td>
<td valign="top" align="left">DGEC</td>
<td valign="top" align="center">94.60%</td>
</tr>
<tr>
<td valign="top" align="left">Atal and Singh (<xref ref-type="bibr" rid="B5">2020</xref>)</td>
<td valign="top" align="center">5</td>
<td valign="top" align="left">Gabor Filter &#x0002B; Wavelet</td>
<td valign="top" align="left">BaROA-DCNN</td>
<td valign="top" align="center">93.19%</td>
</tr>
<tr>
<td valign="top" align="left">The proposed CNN network</td>
<td valign="top" align="center">5</td>
<td valign="top" align="left">Wavalet</td>
<td valign="top" align="left">CNN</td>
<td valign="top" align="center">97.20%</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Statistical variance test is carried out to study the differences of the classification performance of the proposed CNN algorithm as well as the different approaches for the raw and denoised dataset. Analysis of variance is a collection and representation of statistical model. The associated estimation procedures are used to analyze the differences among group means in a sample. The results are recorded in <xref ref-type="table" rid="T11">Tables 11</xref>&#x02013;<xref ref-type="table" rid="T13">13</xref>. The <italic>p</italic>-value is the probability of obtaining test results at least as extreme as the actually observed results, under the assumption that the null hypothesis is correct. In the statistical significance test, the smaller the p-value, the stronger the evidence we should reject the null hypothesis. For analysis in <xref ref-type="table" rid="T11">Table 11</xref>, the null hypothesis is that the denoised data and raw data perform equally well. Significance level is set to 0.05. The <italic>p</italic>-value obtained from <xref ref-type="table" rid="T11">Table 11</xref> shows 0.011 for accuracy, 0.006 for sensitivity, 0.009 for specialty and 0.027 for positive prediction rate. It clearly demonstrates all of the <italic>p</italic>-values are &#x0003C;0.05, therefore it has the full evidence of rejecting the null hypothesis. Thus, the effect of denoising is very significant in the classification of the micro-heartbeats type.</p>
<table-wrap position="float" id="T11">
<label>Table 11</label>
<caption><p>Variance test analysis of the proposed CNN network.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Class</bold></th>
<th valign="top" align="center"><bold>Denoise data</bold></th>
<th valign="top" align="center"><bold>Raw data</bold></th>
<th valign="top" align="center"><bold>p-value</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Acc(%)</td>
<td valign="top" align="center">97.41 &#x000B1; 0.27</td>
<td valign="top" align="center">97.02 &#x000B1; 0.34</td>
<td valign="top" align="center">0.011</td>
</tr>
<tr>
<td valign="top" align="left">Sen(%)</td>
<td valign="top" align="center">97.05 &#x000B1; 0.31</td>
<td valign="top" align="center">96.57 &#x000B1; 0.38</td>
<td valign="top" align="center">0.006</td>
</tr>
<tr>
<td valign="top" align="left">Spe(%)</td>
<td valign="top" align="center">99.35 &#x000B1; 0.07</td>
<td valign="top" align="center">99.25 &#x000B1; 0.08</td>
<td valign="top" align="center">0.009</td>
</tr>
<tr>
<td valign="top" align="left">Ppr(%)</td>
<td valign="top" align="center">97.22 &#x000B1; 0.31</td>
<td valign="top" align="center">96.84 &#x000B1; 0.39</td>
<td valign="top" align="center">0.027</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T12">
<label>Table 12</label>
<caption><p>Variance test analysis of different classifiers on raw data.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Class</bold></th>
<th valign="top" align="center"><bold>The proposed CNN</bold></th>
<th valign="top" align="center"><bold>Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>)</bold></th>
<th valign="top" align="center"><bold>Random forest</bold></th>
<th valign="top" align="center"><bold>BP network</bold></th>
<th valign="top" align="center"><bold><italic>p</italic>-value</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Acc(%)</td>
<td valign="top" align="center">97.02 &#x000B1; 0.34</td>
<td valign="top" align="center">93.22 &#x000B1; 0.49</td>
<td valign="top" align="center">95.08 &#x000B1; 0.39</td>
<td valign="top" align="center">85.97 &#x000B1; 0.57</td>
<td valign="top" align="center">0.0001</td>
</tr>
<tr>
<td valign="top" align="left">Sen(%)</td>
<td valign="top" align="center">96.57 &#x000B1; 0.38</td>
<td valign="top" align="center">92.23 &#x000B1; 0.50</td>
<td valign="top" align="center">94.50 &#x000B1; 0.46</td>
<td valign="top" align="center">84.94 &#x000B1; 0.62</td>
<td valign="top" align="center">0.0001</td>
</tr>
<tr>
<td valign="top" align="left">Spe(%)</td>
<td valign="top" align="center">99.25 &#x000B1; 0.08</td>
<td valign="top" align="center">98.30 &#x000B1; 0.12</td>
<td valign="top" align="center">98.81 &#x000B1; 0.10</td>
<td valign="top" align="center">96.49 &#x000B1; 0.14</td>
<td valign="top" align="center">0.0001</td>
</tr>
<tr>
<td valign="top" align="left">Ppr(%)</td>
<td valign="top" align="center">96.84 &#x000B1; 0.39</td>
<td valign="top" align="center">92.81 &#x000B1; 0.57</td>
<td valign="top" align="center">95.10 &#x000B1; 0.40</td>
<td valign="top" align="center">85.14 &#x000B1; 0.65</td>
<td valign="top" align="center">0.0001</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T13">
<label>Table 13</label>
<caption><p>Variance test analysis of different classifiers on denoised data.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Class</bold></th>
<th valign="top" align="center"><bold>The proposed CNN</bold></th>
<th valign="top" align="center"><bold>Acharya et al. (<xref ref-type="bibr" rid="B2">2017b</xref>)</bold></th>
<th valign="top" align="center"><bold>Random forest</bold></th>
<th valign="top" align="center"><bold>BP network</bold></th>
<th valign="top" align="center"><bold><italic>p</italic>-value</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Acc(%)</td>
<td valign="top" align="center">97.41 &#x000B1; 0.27</td>
<td valign="top" align="center">94.07 &#x000B1; 0.26</td>
<td valign="top" align="center">95.72 &#x000B1; 0.41</td>
<td valign="top" align="center">87.25 &#x000B1; 0.60</td>
<td valign="top" align="center">0.0001</td>
</tr>
<tr>
<td valign="top" align="left">Sen(%)</td>
<td valign="top" align="center">97.05 &#x000B1; 0.31</td>
<td valign="top" align="center">93.17 &#x000B1; 0.25</td>
<td valign="top" align="center">95.01 &#x000B1; 0.44</td>
<td valign="top" align="center">85.51 &#x000B1; 0.65</td>
<td valign="top" align="center">0.0001</td>
</tr>
<tr>
<td valign="top" align="left">Spe(%)</td>
<td valign="top" align="center">99.35 &#x000B1; 0.07</td>
<td valign="top" align="center">98.51 &#x000B1; 0.06</td>
<td valign="top" align="center">98.92 &#x000B1; 0.10</td>
<td valign="top" align="center">96.77 &#x000B1; 0.15</td>
<td valign="top" align="center">0.0001</td>
</tr>
<tr>
<td valign="top" align="left">Ppr(%)</td>
<td valign="top" align="center">97.22 &#x000B1; 0.31</td>
<td valign="top" align="center">93.68 &#x000B1; 0.38</td>
<td valign="top" align="center">95.58 &#x000B1; 0.45</td>
<td valign="top" align="center">87.34 &#x000B1; 0.70</td>
<td valign="top" align="center">0.0001</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>On the other hand, in <xref ref-type="table" rid="T12">Table 12</xref>, the null hypothesis is under raw data, all four classifiers perform equally well. Significance level is still set to 0.05. The obtained <italic>p</italic>-values are too miniscule for accuracy, sensitivity, specialty, and positive prediction rate. All of the <italic>p</italic>-values are &#x0003C; 0.05, therefore it distinctly shows that the null hypothesis should be rejected. Thus, under raw data, it is obvious to interpret that the proposed CNN network performs much better than the rest of the classifiers.</p>
<p>In <xref ref-type="table" rid="T13">Table 13</xref>, the null hypothesis is under denoised data, all four classifiers perform equally well. Significance level is still set to 0.05. The obtained <italic>p</italic>-values are too miniscule for accuracy, sensitivity, specialty and positive prediction rate. All of the <italic>p</italic>-values are &#x0003C; 0.05, therefore it distinctly shows that the null hypothesis should be rejected. Thus, under denoised data, it is obvious to interpret that the proposed CNN network performs much better than the rest of the classifiers.</p>
</sec>
<sec sec-type="conclusions" id="s5">
<title>Conclusion</title>
<p>Cardiovascular disease is a major health problem in today&#x00027;s world. The early diagnosis of cardiac arrhythmia highly relies on the ECG. Unfortunately, the expert level of medical resources is rare, visually identify the ECG signal is challenging and time-consuming. Different from the existing literatures in which most of them classify the five main classes, such as the Non-ectopic, Supraventricular ectopic, Ventricular ectopic, Fusion, and Unknown, in the MIT-BIH Arrhythmia database, our paper pays more attention to specific micro-classes, namely the Normal, Left Bundle Branch Block, Right Bundle Branch Block, Atrial Premature Beats, Premature Ventricular Beats. Compared with the BP neural network, random forests, and other CNN networks, it is worth to highlight that the proposed CNN network has relatively higher accuracy and robustness. The proposed CNN network shows an outstanding performance in the overall classification accuracy of 97.41%, sensitivity of 97.05%, specificity of 99.35%, and positive prediction rate of 97.21% on the classification of the micro-classes of Arrhythmia dataset.</p>
<p>The advantages of the proposed CNN network have been put to evidence. It is endowed with an ability to effectively process the non-filtered dataset with its potential anti-noise features. Besides that, ten-fold cross-validation is implemented in this work to further demonstrate the robustness of the network. In addition, this paper presents an analysis of the classification of micro-classes of the ECG signal with comparison to some techniques of machine learning such as BP and Random Forest. We add values to the research community by discussing the results of the classification of less popular micro-classes of Arrhythmia that can be served as a good source of benchmark literature to other researchers in this field for further research. One possible setback of the proposed solution is that it is computationally intensive to train the network, due to deep learning series is often attributed to large scale data required for training.</p>
<p>As for future work, it would be interesting to explore the use of optimization techniques to find a feasible design and solution. The limitation of our study is that we have yet to apply any optimization techniques to optimize the model parameters and we believe that with the implementation of the optimization, it will be able to further elevate the performance of the proposed solution to the next level.</p>
</sec>
<sec sec-type="data-availability-statement" id="s6">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="https://physionet.org/content/mitdb/1.0.0/">https://physionet.org/content/mitdb/1.0.0/</ext-link>.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>SYW: conceptualization, methods and materials, and supervision and review. MW and YL: data curation. MW, YL, and WY: formal analysis. WY: resources. MW: validation. MW, YL, WY, and SYW: writing. All authors contributed to the article and approved the submitted version.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<ack><p>This work was supported by Xiamen University Malaysia Research Fund under Grant Number IECE/0001.</p></ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Acharya</surname> <given-names>U. R.</given-names></name> <name><surname>Fujita</surname> <given-names>H.</given-names></name> <name><surname>Oh</surname> <given-names>S. L.</given-names></name> <name><surname>Hagiwara</surname> <given-names>Y.</given-names></name> <name><surname>Tan</surname> <given-names>J. H.</given-names></name> <name><surname>Adam</surname> <given-names>M.</given-names></name></person-group> (<year>2017a</year>). <article-title>Application of deep convolutional neural network for automated detection of myocardial infarction using ECG signals</article-title>. <source>Inform. Sci.</source> <volume>415</volume>, <fpage>190</fpage>&#x02013;<lpage>198</lpage>. <pub-id pub-id-type="doi">10.1016/j.ins.2017.06.027</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Acharya</surname> <given-names>U. R.</given-names></name> <name><surname>Oh</surname> <given-names>S. L.</given-names></name> <name><surname>Hagiwara</surname> <given-names>Y.</given-names></name> <name><surname>Tan</surname> <given-names>J. H.</given-names></name> <name><surname>Adam</surname> <given-names>M.</given-names></name> <name><surname>Gertych</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2017b</year>). <article-title>A deep convolutional neural network model to classify heartbeats</article-title>. <source>Comp. Biol. Med.</source> <volume>89</volume>, <fpage>389</fpage>&#x02013;<lpage>396</lpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2017.08.022</pub-id><pub-id pub-id-type="pmid">28869899</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ahlstrom</surname> <given-names>M.</given-names></name> <name><surname>Tompkins</surname> <given-names>W.</given-names></name></person-group> (<year>1985</year>). <article-title>Digital filters for real-time ECG signal processing using microprocessors</article-title>. <source>IEEE Trans. Biomed. Engin.</source> <volume>9</volume>, <fpage>708</fpage>&#x02013;<lpage>713</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.1985.325589</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alfaouri</surname> <given-names>M.</given-names></name> <name><surname>Daqrouq</surname> <given-names>K.</given-names></name></person-group> (<year>2008</year>). <article-title>ECG signal denoising by wavelet transform thresholding</article-title>. <source>Am. J. Appl. Sci.</source> <volume>5</volume>, <fpage>276</fpage>&#x02013;<lpage>281</lpage>. <pub-id pub-id-type="doi">10.3844/ajassp.2008.276.281</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Atal</surname> <given-names>D. K.</given-names></name> <name><surname>Singh</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>Arrhythmia classification with ECG signals based on the optimization-enabled deep convolutional neural network</article-title>. <source>Comp. Methods Prog. Biomed.</source> <volume>196</volume>:<fpage>105607</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2020.105607</pub-id><pub-id pub-id-type="pmid">32593973</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Awal</surname> <given-names>M. A.</given-names></name> <name><surname>Mostafa</surname> <given-names>S. S.</given-names></name> <name><surname>Ahmad</surname> <given-names>M.</given-names></name> <name><surname>Rashid</surname> <given-names>M. A.</given-names></name></person-group> (<year>2014</year>). <article-title>An adaptive level dependent wavelet thresholding for ECG denoising</article-title>. <source>Biocyber. Biomed. Engin.</source> <volume>34</volume>, <fpage>238</fpage>&#x02013;<lpage>249</lpage>. <pub-id pub-id-type="doi">10.1016/j.bbe.2014.03.002</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Bazi</surname> <given-names>Y.</given-names></name> <name><surname>Alajlan</surname> <given-names>N.</given-names></name> <name><surname>AlHichri</surname> <given-names>H.</given-names></name> <name><surname>Malek</surname> <given-names>S.</given-names></name></person-group> (<year>2013</year>). <article-title>Domain adaptation methods for ECG classification</article-title>, in <source>2013 International Conference on Computer Medical Applications (ICCMA)</source> (<publisher-loc>Sousse</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>4</lpage>. <pub-id pub-id-type="doi">10.1109/ICCMA.2013.6506156</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>H.</given-names></name> <name><surname>Lu</surname> <given-names>P.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name></person-group> (<year>2019</year>). <article-title>An effective LSTM recurrent network to detect arrhythmia on imbalanced ECG dataset</article-title>. <source>J. Healthcare Engin.</source> <volume>2019</volume>:<fpage>6320651</fpage>. <pub-id pub-id-type="doi">10.1155/2019/6320651</pub-id><pub-id pub-id-type="pmid">31737240</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guo</surname> <given-names>S.-L.</given-names></name> <name><surname>Han</surname> <given-names>L.-N.</given-names></name> <name><surname>Liu</surname> <given-names>H.-W.</given-names></name> <name><surname>Si</surname> <given-names>Q.-J.</given-names></name> <name><surname>Kong</surname> <given-names>D.-F.</given-names></name> <name><surname>Guo</surname> <given-names>F.-S.</given-names></name></person-group> (<year>2016</year>). <article-title>The future of remote ECG monitoring systems</article-title>. <source>J Geriat. Cardiol.</source> <volume>13</volume>, <fpage>528</fpage>. <pub-id pub-id-type="doi">10.11909/j.issn.1671-5411.2016.06.015</pub-id><pub-id pub-id-type="pmid">27582770</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Homaeinezhad</surname> <given-names>M. R.</given-names></name> <name><surname>Atyabi</surname> <given-names>S. A.</given-names></name> <name><surname>Tavakkoli</surname> <given-names>E.</given-names></name> <name><surname>Toosi</surname> <given-names>H. N.</given-names></name> <name><surname>Ghaffari</surname> <given-names>A.</given-names></name> <name><surname>Ebrahimpour</surname> <given-names>R.</given-names></name></person-group> (<year>2012</year>). <article-title>ECG arrhythmia recognition via a neuro-SVM-KNN hybrid classifier with virtual QRS image-based geometrical features</article-title>. <source>Exp. Syst. Applic.</source> <volume>39</volume>, <fpage>2047</fpage>&#x02013;<lpage>2058</lpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2011.08.025</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>W.</given-names></name> <name><surname>Kong</surname> <given-names>S. G.</given-names></name></person-group> (<year>2007</year>). <article-title>Block-based neural networks for personalized ECG signal classification</article-title>. <source>IEEE Trans. Neural Netw.</source> <volume>18</volume>, <fpage>1750</fpage>&#x02013;<lpage>1761</lpage>. <pub-id pub-id-type="doi">10.1109/TNN.2007.900239</pub-id><pub-id pub-id-type="pmid">18051190</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kandala</surname> <given-names>R. N.</given-names></name> <name><surname>Dhuli</surname> <given-names>R.</given-names></name> <name><surname>P&#x00142;awiak</surname> <given-names>P.</given-names></name> <name><surname>Naik</surname> <given-names>G. R.</given-names></name> <name><surname>Moeinzadeh</surname> <given-names>H.</given-names></name> <name><surname>Gargiulo</surname> <given-names>G. D.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Towards real-time heartbeat classification: evaluation of nonlinear morphological features and voting method</article-title>. <source>Sensors</source> <volume>19</volume>:<fpage>5079</fpage>. <pub-id pub-id-type="doi">10.3390/s19235079</pub-id><pub-id pub-id-type="pmid">31766323</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kiranyaz</surname> <given-names>S.</given-names></name> <name><surname>Ince</surname> <given-names>T.</given-names></name> <name><surname>Gabbouj</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Real-time patient-specific ECG classification by 1-D convolutional neural networks</article-title>. <source>IEEE Trans. Biomed. Engin.</source> <volume>63</volume>, <fpage>664</fpage>&#x02013;<lpage>675</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.2015.2468589</pub-id><pub-id pub-id-type="pmid">26285054</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Q.</given-names></name> <name><surname>Cai</surname> <given-names>W.</given-names></name> <name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Zhou</surname> <given-names>Y.</given-names></name> <name><surname>Feng</surname> <given-names>D. D.</given-names></name> <name><surname>Chen</surname> <given-names>M.</given-names></name></person-group> (<year>2014</year>). <article-title>Medical image classification with convolutional neural network</article-title>, in <source>2014 13th International Conference on Control Automation Robotics &#x00026; Vision (ICARCV)</source> (<publisher-loc>Singapore</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>844</fpage>&#x02013;<lpage>848</lpage>. <pub-id pub-id-type="doi">10.1109/ICARCV.2014.7064414</pub-id></citation></ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>T.</given-names></name> <name><surname>Zhou</surname> <given-names>M.</given-names></name></person-group> (<year>2016</year>). <article-title>ECG classification using wavelet packet entropy and random forests</article-title>. <source>Entropy</source> <volume>18</volume>:<fpage>285</fpage>. <pub-id pub-id-type="doi">10.3390/e18080285</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Luz</surname> <given-names>E. J. D. S.</given-names></name> <name><surname>Nunes</surname> <given-names>T. M.</given-names></name> <name><surname>De Albuquerque</surname> <given-names>V. H. C.</given-names></name> <name><surname>Papa</surname> <given-names>J. P.</given-names></name> <name><surname>Menotti</surname> <given-names>D.</given-names></name></person-group> (<year>2013</year>). <article-title>ECG arrhythmia classification based on optimum-path forest</article-title>. <source>Exp. Syst. Applic.</source> <volume>40</volume>, <fpage>3561</fpage>&#x02013;<lpage>3573</lpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2012.12.063</pub-id></citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Martis</surname> <given-names>R. J.</given-names></name> <name><surname>Acharya</surname> <given-names>U. R.</given-names></name> <name><surname>Lim</surname> <given-names>C. M.</given-names></name> <name><surname>Mandana</surname> <given-names>K.</given-names></name> <name><surname>Ray</surname> <given-names>A. K.</given-names></name> <name><surname>Chakraborty</surname> <given-names>C.</given-names></name></person-group> (<year>2013</year>). <article-title>Application of higher order cumulant features for cardiac health diagnosis using ECG signals</article-title>. <source>Int. J. Neural Syst.</source> <volume>23</volume>:<fpage>1350014</fpage>. <pub-id pub-id-type="doi">10.1142/S0129065713500147</pub-id><pub-id pub-id-type="pmid">23746287</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Masko</surname> <given-names>D.</given-names></name> <name><surname>Hensman</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <source>The Impact of Imbalanced Training Data for Convolutional Neural Networks</source>. Dissertation. KTH, Skolan f&#x000F6;r datavetenskap och kommunikation (CSC). Retrieved from: <ext-link ext-link-type="uri" xlink:href="http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-166451">http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-166451</ext-link></citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mc Namara</surname> <given-names>K.</given-names></name> <name><surname>Alzubaidi</surname> <given-names>H.</given-names></name> <name><surname>Jackson</surname> <given-names>J. K.</given-names></name></person-group> (<year>2019</year>). <article-title>Cardiovascular disease as a leading cause of death: how are pharmacists getting involved?</article-title> <source>Integr. Pharm. Res. Pract.</source> <volume>8</volume>:<fpage>1</fpage>. <pub-id pub-id-type="doi">10.2147/IPRP.S133088</pub-id><pub-id pub-id-type="pmid">30788283</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Moody</surname> <given-names>G. B.</given-names></name> <name><surname>Mark</surname> <given-names>R. G.</given-names></name></person-group> (<year>2001</year>). <article-title>The impact of the MIT-BIH arrhythmia database</article-title>. <source>IEEE Engin. Med. Biol. Mag.</source> <volume>20</volume>, <fpage>45</fpage>&#x02013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.1109/51.932724</pub-id><pub-id pub-id-type="pmid">11446209</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Osowski</surname> <given-names>S.</given-names></name> <name><surname>Hoai</surname> <given-names>L. T.</given-names></name> <name><surname>Markiewicz</surname> <given-names>T.</given-names></name></person-group> (<year>2004</year>). <article-title>Support vector machine-based expert system for reliable heartbeat recognition</article-title>. <source>IEEE Trans. Biomed. Engin.</source> <volume>51</volume>, <fpage>582</fpage>&#x02013;<lpage>589</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.2004.824138</pub-id><pub-id pub-id-type="pmid">15072212</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pan</surname> <given-names>J.</given-names></name> <name><surname>Tompkins</surname> <given-names>W. J.</given-names></name></person-group> (<year>1985</year>). <article-title>A real-time QRS detection algorithm</article-title>. <source>IEEE Trans. Biomed. Engin.</source> <volume>3</volume>, <fpage>230</fpage>&#x02013;<lpage>236</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.1985.325532</pub-id><pub-id pub-id-type="pmid">3997178</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>P&#x00142;awiak</surname> <given-names>P.</given-names></name> <name><surname>Abdar</surname> <given-names>M.</given-names></name> <name><surname>Acharya</surname> <given-names>U. R.</given-names></name></person-group> (<year>2019</year>). <article-title>Application of new deep genetic cascade ensemble of SVM classifiers to predict the Australian credit scoring</article-title>. <source>Appl. Soft Comput.</source> <volume>84</volume>:<fpage>105740</fpage>. <pub-id pub-id-type="doi">10.1016/j.asoc.2019.105740</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>P&#x00142;awiak</surname> <given-names>P.</given-names></name> <name><surname>Abdar</surname> <given-names>M.</given-names></name> <name><surname>P&#x00142;awiak</surname> <given-names>J.</given-names></name> <name><surname>Makarenkov</surname> <given-names>V.</given-names></name> <name><surname>Acharya</surname> <given-names>U. R.</given-names></name></person-group> (<year>2020</year>). <article-title>DGHNL: A new deep genetic hierarchical network of learners for prediction of credit scoring</article-title>. <source>Inform. Sci.</source> <volume>516</volume>, <fpage>401</fpage>&#x02013;<lpage>418</lpage>. <pub-id pub-id-type="doi">10.1016/j.ins.2019.12.045</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>P&#x00142;awiak</surname> <given-names>P.</given-names></name> <name><surname>Acharya</surname> <given-names>U. R.</given-names></name></person-group> (<year>2020</year>). <article-title>Novel deep genetic ensemble of classifiers for arrhythmia detection using ECG signals</article-title>. <source>Neural Comp. Applic.</source> <volume>32</volume>, <fpage>11137</fpage>&#x02013;<lpage>11161</lpage>. <pub-id pub-id-type="doi">10.1007/s00521-018-03980-2</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Sarfraz</surname> <given-names>M.</given-names></name> <name><surname>Khan</surname> <given-names>A. A.</given-names></name> <name><surname>Li</surname> <given-names>F. F.</given-names></name></person-group> (<year>2014</year>). <article-title>Using independent component analysis to obtain feature space for reliable ECG Arrhythmia classification</article-title>, in <source>2014 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)</source> (<publisher-loc>Belfast</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>62</fpage>&#x02013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1109/BIBM.2014.6999249</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Singh</surname> <given-names>B. N.</given-names></name> <name><surname>Tiwari</surname> <given-names>A. K.</given-names></name></person-group> (<year>2006</year>). <article-title>Optimal selection of wavelet basis function applied to ECG signal denoising</article-title>. <source>Dig. Sign. Proc.</source> <volume>16</volume>, <fpage>275</fpage>&#x02013;<lpage>287</lpage>. <pub-id pub-id-type="doi">10.1016/j.dsp.2005.12.003</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>G. J.</given-names></name> <name><surname>Yuan</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>K.</given-names></name></person-group> (<year>2014</year>). <article-title>Simulation of arrhythmia mechanism based on ventricular electrical conduction model</article-title>. <source>J. Harbin Inst. Technol.</source> <volume>1</volume>:<fpage>8</fpage>.</citation></ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>J.</given-names></name> <name><surname>Ye</surname> <given-names>Y.</given-names></name> <name><surname>Pan</surname> <given-names>X.</given-names></name> <name><surname>Gao</surname> <given-names>X.</given-names></name></person-group> (<year>2015</year>). <article-title>Parallel-type fractional zero-phase filtering for ECG signal denoising</article-title>. <source>Biomed. Sign. Proc. Control</source> <volume>18</volume>, <fpage>36</fpage>&#x02013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2014.10.012</pub-id></citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>Y.</given-names></name> <name><surname>Xia</surname> <given-names>W.</given-names></name> <name><surname>Lin</surname> <given-names>M.</given-names></name> <name><surname>Huang</surname> <given-names>J.</given-names></name> <name><surname>Ni</surname> <given-names>B.</given-names></name> <name><surname>Dong</surname> <given-names>J.</given-names></name></person-group> (<year>2015</year>). <article-title>HCP: A flexible CNN framework for multi-label image classification</article-title>. <source>IEEE Transac. Patt. Anal. Mach. Intell.</source> <volume>38</volume>, <fpage>1901</fpage>&#x02013;<lpage>1907</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2015.2491929</pub-id><pub-id pub-id-type="pmid">26513778</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wong</surname> <given-names>S. Y.</given-names></name> <name><surname>Yap</surname> <given-names>K. S.</given-names></name> <name><surname>Yap</surname> <given-names>H. J.</given-names></name></person-group> (<year>2016</year>). <article-title>A Constrained optimization based extreme learning machine for noisy data regression</article-title>. <source>Neurocomputing</source> <volume>171</volume>, <fpage>1431</fpage>&#x02013;<lpage>1443</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2015.07.065</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wong</surname> <given-names>S. Y.</given-names></name> <name><surname>Yap</surname> <given-names>K. S.</given-names></name> <name><surname>Yap</surname> <given-names>H. J.</given-names></name> <name><surname>Tan</surname> <given-names>S. C.</given-names></name></person-group> (<year>2015a</year>). <article-title>&#x0201C;A truly online learning algorithm using hybrid fuzzy ARTMAP and online extreme learning machine for pattern classification</article-title>. <source>Neural Process. Lett</source>. <volume>42</volume>, <fpage>585</fpage>&#x02013;<lpage>602</lpage>. <pub-id pub-id-type="doi">10.1007/s11063-014-9374-5</pub-id></citation></ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wong</surname> <given-names>S. Y.</given-names></name> <name><surname>Yap</surname> <given-names>K. S.</given-names></name> <name><surname>Yap</surname> <given-names>H. J.</given-names></name> <name><surname>Tan</surname> <given-names>S. C.</given-names></name> <name><surname>Chang</surname> <given-names>S. W.</given-names></name></person-group> (<year>2015b</year>). <article-title>On Equivalence of FIS and ELM for interpretable rule-based knowledge representation</article-title>. <source>IEEE Trans. Neural Netw. Learn. Syst.</source> <volume>27</volume>, <fpage>1417</fpage>&#x02013;<lpage>1430</lpage>. <pub-id pub-id-type="doi">10.1109/TNNLS.2014.2341655</pub-id><pub-id pub-id-type="pmid">25134093</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yadav</surname> <given-names>S. K.</given-names></name> <name><surname>Sinha</surname> <given-names>R.</given-names></name> <name><surname>Bora</surname> <given-names>P. K.</given-names></name></person-group> (<year>2015</year>). <article-title>Electrocardiogram signal denoising using non-local wavelet transform domain filtering</article-title>. <source>IET Signal Proc.</source> <volume>9</volume>, <fpage>88</fpage>&#x02013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.1049/iet-spr.2014.0005</pub-id></citation></ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yildirim</surname> <given-names>&#x000D6;.</given-names></name> <name><surname>P&#x00142;awiak</surname> <given-names>P.</given-names></name> <name><surname>Tan</surname> <given-names>R.-S.</given-names></name> <name><surname>Acharya</surname> <given-names>U. R.</given-names></name></person-group> (<year>2018</year>). <article-title>Arrhythmia detection using deep convolutional neural network with long duration ECG signals</article-title>. <source>Comp. Biol. Med.</source> <volume>102</volume>, <fpage>411</fpage>&#x02013;<lpage>420</lpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2018.09.009</pub-id><pub-id pub-id-type="pmid">30245122</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yin</surname> <given-names>W.</given-names></name> <name><surname>Yang</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Oki</surname> <given-names>E.</given-names></name></person-group> (<year>2016</year>). <article-title>ECG monitoring system integrated with IR-UWB radar based on CNN</article-title>. <source>IEEE Access</source> <volume>4</volume>, <fpage>6344</fpage>&#x02013;<lpage>6351</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2016.2608777</pub-id></citation></ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yu</surname> <given-names>S.-N.</given-names></name> <name><surname>Chen</surname> <given-names>Y.-H.</given-names></name></person-group> (<year>2007</year>). <article-title>Electrocardiogram beat classification based on wavelet transformation and probabilistic neural network</article-title>. <source>Patt. Recogn. Lett.</source> <volume>28</volume>, <fpage>1142</fpage>&#x02013;<lpage>1150</lpage>. <pub-id pub-id-type="doi">10.1016/j.patrec.2007.01.017</pub-id><pub-id pub-id-type="pmid">17946461</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zubair</surname> <given-names>M.</given-names></name> <name><surname>Kim</surname> <given-names>J.</given-names></name> <name><surname>Yoon</surname> <given-names>C.</given-names></name></person-group> (<year>2016</year>). <article-title>An automated ECG beat classification system using convolutional neural networks</article-title>, in <source>2016 6th International Conference on IT Convergence and Security (ICITCS)</source> (<publisher-loc>Prague</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1109/ICITCS.2016.7740310</pub-id></citation></ref>
</ref-list> 
</back>
</article> 
