Computers in Biology and Medicine: Yong Xia, Naren Wulan, Kuanquan Wang, Henggui Zhang
Computers in Biology and Medicine: Yong Xia, Naren Wulan, Kuanquan Wang, Henggui Zhang
Computers in Biology and Medicine: Yong Xia, Naren Wulan, Kuanquan Wang, Henggui Zhang
A R T I C L E I N F O A B S T R A C T
Keywords: Background: Atrial fibrillation (AF) is the most common cardiac arrhythmia. The incidence of AF increases with
Atrial fibrillation age, causing high risks of stroke and increased morbidity and mortality. Efficient and accurate diagnosis of AF
Short-term Fourier transform based on the ECG is valuable in clinical settings and remains challenging. In this paper, we proposed a novel
Stationary wavelet transform method with high reliability and accuracy for AF detection via deep learning.
Deep convolutional neural networks Method: The short-term Fourier transform (STFT) and stationary wavelet transform (SWT) were used to analyze
ECG segments to obtain two-dimensional (2-D) matrix input suitable for deep convolutional neural networks.
Then, two different deep convolutional neural network models corresponding to STFT output and SWT output
were developed. Our new method did not require detection of P or R peaks, nor feature designs for classification,
in contrast to existing algorithms. Finally, the performances of the two models were evaluated and compared with
those of existing algorithms.
Results: Our proposed method demonstrated favorable performances on ECG segments as short as 5 s. The deep
convolutional neural network using input generated by STFT, presented a sensitivity of 98.34%, specificity of
98.24% and accuracy of 98.29%. For the deep convolutional neural network using input generated by SWT, a
sensitivity of 98.79%, specificity of 97.87% and accuracy of 98.63% was achieved.
Conclusion: The proposed method using deep convolutional neural networks shows high sensitivity, specificity and
accuracy, and, therefore, is a valuable tool for AF detection.
* Corresponding author.
E-mail address: xiayong@hit.edu.cn (Y. Xia).
1
Joint first authors.
https://doi.org/10.1016/j.compbiomed.2017.12.007
Received 7 July 2017; Received in revised form 13 December 2017; Accepted 13 December 2017
85
Y. Xia et al. Computers in Biology and Medicine 93 (2018) 84–92
variation of the signal within the window function. Fig. 2 presents a tree-structure diagram of stationary wavelet trans-
form. In Fig. 2, XðzÞ is the zero-padded signal in z-dimension. GðzÞ and
2.3.2. Wavelet transform HðzÞ are high-pass and low-pass filters. Note that in each SWT decom-
Although STFT is a good way to analyze signals in the time-frequency position level, the impulse responses of the high pass and low pass filters
domain, its window is unchanged. An alternative approach to over- are up-sampled by a factor of two. The detail coefficients Dj ðnÞ and coarse
coming the problem in STFT is wavelet transform (WT). The advantage of coefficients Cj ðnÞ in the time domain can be recursively computed as:
the wavelet transform is multiresolution analysis, which means a window X
automatically narrows when focusing on high frequencies and auto- Dj ðnÞ ¼ gðmÞDj1 n 2j1 m (4)
matically widens when focusing on low frequencies. Using wavelet m
1 tb 2.3.3. Comparison between input representations
Tða; bÞ ¼ ∫ f ðtÞψ dt (3)
aR a At the end of the pre-processing step, we employed STFT and SWT,
respectively, on each 5-s signal segment (mentioned in Section 2.2) to
where a is the scale factor, b is the translation. The analyzing wavelet is ψ .
For small a, the wavelet has a good localization because it is contracted in
the time domain; so wavelet transform gives detailed information (or high
frequency) of the signal. For large a, the wavelet is expanded and the
wavelet transform gives a global view of the signal.
Traditionally, wavelet transform has been used for de-noising,
delineation, and compression of the signal. The continuous wavelet
transform (CWT) and discrete wavelet transform (DWT) are not time-
invariant, which means that temporal shifts in the signal will generate
different sets of wavelet transform coefficients [34]. To the contrary, the
stationary wavelet transform (SWT) is time-invariant at each decompo-
sition level. Since the SWT with J-levels on a signal requires the signal
length to be a multiple of 2J, the filtered data segment needs to be
zero-padded.
Fig. 2. Tree-structure of 3-level stationary wavelet transform.
86
Y. Xia et al. Computers in Biology and Medicine 93 (2018) 84–92
87
Y. Xia et al. Computers in Biology and Medicine 93 (2018) 84–92
Since the sampling frequency is 250 Hz, this signal contains 1250 sam-
pling points. From Fig. 5, it can be observed that the atrial activity can be
perceived in the wavelet domain, especially in the detail coefficients.
Fig. 6(A) shows the feature maps extracted by the first convolutional Fig. 7. Feature maps and weight map of DeepNet2. A: Feature map; and B: Weight map.
layer of DeepNet1 with 16 convolutional kernels of size 5 5. Fig. 6(B)
shows the weight maps (called kernel matrices) of the first convolutional RGB image; so 32 convolutional kernels can only generate 32 not 96 wt
layer. A RGB spectrogram has three feature dimensions of red, blue, and maps as shown in Fig. 7(B).
green; each dimension is executed by 16 kernels to perform a convolution It is not hard to find out that traditional feature descriptors usually
operation so that a total of 48 wt maps are generated. It is worth noting rely on the absence of P-waves or R-R irregularities, or hand-crafted
that for RGB three channels, the actual size of a convolutional kernel is features. For example, many R-R interval-based methods compare the
K K 3 (K means kernel size), so a convolutional kernel has three density histogram of R-R intervals. For traditional features, P and/or R
K K weight maps. In DeepNet1, its first convolutional layer has 16 peaks need to be located; or researchers need to have very high expertise
kernels with kernel size of 5 5, so its first convolutional layer has 48 to manually select features for AF detection. As a result, their perfor-
5 5 wt maps. The corresponding 5 5 kernel on each channel is used to mances depend on the accuracy of peak detection or the reliability of the
convolute the image on each channel, and then the output of the three hand-crafted features. When missing peaks or erroneously detected
channels are added to get 16 feature maps instead of 48 feature maps. peaks, or hand-crafted features is not appropriate, their performance will
Fig. 7(A) shows the feature maps extracted by the first convolutional be compromised. In contrast, deep features, shown in Figs. 6 and 7, make
layer of DeepNet2 with 32 convolutional kernels of size 3 11. Different a comprehensive description of a signal segment from different angles
from the feature maps and weight maps in DeepNet1, those in DeepNet2 and multiple layers. This demonstrates that deep learning methods can
are more complex and present more detail of a 5-s signal segment. This obtain more detailed feature information from the signal. Therefore,
also demonstrates, on another perspective, that the input type based on deep features can produce more robust and stable performance than
SWT has a better explanation of the original signal in the time domain traditional methods. Moreover, deep features have diversity with
and frequency domain than the input form based on STFT. Another point different levels. From low layer to high layer, the features change from
to note is that the input type based on SWT is just a 2-D matrix of coef- concrete and delicate to abstract and visual, which is a process from local
ficient time series. It can be regarded as a grayscale “image” but not a to global.
Table 1
The comparisons of different DCNN architectures which use SWT coefficient time series as
input.
1 1 1 92.18%
2 1 2 95.21%
3 2 2 –
Fig. 5. An ECG signal prior to and during AF and its detail wavelet coefficients from levels
4 2 (without ReLU) 2 98.63%
4 to 6. (A) Original signal; (B) detail coefficient at level 4; (C) detail coefficient at level 5;
5 3 2 –
and (D) detail coefficient at level 6.
88
Y. Xia et al. Computers in Biology and Medicine 93 (2018) 84–92
record the test accuracy since the DCNN did not work in these three
3.5. The architecture of DeepNet2
experiments. The training error did not decrease substantially; that is, the
parameters in the network are not valid for classification. Specially, the
Table 3 shows the architecture of DeepNet2. Since the architecture of
second column (the number of convolutional layers) in Tables 1 and 2
DeepNet2 is similar to the architecture of DeepNet1 as shown in Fig. 8,
represents a combination layer of a convolutional layer, a ReLU layer and
the architecture of DeepNet2 is illustrated using a table. The DeepNet2 is
a pooling layer, not just a convolutional layer. For example, the second
composed of two convolutional layers, two max-pooling layers, two ReLU
column of Experiment 1 in Table 1 is one, which means a convolutional
layers, one dropout layer, two fully-connected layers and one softmax
layer, a ReLU layer and a pooling layer. The second column of Experi-
layer. The first convolutional layer with kernel size of 3 11 has 32
ment 4 in Table 2 is three, which means three convolutional layers, three
hidden units followed by a ReLU layer. The pooling size of the first
ReLU layers and three pooling layers. Although the second column re-
pooling layer is 2 3. The second convolutional layer with a kernel size
sults of Experiment 3 and Experiment 4 in Table 1 are both two, Exper-
of 2 11 has 32 hidden units. The pooling size of the second pooling
iment 4 means that there is no ReLU layer between the second
layer is 2 3. The first fully-connected layer has 100 hidden units fol-
convolutional layer and the second pooling layer.
lowed by a ReLU layer and then a dropout layer. The second fully-
As a result, the best DCNN architecture using SWT coefficient time
connected layer has two hidden units. Lastly, there is a softmax layer
series as input is two combination layers (convolutional layer, ReLU layer
with two outputs. The initial learning rate is set to 0.01 and it is reduced
and pooling layer), where there is no ReLU layer between the second
by a factor of 0.1 every 5000 iterations. Momentum and weight decay
convolutional layer and the second pooling layer, and two fully-
rates are set to 0.9 and 0.0005, respectively.
connected layers. The best DCNN architecture using a spectrogram as
input is three combination layers and two fully-connected layers. It can
be observed in Tables 1 and 2 that, in the current best architecture, 3.6. Comparison of performance between DeepNet1 and DeepNet2
increasing the number of combination layers or fully-connected layers
will lead to poor test accuracy. Fig. 9 shows the detection results of the proposed method DeepNet1
on ECG signal segments where 1 means AF, and 0 means non-AF. Fig. 10
3.4. The architecture of DeepNet1 shows the detection results of the proposed method DeepNet2 on the
same ECG signal segments as in Fig. 9. The left side of Figs. 9 and 10
Through Section 3.3 and other experiments, we get the detailed
parameter settings for DeepNet1 and DeepNet2. Fig. 8 shows the archi- Table 3
The detailed architecture of DeepNet2.
tecture of DeepNet1. The DeepNet1 is composed of three convolutional
layers, three max-pooling layers, two ReLU layers, one drop-out layer, Layer # of output Kernel size/Pooling size
two fully-connected layers and one softmax layer. The first convolutional Conv1 32 3 11
layer with a kernel size of 5 5 has 16 hidden units followed by a ReLU ReLU1 32
layer. The pooling size of the first pooling layer is 3 3. The settings of Pool1 32 23
Conv2 32 2 11
the following convolutional, ReLU and pooling layers are the same as
Pool2 32 23
those for the first layer. The first fully-connected layer has 50 hidden Fc1 100
units followed by a ReLU layer and then a dropout layer. The second ReLU2 100
fully-connected layer has two hidden units. Lastly, there is a softmax Dropout1
layer with two outputs. The learning rate is fixed to 0.0008. Momentum Fc2 2
Softmax 2
and weight decay rates are set to 0.9 and 0.0005 according to [40],
Fig. 8. The DCNN architecture for the detection of atrial fibrillation using the spectrogram as input.
89
Y. Xia et al. Computers in Biology and Medicine 93 (2018) 84–92
Fig. 9. Detection results using the proposed method DeepNet1. (A) Original signal; (B) true annotation; (C) score (output of DCNNs classifier); (D) detection result; (E) Original signal; (F)
true annotation; (G) score (output of DCNNs classifier); (H) detection result.
Fig. 10. Detection results using the proposed method DeepNet2. (A) Original signal; (B) true annotation; (C) score (output of DCNNs classifier); (D) detection result; (E) Original signal; (F)
true annotation; (G) score (output of DCNNs classifier); (H) detection result.
shows the detection result of the same signal segment of record 04043
containing an AF episode. From Figs. 9(D) and Fig. 10(D), we can observe
that there is an erroneous detection around 35 s–50 s (Fig. 9(D)), but not
in (Fig. 10(D)). The boundary of the AF segment, however, is accurately
positioned in Fig. 9 (D). Whereas it is advanced by approximately 5 s in
Fig. 10(D). The right side of Figs. 9 and 10 shows the detection results of
the same signal segment of record 04936 of a non-AF episode. From
Figs. 9(H) and 10(H), we can observe that the performance is satisfactory
in detecting the non-AF segment. The detection score is slightly higher in
Fig. 10 (G) during about 10 s–25 s than in Fig. 9(G).
Based on a ten-fold cross-validation evaluation, the DeepNet1 pre-
sented a sensitivity of 98.34 0.13%, specificity of 98.24 0.11% and
accuracy of 98.29 0.17% on average. For DeepNet2, a sensitivity of Fig. 11. Test accuracy of the proposed method DeepNet1 and DeepNet2 with respect to
98.79 0.12%, specificity of 97.87 0.15% and accuracy of iterations.
98.63 0.17% on average was achieved. Fig. 11 shows test accuracy of
the proposed method DeepNet1 and DeepNet2 along the iteration. It can 3.7. Training time
be observed that the accuracy of their classification is basically the same
and very close to one. The experiments were performed on a computer with 2 CPUs at
2.1 GHz, 2 NVIDIA Tesla K40c GPU and 32-Gb memory. We ran all the
90
Y. Xia et al. Computers in Biology and Medicine 93 (2018) 84–92
proposed Deep models over our highly efficient GPU using the Caffe deep Table 4
learning framework [39]. The total training time of DeepNet1 was about Comparison of the performances of AF detection algorithms that have been validated on
MIT-BIH AFIB database.
40 min for 30,000 iterations; so the average time for 1 BP iteration was
about 83 msec. The total training time of DeepNet2 was about 8 min for Algorithm Sensitivity Specificity Accuracy Methodology
(%) (%) (%)
30,000 iterations; so the average time for 1 BP iteration was about
16 msec. Slocum et al. 62.80 77.46 – P-wave absence
[15]
Tateno et al. 94.40 97.20 – RR interval
4. Discussion [18] irregularity
Dash et al. 94.40 95.10 – RR interval
We performed a ten-fold cross-validation on the data obtained by [19] irregularity
SWT and STFT to train and test the deep convolutional neural networks. Huang et al. 96.10 98.10 – RR interval
[20] irregularity
Classification performance was measured using accuracy (Acc), sensi-
Lee et al. [22] 98.20 97.70 – RR interval
tivity (Se) and specificity (Sp). Table 4 summarizes the performance of irregularity
our methods and the existing algorithms validated on the MIT-BIH AF Babaeizadeh 92.00 95.50 – RR interval
database. It is apparent that the specificity of our proposed method et al. [24] irregularity þ P-
DeepNet1, with 98.24%, is the highest of all algorithms listed in Table 4. wave absence
Jiang et al. 98.20 97.50 – RR interval
The specificity of our proposed method DeepNet2 is just 0.13% lower [25] irregularity þ P-
than that of Huang [20]. The sensitivity of our proposed method Deep- wave absence
Net1 and DeepNet2 are 98.34% and 98.79%, respectively, which are Asgari et al. 97.00 97.10 – Peak-to-
both higher than any other algorithms. The accuracy of our proposed [26] averagepower
ration þ log-energy
method DeepNet1 and DeepNet2 is 98.29%, 98.63%, respectively.
entropy
For STFT, some researchers used a grayscale spectrogram instead of a DeepNet1 98.34 98.24 98.29 STFT þ DCNNs
RGB spectrogram. Here, we also evaluate DeepNet1 by using a grayscale (RGB)
image of a spectrogram, and the result is also provided in Table 4 for DeepNet1 98.60 97.17 97.74 STFT þ DCNNs
comparison. It can be observed that the performance on sensitivity with (grayscale)
DeepNet2 98.79 97.87 98.63 SWT þ DCNNs
grayscale spectrogram is a little better than that of a RGB spectrogram.
However, in general, a RGB spectrogram outweighs a grayscale spec- Bold refers proposed methods in this paper.
trogram for AF detection. As mentioned in Section 3.2 (regarding the
feature maps and weight maps of STFT þ DCNN), a RGB spectrogram has
three feature dimensions of red, blue, green, and if the first convolutional DCNNs will extract the intrinsic characteristics of the signal and use
layer has 16 kernels, then 16 feature maps and 48 wt maps (called weight these learned features to classify.
matrices) will be produced. However, a grayscale spectrogram has only ● Furthermore, some existing methods such as the Jiang [25] method,
one channel, if the first convolutional layer has 16 kernels, then only 16 need a long data segment (50 beats, about 40 s) to ensure a high ac-
feature maps and 16 wt maps will be produced. It can be seen that the curacy on AF detection which will cause some short AF episodes to be
weight matrix from RGB spectrogram is three times larger than the undetectable. But our proposed methods can achieve a good accuracy
weight matrix from a grayscale spectrogram. Furthermore, as the number on an ECG of 5 s.
of layers increase, the difference in the weight matrix will increase
greatly. As for the convolutional neural network, the weight matrix can 5. Conclusion
be considered to describe and analyze the data from different aspects or
angles. So, in a way, the increase in the weight matrix is beneficial for the In this paper, we convert the one-dimensional ECG signal into two-
network to capture the intrinsic characteristics of data and, accordingly, dimensional form by short-time Fourier transform and stationary
can improve the accuracy of classification. wavelet transform respectively, to perform the detection of atrial fibril-
Overall, our proposed methods, DeepNet1 and DeepNet2, have a lation using a deep convolution neural network. The proposed method
better performance than Slocum [15], Tateno [18], Dash [19], Huang does not require the detection of P or R peak and the extraction of
[20], Lee [21], Babaeizadeh [24], Jiang [25] and Asgari [26] algorithms. handcrafted manual feature. The results of the experiments, which are
Specifically, the methods we propose have the following advantages for performed over the MIT-BIH Atrial fibrillation (MIT-BIH AFIB) database,
detecting AF in clinical applications: show that our DCNNs with the input form based on STFT has sensitivity
of 98.34%, specificity of 98.24%, accuracy of 98.29%, and DCNNs with
● Tateno [18], Dash [19], Huang [20] and Lee [22] methods need to the input form based on SWT has sensitivity of 98.79%, specificity of
detect R-R intervals. Slocum [15] method needs to detect P-wave. 97.87%, accuracy of 98.63%, both outperform the majority of the
Babaeizadeh [24] and Jiang [25] methods need to detect both R-R existing algorithms. Furthermore, the proposed methods achieve satis-
intervals and P-wave. Their performances depend on the accuracy of factory performance for data segment as short as 5 s. Therefore, the
peak detection and when missing peaks or erroneously detected proposed approach is a fast, accurate, efficient method for atrial fibril-
peaks, their performance may be poor. In contrast, our proposed lation detection.
methods do not rely on peak detection.
● Although the Asgari [26] method does not need to detect R-R in- Conflicts of interest
tervals or P-wave, they manually selected the peak-to-average power
ratio and log-energy entropy of the signal as features of AF detection. None.
However, the features used in Asgari [26] method may not represent
the optimal characteristics of the signal. Moreover, for the AF Acknowledgment
detection algorithms using hand-crafted features, the detection ac-
curacy is largely dependent on the quality of the feature selection. In This work was financially supported by Shandong Province Natural
contrast, our proposed methods, DeepNet1 and DeepNet2, do not Science Foundation (ZR2015FM028) and Shandong Province Science
have that limitation. We simply use the original signal and transform and Technology Development Plan (2014GSF118152).
it to a 2-D representation with STFT and SWT respectively. Then the
91
Y. Xia et al. Computers in Biology and Medicine 93 (2018) 84–92
References [19] S. Dash, K.H. Chon, S. Lu, E.A. Raeder, Automatic real time detection of atrial
fibrillation, Ann. Biomed. Eng. (2009) 1701–1709.
[20] C. Huang, S. Ye, H. Chen, D. Li, F. He, Y. Tu, A novel method for detection of the
[1] V. Fuster, L.E. Ryden, D.S. Cannom, H.J. Crijns, A.B. Curtis, K.A. Ellenbogen,
transition between atrial fibrillation and sinus rhythm, IEEE Trans. Biomed. Eng.
J.L. Halperin, G.N. Kay, J.-Y. Le Huezey, J.E. Lowe, S.B. Olsson, E.N. Prystowsky,
(2011) 1113–1119.
J.L. Tamargo, L.S. Wann, 2011 ACCF/AHA/HRS focused updates incorporated into
[21] D.E. Lake, J.R. Moorman, Accurate estimation of entropy in very short physiological
the ACC/AHA/ESC 2006 guidelines for the management of patients with atrial
time series: the problem of atrial fibrillation detection in implanted ventricular
fibrillation, J. Am. Coll. Cardiol. (2011) 101–198.
devices, AJP Hear. Circ. Physiol (2011) 319–325.
[2] M. Lainscak, N. Dagres, G.S. Filippatos, S.D. Anker, D.T. Kremastinos, Atrial
[22] J. Lee, B.A. Reyes, D.D. McManus, O. Mathias, K.H. Chon, Atrial fibrillation
fibrillation in chronic non-cardiac disease: where do we stand? Int. J. Cardiol.
detection using an iphone 4S, IEEE Trans. Biomed. Eng. (2013) 203–206.
(2008) 311–315.
[23] X. Zhou, H. Ding, B. Ung, E. Pickwell-MacPherson, Y. Zhang, Automatic online
[3] A. Camm, G. Lip, R. De Caterina, I. Savelieva, D. Atar, S.H. Hohnloser, 2012 focused
detection of atrial fibrillation based on symbolic dynamics and Shannon entropy,
update of the ESC Guidelines for the management of atrial fibrillation, Eur. Heart J.
Biomed. Eng. Online (2014) 18.
(2012) 2719–2747.
[24] S. Babaeizadeh, R.E. Gregg, E.D. Helfenbein, J.M. Lindauer, S.H. Zhou,
[4] S. Colilla, A. Crow, W. Petkun, D.E. Singer, T. Simon, X. Liu, Estimates of current
Improvements in atrial fibrillation detection for real-time monitoring,
and future incidence and prevalence of atrial fibrillation in the U.S. adult
J. Electrocardiol. (2009) 522–526.
population, Am. J. Cardiol. (2013) 1142–1147.
[25] K. Jiang, C. Huang, S. Ye, H. Chen, High accuracy in automatic detection of atrial
[5] B.P. Krijthe, A. Kunst, E.J. Benjamin, G.Y.H. Lip, O.H. Franco, A. Hofman,
fibrillation for Holter monitoring, J. Zhejiang Univ. - Sci. B. (2012) 751–756.
J.C.M. Witteman, B.H. Stricker, J. Heeringa, Projections on the number of
[26] S. Asgari, A. Mehrnia, M. Moussavi, Automatic detection of atrial fibrillation using
individuals with atrial fibrillation in the European Union, from 2000 to 2060, Eur.
stationary wavelet transform and support vector machine, Comput. Biol. Med.
Heart J. (2013) 2746–2751.
(2015) 132–142.
[6] A.M. Gillis, A.D. Krahn, A.C. Skanes, S. Nattel, Management of atrial fibrillation in
[27] S.C. Lee, Using a translation-invariant neural network to diagnose heart arrhythmia,
the year 2033: new concepts, tools, and applications leading to personalized
in: images Twenty-First Century, Proc. Annu. Int. Eng. Med. Biol. Soc (1996)
medicine, Can. J. Cardiol. (2013) 1141–1146.
2025–2026.
[7] C.D. Furberg, B.M. Psaty, T.A. Manolio, J.M. Gardin, V.E. Smith, P.M. Rautaharju,
[28] P. De Chazal, R.B. Reilly, A patient-adapting heartbeat classifier using ECG
Prevalence of atrial fibrillation in elderly subjects (the Cardiovascular Health
morphology and heartbeat interval features, IEEE Trans. Biomed. Eng. (2006)
Study), Am. J. Cardiol. (1994) 236–241.
2535–2543.
[8] P. a Wolf, R.D. Abbott, W.B. Kannel, Atrial fibrillation as an independent risk factor
[29] H. Goh, N. Thome, M. Cord, J.H. Lim, Learning deep hierarchical visual feature
for stroke: the Framingham Study, Stroke (1991) 983–988.
coding, IEEE Trans. Neural Networks Learn. Syst (2014) 2212–2225.
[9] C. a Sanoski, Clinical, economic, and quality of life impact of atrial fibrillation,
[30] Y. LeCun, Y. Bengio, G. Hinton, Deep learning, Deep Learn (2015) 436–444.
J. Manag. Care Pharm. (2009) 4–9.
[31] P. Xiong, H. Wang, M. Liu, S. Zhou, Z. Hou, X. Liu, ECG signal enhancement based
[10] V. Markides, R.J. Schilling, Atrial fibrillation: classification, pathophysiology,
on improved denoising auto-encoder, Eng. Appl. Artif. Intell. (2016) 194–202.
mechanisms and drug treatment, Heart (2003) 939–943.
[32] L. Cohen, Time-frequency analysis: theory and applications, J. Acoust. Soc. Am. 134
[11] C.T. January, L.S. Wann, J.S. Alpert, H. Calkins, J.E. Cigarroa, J.C.J. Cleveland,
(5) (1995), 4002–4002.
J.B. Conti, P.T. Ellinor, M.D. Ezekowitz, M.E. Field, K.T. Murray, R.L. Sacco,
[33] P.R. Gomes, F.O. Soares, J.H. Correia, C.S. Lima, ECG data-acquisition and
W.G. Stevenson, P.J. Tchou, C.M. Tracy, C.W. Yancy, 2014 AHA/ACC/HRS
classification system by using Wavelet-domain Hidden Markov Models, in: 2010
guideline for the management of patients with atrial fibrillation: executive
Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. EMBC’10, 2010, pp. 4670–4673.
summary, J. Am. Coll. Cardiol. (2014) 2246–2280.
[34] A. Salmanpour, L.J. Brown, J.K. Shoemaker, Performance analysis of stationary and
[12] Q. Xiong, M. Proietti, K. Senoo, G.Y.H. Lip, Asymptomatic versus symptomatic atrial
discrete wavelet transform for action potential detection from sympathetic nerve
fibrillation: a systematic review of age/gender differences and cardiovascular
recordings in humans, in: 2008 30th Annu. Int. Conf. IEEE Eng. Med. Biol. Soc,
outcomes, Int. J. Cardiol. (2015) 172–177.
2008, pp. 2932–2935.
[13] S.A. Strickberger, J. Ip, S. Saksena, K. Curry, T.D. Bahnson, P.D. Ziegler,
[35] I.J. Goodfellow, D. Wardefarley, M. Mirza, A. Courville, Y. Bengio, Maxout
Relationship between atrial tachyarrhythmias and symptoms, Heart Rhythm (2005)
networks, Comput. Sci (2013) 1319–1327.
125–131.
[36] A.L. Goldberger, L.A.N. Amaral, L. Glass, J.M. Hausdorff, P.C. Ivanov, R.G. Mark,
[14] C.W. Israel, G. Gr€onefeld, J.R. Ehrlich, Y.G. Li, S.H. Hohnloser, Long-term risk of
J.E. Mietus, G.B. Moody, C.-K. Peng, H.E. Stanley, PhysioBank, PhysioToolkit, and
recurrent atrial fibrillation as documented by an implantable monitoring device:
PhysioNet, Circulation (2000) 215–220.
implications for optimal patient care, J. Am. Coll. Cardiol. (2004) 47–52.
[37] B. Weng, J.J. Wang, F. Michaud, M. Blanco-Velasco, Atrial fibrillation detection
[15] J. Slocum, A. Sahakian, S. Swiryn, Diagnosis of atrial fibrillation from surface
using stationary wavelet transform analysis, Conf Proc IEEE Eng Med Biol Soc.
electrocardiograms based on computer-detected atrial activity, J. Electrocardiol.
(2008) 1128–1131.
(1992) 1–8.
[38] M. Stridh, L. S€ornmo, C.J. Meurling, S.B. Olsson, Sequential characterization of
[16] S. Ladavich, B. Ghoraani, Rate-independent detection of atrial fibrillation by
atrial tachyarrhythmias based on ECG time-frequency analysis, IEEE Trans. Biomed.
statistical modeling of atrial activity, Biomed. Signal Process Contr. (2015)
Eng. (2004) 100–114.
274–281.
[39] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama,
[17] N. Larburu, T. Lopetegi, I. Romero, Comparative study of algorithms for Atrial
T. Darrell, Caffe: convolutional architecture for fast feature embedding, ACM Int.
Fibrillation detection, 2011 Comput. Cardiol (2011) 265–268.
Conf. Multimed (2014) 675–678.
[18] K. Tateno, L. Glass, Automatic detection of atrial fibrillation using the coefficient of
[40] A. KrizhKrizhevsky, I. Sutskever, G.E. Hinton, ImageNet classification with deep
variation and density histograms of RR and ΔRR intervals, Med. Biol. Eng. Comput.
convolutional neural networks, Adv. Neural Inf. Process. Syst. (2012) 1–9.
(2001) 664–671.
92