ArticlePDF Available

Convolutive Blind Source Separation Algorithms Applied to the Electrocardiogram of Atrial Fibrillation: Study of Performance

Authors:

Abstract and Figures

The analysis of the surface electrocardiogram (ECG) is the most extended noninvasive technique in medical diagnosis of atrial fibrillation (AF). In order to use the ECG as a tool for the analysis of AF, we need to separate the atrial activity (AA) from other cardioelectric signals. In this matter, statistical signal processing techniques, like blind source separation (BSS), are able to perform a multilead statistical analysis with the aim to obtain the AA. Linear BSS techniques can be divided in two groups depending on the mixing model: algorithms where instantaneous mixing of sources is assumed, and convolutive BSS (CBSS) algorithms. In this work, a comparison of performance between one relevant CBSS algorithm, namely Infomax, and one of the most effective independent component analysis (ICA) algorithms, namely FastICA, is developed. To carry out the study, pseudoreal AF ECGs have been synthesized by adding fibrillation activity to normal sinus rhythm. The algorithm performances are expressed by two indexes: the signal to interference ratio (SIRAA) and the cross-correlation (RAA) between the original and the estimated AA. Results empirically prove that the instantaneous mixing model is the one that obtains the best results in the AA extraction, given that the mean SIRAA obtained by the FastICA algorithm (37.6 +/- 17.0 dB) is higher than the main SIRAA obtained by Infomax (28.5 +/- 14.2 dB). Also the RAA obtained by FastICA (0.92 +/- 0.13) is higher than the one obtained by Infomax (0.78 +/- 0.16).
Content may be subject to copyright.
1530 IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 54, NO. 8, AUGUST 2007
Communications
Convolutive Blind Source Separation Algorithms Applied
to the Electrocardiogram of Atrial Fibrillation: Study of
Performance
Carlos Vayá*, José J. Rieta, César Sánchez, and David Moratal
Abstract—The analysis of the surface electrocardiogram (ECG) is the
most extended noninvasive technique in medical diagnosis of atrial fibril-
lation (AF). In order to use the ECG as a tool for the analysis of AF, we
need to separate the atrial activity (AA) from other cardioelectric signals.
In this matter, statistical signal processing techniques, like blind source sep-
aration (BSS), are able to perform a multilead statistical analysis with the
aim to obtain the AA. Linear BSS techniques can be divided in two groups
depending on the mixing model: algorithms where instantaneous mixing of
sources is assumed, and convolutive BSS (CBSS) algorithms. In this work, a
comparison of performance between one relevant CBSS algorithm, namely
Infomax, and one of the most effective independent component analysis
(ICA) algorithms, namely FastICA, is developed. To carry out the study,
pseudoreal AF ECGs have been synthesized by adding fibrillation activity
to normal sinus rhythm. The algorithm performances are expressed by two
indexes: the signal to interference ratio (
) and the cross-correla-
tion (
) between the original and the estimated AA. Results empirically
prove that the instantaneous mixing model is the one that obtains the best
results in the AA extraction, given that the mean
obtained by the
FastICA algorithm (
) is higher than the main ob-
tained by Infomax (
). Also the obtained by FastICA
(
) is higher than the one obtained by Infomax ( ).
Index Terms—Atrial fibrillation, convolutive blind source separation,
ECG, independent component analysis.
I. INTRODUCTION
The prevalence of atrial fibrillation (AF) is less than 1% in people
under 60 years old, but it increases significantly in those over 70, ap-
proximating to 10% in those over 80 [1]. The exhaustive analysis of AF
episodes requires to separate previously the atrial activity (AA) com-
ponent from other bioelectric signals that contribute to the electrocar-
diogram (ECG) formation [2]. Some of these signals are ventricular
activity (VA), muscular activity, noise and artifacts introduced by elec-
trodes, and the powerline interference [3].
Blind Source Separation (BSS) techniques have been successfully
applied to the extraction of the AA from the ECG of AF episodes by as-
suming instantaneous and linear mixing of the bioelectric signals in the
human body [3]. The feasibility in this matter of the FastICA algorithm,
based on independent components analysis (ICA), is proved in [3] and
[4]. Nonetheless, the propagation delays of the bioelectric signal in the
Manuscript received July 26, 2006; revised October 28, 2006. This work was
supported in part by the the Generalitat Valenciana under Project GV06/299.
Asterisk indicates corresponding author.
*C. Vayá is with the Department of Innovation in Bioengineering, Castilla-la
Mancha University, Escuela Politécnica Superior de Cuenca, Camino del
Pozuelo s/n, 16071 Cuenca, Spain (e-mail: carlos.vaya@uclm.es).
J. J. Rieta and D. Moratal are with Biomedical Synergy, Polytechnic Univer-
sity of Valencia, EPSG, Carretera Nazaret Oliva s/n, 46730 Gandía, Valencia,
Spain.
C. Sánchez is with the Department of Innovation in Bioengineering,
Castilla-la Mancha University, Escuela Politécnica Superior de Cuenca,
Camino del Pozuelo s/n, 16071 Cuenca, Spain (e-mail:carlos.vaya@uclm.es).
Digital Object Identifier 10.1109/TBME.2006.889778
body and specially the movements of the heart could violate the ICA
assumptions of spatial stationarity and instantaneous mixing of phys-
ical sources [3]. On the contrary, in the convolutive model, these cir-
cumstances can be considered by using convolutive BSS (CBSS) algo-
rithms. CBSS algorithms have not been applied yet to the extraction
of the AA. Our objective is to compare the performance of one rele-
vant CBSS algorithm, namely Infomax [5], and the FastICA algorithm
[6]. The main goals of this contribution are, on the one hand, to reveal
whether CBSS algorithms are reliable to extract the AA from the ECG
of AF episodes when the typical ECG sampling rate of 1 kHz is used
and, on the other hand, to study their possible improvement in the AA
estimation with respect to the instantaneous BSS algorithms.
II. B
LIND
SOURCE SEPARATION
EXTRACTION TECHNIQUES
Time-domain-based techniques, like Average Beat Substraction [7],
have been well accepted and used in clinical applications [8] to cancel
out the QRS complex and the T wave. Nonetheless, they are only ap-
plied to the leads where atrial fibrillation is more easily distinguish-
able, e.g., V1. This means that if we apply QRST cancellation tech-
niques to different leads, we would obtain different atrial activities as
well. Consequently, they do not make use of the information included
in every lead in an unified way. On the contrary, BSS techniques make
a multilead statistical analysis by exploiting the spatial diversity that
multiple spatially separated electrodes introduce [3], [9]. When linear
mixing of sources is considered, we talk about linear BSS techniques.
Three main conditions must be fulfilled in order to apply linear BSS
techniques: 1) independence of sources; 2) nongaussianity of sources;
3) linear mixing of sources [10]. Within the context of AF episodes,
these three conditions are applicable. First, in AF episodes the bioelec-
tric sources of the heart generating AA and VA can be considered to
be statistically independent. This is justified because only a portion of
atrial waveforms traverse the atrioventricular node and provoke ven-
tricular depolarization [3], [4]. Furthermore, the physical origin of the
atrial wavefront that has been able to produce ventricular depolariza-
tion could be very variable. This uncoordinated operation of AA and
VA during an AF episode allows us to consider them as statistically
independent [4]. Second, AA has a sub-Gaussian probability distribu-
tion whereas the VA is clearly super-Gaussian [3]. Finally, the mixing
process of the bioelectric sources in the human body can be considered
to be linear [11], [12]. Therefore, linear BSS techniques can be applied
to extract the AA in AF episodes.
These are the most important linear models used in BSS:
1) Instantaneous Linear Mixing Model: Within the assumption of
linearity, the simplest situation is the instantaneous linear model. In
this case, the instantaneous mixing of the cardioelectric sources in the
human body is assumed. This can be expressed as [6]
(1)
where
is a matrix that represents the linear transformation of source
mixing,
is the column vector that contains the observations at the
th sampling time, contains the sources, and the additive
noise in every source. The BSS algorithms based on the instantaneous
linear mixing model try to find the matrix
as the pseudoinverse of
the
that obtains the best estimation of the sources [6]:
(2)
0018-9294/$25.00 © 2007 IEEE
IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 54, NO. 8, AUGUST 2007 1531
The global mixing-unmixing system is characterized by the matrix ,
so that
.
The FastICA algorithm [6], based on the instantaneous mixing
model, has already been applied to the extraction of the AA from
ECGs of AF episodes [3]. In FastICA the instantaneous mixing of the
sources is assumed and quite good results in the extraction of the AA
have been obtained by using this algorithm [2].
2) Convolutive Linear Mixing Model: In the convolutive model, a
more realistic case is considered. Here, weighted and, besides, delayed
contributions in the generation of the observations are taken into ac-
count [6].
The convolutive model is a more general linear mixing BSS model
where coefcients of
are substituted by nite-duration impulse re-
sponse (FIR) lters and the product operator of (1) and (2) by the con-
volution operator
. Indeed, the previous instantaneous model could
be considered as a particular case of the convolutive model where the
length of the lters is one.
The performance of several CBSS algorithms has been tested, but
only the results obtained by Infomax [5] are presented here. Infomax
belongs to an important family of algorithms based on information
theory. Mutual information (MI) is a natural measure of the depen-
dence between random signals [6]. MI is always nonnegative, and zero
if and only if the variables are statistically independent. MI can be used
as a criterion for obtaining the original sources of the BSS problem [6].
The other algorithms that have been tested are the Multi-channel
Blind Least-Mean Square (MBLMS) algorithm [13], the time-delayed
decorrelation (TDD) algorithm [14], and the Convolutive Blind Signal
Separation (CoBliSS) algorithm [15]. Results obtained by TDD are
very similar to results obtained by Infomax, and results obtained by
MBLMS and TDD are quite poorer. Consequently, only results of In-
fomax have been reported in this paper.
III. P
ERFORMANCE ESTIMATION AND
DATABASE
A. Spectral Concentration (SC)
Considering that the typical spectral morphology for AA is charac-
terized by a very pronounced peak in frequencies from 5 to 8 Hz, with
no harmonics and insignicant amplitudes above 15 Hz [16], it is pos-
sible to dene a performance extraction index capable of evaluating
AA extraction quality based on the spectral concentration [9]. The SC
can be dened as
(3)
where
is the power spectral density of the AA signal,
is the fre-
quency,
is the sampling rate (1 kHz), and is the main frequency
peak of the AA. Experimental observations prove that the AA is the
estimated independent component with the highest
when the limits
of the numerator summation are those specied in (3) [8]. Therefore,
the
can be used as an indicator to select, in an automated way, the
estimated source that best matches the AA. A wider distance between
frequency limits introduces high frequency noise components in the
calculation of the
indicator so that the probability of discriminating
correctly the AA is reduced. A shorter distance between limits reduces
the AA power considered in
so that the probability of AA discrim-
ination is reduced too.
B. Performance Indexes
1) Signal-to-Interference Ratio (SIR): Considering
as the one of
the
observations with the highest contribution of AA and as the
one of the
sources that contains the AA, the SIR of
expressed in
decibels is dened as [17]
(4)
where
are the FIR lters of the mixing matrix
, which represent
the contribution of the
th source on the th observation.
stands for
the mathematical expectation. In the same way, considering that
is
the estimated source with the highest contribution of AA (i.e, the esti-
mated source with the highest
), the SIR of
is [17]
(5)
where
are the FIR lters of the global system matrix
. Finally,
the signal-to-interference ratio
is dened as
(6)
2) Cross-Correlation: The cross-correlation between the original
AA and the estimated AA is calculated as
(7)
where
and
are the original and the estimated AA, respectively.
C. ECG Database
The calculation of the previously dened indexes needs the original
sources and the mixing matrix to be known. Given that the sources that
contribute to the generation of real ECGs are unknown, we needed to
establish two environments of synthesized ECGs, so that the perfor-
mance indexes could be applied.
1) First Environment: In the rst environment, 15 pairs of separated
AA and VA of AF episodes are mixed using the convolutive model.
These separated AA and VA have been previously obtained from the
V1 lead of real ECGs by using QRST cancellation [7]. The mixing is
made by random mixing matrices
. Obviously, this is not the real
mixing process of the bioelectric signals in the human body. Nonethe-
less, it can be considered as a valid and objective way of testing and
discarding those algorithms (CoBliSS and MBLMS) that were clearly
useless to extract the AA. The FIR lters length of these mixing ma-
trices (
) has been changed from 1 to 8. All ECGs are 12 s in length,
and they all were obtained at a sampling rate of 1 kHz. In the sepa-
ration process, the lter length of the separation matrix
is an ad-
justable parameter in the CBSS algorithms (
). It has been changed
in the tests from 2 to 32. The value of two is the lowest value allowed
by all the tested CBSS algorithms. The value of 32 has been chosen so
that the controlled maximum delay due to the mixing-unmixing process
equals to 40 ms in duration, accordingly to the aforementioned max-
imum value of
and the sampling rate [13].
2) Second Environment: In the second environment, AF 12-leads
ECG are synthesized by adding AA and VA of every lead. AA and VA
were previously separated from real ECGs of AF episodes by using
QRST cancelation [7]. Therefore, the synthesized ECGs are exactly the
original real recorded signals. In this way we know their AA and VA
components, so that the previously dened parameters can be applied.
The convolutive mixing is assumed to be present in the generation of
the real ECGs and, consequently, in the separated AA and VA. This
synthesis process is depicted in Fig. 1. All resulting ECG recordings
last for 8 s and are sampled at 1 KHz. This second environment com-
prises 20 synthesized ECGs. The maximum value of
has been xed
1532 IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 54, NO. 8, AUGUST 2007
Fig. 1. Second environment synthetic ECG generation. The AF 12-leads ECGs
are synthesized by adding the AA and the VA of every lead, previously separated
from real ECGs of AF episodes.
Fig. 2. and mean values of Infomax and FastICA algorithms in
the rst environment as a function of the FIR lters length of the random mixing
matrices (
).
to 128 by considering 128 ms as a reasonable maximum propagation
delay for all the bioelectric signals in the human body [11]. The res-
piratory induced axis-shifts of all the analyzed ECG recordings have
been previously eliminated by signal ltering. All these recordings are
free from ectopic beats or other nonideal characteristics. This allows
us to develop an initial work focused in the performance of the CBSS
algorithms when regular signals of AF are considered.
Note that the presence of previous convolution in the initial AA and
VA is not an obstacle to the study of performance of CBSS algorithms.
If previous convolution is present, the CBSS algorithms would try to
invert the global linear system composed by the previous convolution
and the additional simulated convolution. On the other hand, given that
the mixing process of sources is unknown, the only way to test the
performance of extraction when no direct access to the atrial activity is
available (we are using surface recordings) is to dene a set of synthetic
signals so that we can compare the signals estimated by BSS algorithms
with the original signals.
IV. R
ESULTS
A. First Environment Results
The global mean
obtained in this environment by FastICA
algorithm (
) is higher than the global mean
obtained by Infomax (
). Also the obtained by
FastICA (
) is higher than this obtained by Infomax (
).
The inuence of
in the rst environment is shown in Fig. 2. Both
performance indexes,
and , decrease when mixing lters
length (
) increases. Mean values of
obtained by Infomax
are several decibels lower than the respective FastICA mean values,
regardless of
. Furthermore, FastICA
mean values are always
nearer to one than those of Infomax.
Fig. 3 illustrates the inuence of the separation lters length (
).
The mean values of FastICA are included only as a reference constant
Fig. 3.
and mean values of Infomax and FastICA algorithms in
the rst environment as a function of the FIR lters length of the separation
matrices (
).
Fig. 4. and mean values of Infomax and FastICA algorithms in
the second environment as a function of the FIR lters length of the separation
matrices (
).
value, given that FastICA does not match the convolutive model and,
therefore, the lters length parameter cannot be chosen. Mean
and
of Infomax are slightly lower than the respective indexes of
FastICA.
decreases when
increases. Thus, the best separation
performance of Infomax is obtained when the instantaneous model is
considered in this environment.
B. Second Environment Results
Also in this environment, the global mean
obtained by Fas-
tICA algorithm (
) is several decibels higher than the one
obtained by Infomax (
). In the same way, the global mean
obtained by FastICA (
) is higher than the one obtained
by Infomax (
).
The inuence of
in the second environment is shown in Fig. 4.
Note that the highest values of
and
, in the Infomax algo-
rithm, are obtained when
equals to eight.
V. C
ONCLUSION
Infomax obtained acceptable results but poorer than those of Fas-
tICA. With regard to the
parameter in the rst environment, Fas-
tICA always exceeds the results obtained by the Infomax algorithm in
the case of instantaneous mixing (
). Hence the instantaneous
linear mixing model can be considered as an acceptable approximation
for the bioelectric mixing in the human body. Given that the instan-
taneous linear mixing model is the particular case of the convolutive
mixing model when
is equal to one, we can conclude that CBSS
algorithms need an improvement to reach at least the performance of
the FastICA algorithm in the case of instantaneous mixing and to ob-
tain better results in the convolutive case.
The Infomax algorithm has been easily adapted to the standard
12-lead real ECG of the second environment, reaching its best perfor-
mance when a convolutive model (
) is considered. An in-depth
analysis of the Infomax algorithm and its adjustment to the special
features of ECG signals could obtain an appropriate algorithm for the
extraction of the AA.
REFERENCES
[1] V. Fuster et al., CC/AHA/ESC guidelines for the management of pa-
tients with atrial brillation,J. Am. Coll. Cardiol., vol. 48, no. 4, pp.
854906, 2006.
IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 54, NO. 8, AUGUST 2007 1533
[2] P. Langley, J. J. Rieta, M. Stridh, J. Millet, L. Sörnmo, and A. Murray,
Comparison of atrial signal extraction algorithms in 12-lead ECGS
with atrial brillation,
IEEE Trans. Biomed. Eng., vol. 53, no. 2, pp.
343346, Feb. 2006.
[3] J. J. Rieta, F. Castells, C. Sánchez, and V. Zarzoso, Atrial activity
extraction for atrial brillation analysis using blind source separation,
IEEE Trans. Biomed. Eng., vol. 51, no. 7, pp. 11761186, Jul. 2004.
[4] J. J. Rieta, F. Castells, C. Sánchez, D. Moratal-Pérez, and J. Millet,
Bioelectric model of atrial brillation: Applicability of blind source
separation techniques for atrial activity estimation in atrial brillation
episodes,in Proc. Int. Conf. Computers in Cardiology, 2003, vol. 30,
pp. 525528.
[5] S. Amari, A. Cichocki, and H. Yang, A new learning algorithm for
blind source separation, Adv. Neural Inf. Process. Syst., no. 8, pp.
757763, 1996.
[6] A. Hyvärinen, J. Karhunen, and E. Oja, Independent Component Anal-
ysis. New York: Wiley, 2001.
[7] J. Slocum, A. Sahakian, and S. Swiryn, Diagnosis of atrial brillation
from surface electrocardiogram based on computer-detected atrial ac-
tivity,J. Electrocardiol., vol. 25, no. 1, pp. 18, 1992.
[8] A. Bollmann, K. Sonne, H. D. Esperer, I. Toepffer, J. J. Langberg, and
H. U. Klein, Non-invasive assessment of brillatory activity in pa-
tients with paroxysmal and persistent atrial brillation using the Holter
ECG,Cardiovasc. Res., vol. 44, pp. 6066, 1999.
[9] F. Castells, J. J. Rieta, J. Millet, and V. Zarzoso, Spatiotemporal blind
source separation approach to atrial activity estimation in atrial tach-
yarrhythmias,IEEE Trans. Biomed. Eng., vol. 52, no. 2, pp. 258267,
Feb. 2005.
[10] J. F. Cardoso, Blind signal separation: Statistical principles, Proc.
IEEE, vol. 9, pp. 20092025, 1981.
[11] R. Plonsey and D. B. Heppner, Considerations of quasi-stationarity in
electrophysiological systems,Bull. Math. Biophys., vol. 29, no. 4, pp.
657664, 1967.
[12] R. M. Gulrajani, The forward and the inverse problem of electrocar-
diography, IEEE Eng. Med. Biol. Mag., vol. 17, no. 5, pp. 84101,
Sep.-Oct. 1998.
[13] R. H. Lambert, Multichanel blind deconvolution: FIR matrix algebra
and separation of multipath mixtures, Ph.D. dissertation, Univ.
Southern California, Los Angeles, 1996.
[14] N. Murata, S. Ikeda, and A. Ziehe, An approach to blind source sepa-
ration based on temporal structure of speech signals, Neurocomputing,
vol. 41, no. 1-4, pp. 124, 2001.
[15] D. W. E. Schobben and P. W. Sommen, A frequency domain blind
signal separation method based on decorrelation,IEEE Trans. Signal
Process., vol. 50, no. 8, pp. 18551865, Aug. 2002.
[16] P. Langley, M. Stridh, J. J. Rieta, L. Sörnmo, J. Millet-Roig, and A.
Murray, Comparison of atrial rhythm extraction techniques for the
estimation of the main atrial frequency from the 12-lead electrocardio-
gram in atrial brillation, in Proc. Int. Conf. Computers in Cardiology,
2002, pp. 2932.
[17] J. M. Sanchis, F. Castells, and J. J. Rieta, Convolutive acoustic mix-
tures approximation to an instantaneous model using a stereo boundary
microphone conguration, in Lecture Notes in Computer Science.
Berlin, Germany: Springer-Verlag, 2004, vol. 3195, pp. 816823.
Use of Genetic Algorithms to Optimize Fiber Optic Probe
Design for the Extraction of Tissue Optical Properties
Gregory M. Palmer and Nirmala Ramanujam*
AbstractThis paper outlines a framework by which the optimal illu-
mination/collection geometry can be identied for a particular biomedical
application. In this paper, this framework was used to identify the optimal
probe geometry for the accurate determination of tissue optical properties
representative of that in the ultraviolet-visible (UV-VIS) spectral range. An
optimal probe geometry was identied which consisted of a single illumina-
tion and two collection bers, one of which is insensitive to changes in scat-
tering properties, and the other is insensitive to changes in the attenuation
coefcient. Using this probe geometry in conjunction with a neural network
algorithm, the optical properties could be extracted with root-mean-square
errors of 0.30
for the reduced scattering coefcient (tested range of
340
), and 0.41 for the absorption coefcient (tested range of
080
).
I. I
NTRODUCTION
The diffuse reectance spectrum, which reects the absorption and
scattering properties of a turbid medium, is sensitive to a number of im-
portant physiological indicators and thus, is a useful tool for the early
diagnosis of precancers and cancers. The illumination/collection geom-
etry is a critical aspect of tissue diffuse reectance spectroscopy in that
it affects sensitivity to the absorption and scattering properties of the
medium, the sensing depth and the signal-to-noise ratio. Specialized
probe geometry designs have been previously shown to be useful in
characterizing tissue optical properties from diffuse reectance spectra
[1], [2].
This paper outlines a framework by which the optimal illumina-
tion/collection geometry can be identied given a particular design ob-
jective. Here, the framework is used to identify the optimal probe ge-
ometry for the accurate determination of tissue optical properties rep-
resentative of that in the ultraviolet-visible (UV-VIS) spectral range.
The unique benets of this approach are 1) no
a priori information is
needed about the tissue absorbers and scatterers and 2) there is no re-
quirement for complex multiple source-detector separation ber probe
geometries.
II. M
ETHODS
A. General Optimization Methodology
The optimization methodology proceeds in the following manner.
First, a population of ber probe geometries was randomly initialized.
Next, diffuse reectance measurements were simulated for each of
these probe geometries for a wide range of tissue optical properties
using Monte Carlo modeling. The training data set for each of the
probe geometries was used to optimize a neural network algorithm
to take the diffuse reectance as an input, and output the optical
properties. The optimized neural network algorithm was applied to an
Manuscript received November 9, 2005; revised November 15, 2006. This
work was supported in part by University of Wisconsin through Radiological
Sciences Training Grant 5T32CA009206-27, sponsored by the National Insti-
tutes of Health, and in part by the National Institutes of Health (NIH) under
Grant 1R01CA100559-01A1. Asterisk indicates corresponding author.
G. M. Palmer is with the Department of Radiation Oncology, Duke Univer-
sity, Durham, NC 27710 USA (e-mail: greg.palmer@duke.edu).
*N. Ramanujam is with the Department of Biomedical Engineering, Duke
University, Durham, NC 27703 USA. (e-mail: nimmi@duke.edu).
Digital Object Identier 10.1109/TBME.2006.889779
0018-9294/$25.00 © 2007 IEEE
... This method requires R-wave positions for accurate template localization and often utilizes dynamic time warping to adapt the template shape to account for respiration and heart rate variations. Rieta et al. (2004) and Vaya et al. (2007) proposed blind source separation using linear independent component analysis to extract atrial activity from 12-lead ECG recordings, assuming statistical independence with non-Gaussian distribution between atrial and ventricular activity. Deep learning methods have shown potential in ECG analysis tasks, but existing P-wave segmentation algorithms often adapt successful architectures from computer vision without addressing limitations like poor generalization on ECGs with pathological rhythms (Saclova et al., 2022;Portet, 2008). ...
Preprint
Full-text available
Electrocardiography (ECG) wave morphology and timing provide critical information for diagnosing arrhythmias and conduction abnormalities, allowing risk stratification for various cardiac diseases. However, extraction of these features becomes challenging in the presence of superimposed waves from distinct cardiac chambers, a common occurrence during pathological rhythms. This work proposes a novel Surrogate-boosted Temporal Contrastive Representation Learning (S-TCRL) frame-work to address this challenge. S-TCRL leverages weak labels, readily obtainable from invasive catheter examinations, to extract latent representations of superimposed P waves. We reformulate the problem from object-wise to sample-wise incomplete information by employing surrogate labels. A 1D fully-convolutional feature pyramid network (FPN) extracts multi-scale features from ECG signals. These features are segmented into equal-sized temporal regions, whose labels are inferred from individual samples using a multiple-instance learning (MIL) paradigm. Non-sequential embeddings are generated to facilitate alignment-free cosine similarity estimation. A temperature-scaled cross-entropy loss function minimizes the distance between embeddings of similar regions (likely containing P waves) while maximizing the distance between dissimilar ones. The framework’s efficacy is evaluated on a custom ECG dataset comprising 3265 short-term recordings from 708 individuals undergoing catheter ablation. S-TCRL achieves significant improvement in the downstream P wave segmentation task compared to two baseline MIL methods. The average re-call and precision for both P wave boundaries reach 70.0% and 80.0%, respectively, exceeding the base-lines’ 63.5% and 67.5%. The results demonstrate the potential of S-TCRL for embedding representation of superimposed P waves and its generalizability to tasks such as arrhythmia classification.
... A number of approaches have been proposed for P wave segmentation, such as correlation-based methods [4,5], QRS-T suppression employing subtraction of a template [6] or source separation techniques [7], and deep learning approaches extracting P wave features by means of supervised learning [8]. To date, only two studies [3,9] have investigated the detection of P waves in pathological cardiac rhythms, however, excluding obfuscated P waves due to the aforementioned challenges. ...
... Thus, many methods have been proposed to separate f-waves from QRST complexes, which can be divided into two categories: (1) multi-lead approaches, such as the blind source separation algorithms [8][9][10][11], and the spatio-temporal cancellation algorithms [12][13][14]; (2) single-lead approaches, including the weighted average beat subtraction (WABS) [15], the diffusion distance non-local Euclidean median filtering (DD-NLEM) [16], and the extended Kalman smoother (EKS) [7]. The multi-lead approaches can exploit the correlation information between channels, and these approaches are usually more accurate than the single-lead approaches. ...
Article
Full-text available
The f-wave extraction (FE) is essential for analysis of atrial fibrillations. However, the state-of-the-art FE methods are model-based, and they cannot well adapt to the QRST complexes with high morphological variabilities which often appear in clinical electrocardiogram (ECG). Recently, the encoder-decoder based deep learning networks have been successfully applied to separate variable speech waveforms. However, how these networks are exploited to extract f-waves from ECG recordings is still unclear. Moreover, these networks require different sources to share the common encoder and decoder, which restricts the effectiveness of source representations. To address these issues, a dual temporal convolutional network called DT-FENet is proposed for single lead FE, which integrates the source-specific encoder-decoder mappings and the information fused attention mechanism to respectively learn the latent masks of f-waves and QRST complexes. The proposed DT-FENet can be considered as a dual-stream extension of the famous Conv-TasNet. Compared to Conv-TasNet, the source-specific encoder-decoder mappings of the DT-FENet can obtain more representative bases and sparser activations in latent feature spaces to facilitate the FE task. To the best of our knowledge, this is the first deep learning method for single-lead FE. Extensive experiments were conducted on the clinical ECG records, and the experimental results show that the proposed DT-FENet performs significantly better than the state-of-the-art FE methods with less than one tenth of unnormalized ventricular residuals and about twice spectral concentrations. The proposed DT-FENet can provide accurate f-wave information for atrial fibrillations detection using single-lead ECGs.
... Convolutive Blind Source Separation (BSS) is a generic inverse problem with broad impact as the extraction of multiple sources from their space-time mixtures is a common problem in various engineering areas. The acoustic source separation in reverberant environments [1,2] can be considered as the signature case, however, the applications span a much wider range including digital communications [3], electrocardiogram measurements [4], radar signal processing [5]. ...
Thesis
Full-text available
In this thesis, a new class of novel Bounded Component Analysis (BCA) algorithms based on sparsity assumption, and their applications on some practical problems are introduced. As an application, the BCA algorithm proposed by Erdogan is demonstrated to separate the direct and reflected echoes which can improve the performance of classical direction of arrival estimation methods based on free space propagation theory. Following, we propose a new BCA framework for the separation of the instantaneous mixtures of sparse and bounded sources. Based on this framework, our first proposed algorithm is named Sparse Bounded Component Analysis (SBCA) which is derived from a geometric objective function defined over a completely deterministic setting. Since the framework is not related to statistical properties of source signal model, it is applicable to sources which can be statistically independent or dependent in both spatial and temporal domains. Then, SBCA framework is extended to the convolutive mixtures of sparse sources and propose time and frequency domain convolutive signal separation algorithms. Finally, a space-time analysis tool is provided to detect and identify brain activity signals that have a non-stationary nature. This tool mainly relies on the short time convergence property of our SBCA framework.
... In general, all artifacts like eye movements, eye blinks, cardio activities are generated by mutually independent source and volume conduction is linear, which makes the first premise is reasonable. However, the acquisition of biomedical signal is usually not linear instantaneous, and therefore, a Convolutive ICA (CICA) approach considering weighted and delayed contributions of signals was proposed in [61]. Additionally, ICA can estimate original signals which are non-Gaussian, but unfortunately sources are usually not known to be Gaussian or non-Gaussian. ...
Article
Full-text available
Electroencephalogram (EEG) plays an important role in identifying brain activity and behavior. However, the recorded electrical activity always be contaminated with artifacts and then affect the analysis of EEG signal. Hence, it is essential to develop methods to effectively detect and extract the clean EEG data during encephalogram recordings. Several methods have been proposed to remove artifacts, but the research on artifact removal continues to be an open problem. This paper tends to review the current artifact removal of various contaminations. We first discuss the characteristics of EEG data and the types of different artifacts. Then, a general overview of the state-of-the-art methods and their detail analysis are presented. Lastly, a comparative analysis is provided for choosing a suitable methods according to particular application.
Article
Electrocardiogram (ECG) signals are normally contaminated by various physiological and nonphysiological artifacts. Among these artifacts baseline wandering, electrode movement and muscle artifacts are particularly difficult to remove. Independent component analysis (ICA) is a well-known technique of blind source separation (BSS) and is extensively used in literature for ECG artifact elimination. In this article, the independent vector analysis (IVA) is used for artifact removal in the ECG data. This technique takes advantage of both the canonical correlation analysis (CCA) and the ICA due to the utilization of second-order and high order statistics for un-mixing of the recorded mixed data. The utilization of recorded signals along with their delayed versions makes the IVA-based technique more practical. The proposed technique is evaluated on real and simulated ECG signals and it shows that the proposed technique outperforms the CCA and ICA because it removes the artifacts while altering the ECG signals minimally.
Article
Background Electroencephalographic (EEG) recordings are used to trace neural activity within the cortex to study brain functioning over time. Introduction During data acquisition, the unequivocal way to reduce artifact is to avoid artifact stimulating events. Though there are certain artifacts that make this task challenging due to their association with the internal human mechanism, in the human-computer interface, these physiological artifacts are of great assistance and act as a command signal for controlling a device or an application (communication). That is why pre-processing of electroencephalographic readings has been a progressive area of exploration, as none of the published work can be viewed as a benchmark for constructive artifact handling. Method This review offers a comprehensive insight into state of the art physiological artifact removal techniques listed so far. The study commences from the single-stage traditional techniques to the multistage techniques, examining the pros and cons of each discussed technique. Also, this review paper gives a general idea of various datasets available and briefs the topical trend in EEG signal processing. Result Comparing the state of the art techniques with hybrid ones on the basis of performance and computational complexity, it has been observed that the single-channel techniques save computational time but lack in effective artifact removal especially physiological artifacts. On the other hand, hybrid techniques merge the essential characteristics resulting in increased performance, but time consumption and complexity remain an issue. Conclusion Considering the high probability of the presence of multiple artifacts in EEG channels, a trade-off between performance, time and computational complexity is the only key for effective processing of artifacts in the time ahead. This paper is anticipated to facilitate upcoming researchers in enriching the contemporary artifact handling techniques to mitigate the expert’s burden.
Conference Paper
Full-text available
In this work it is demonstrated that, taking into account the condi- tions of a convolutive mixture of acoustic signals, it is possible to configure a mixture system whose separation model accomplishes the conditions of an a in- stantaneous mixture. This system is achievable by using stereo boundary mi- crophones; this type of coincident microphones can be used to uniform delays of the propagation channels and to reduce the number of reflections that charac- terize the impulse response of the system. By means of coincident boundary microphones techniques, instantaneous BSS algorithms are applicable, thus providing optimal results with less computational cost. This system is validated in both anechoic and reverberant chambers.
Article
ACC/AHA/ESC guidelines for the management of patients with supraventricular arrhythmias--executive summary. a report of the American college of cardiology/American heart association task force on practice guidelines and the European society of cardiology committee for practice guidelines (writing committee to develop guidelines for the management of patients with supraventricular arrhythmias) developed in collaboration with NASPE-Heart Rhythm Society.
Book
A comprehensive introduction to ICA for students and practitionersIndependent Component Analysis (ICA) is one of the most exciting new topics in fields such as neural networks, advanced statistics, and signal processing. This is the first book to provide a comprehensive introduction to this new technique complete with the fundamental mathematical background needed to understand and utilize it. It offers a general overview of the basics of ICA, important solutions and algorithms, and in-depth coverage of new applications in image processing, telecommunications, audio signal processing, and more.Independent Component Analysis is divided into four sections that cover:* General mathematical concepts utilized in the book* The basic ICA model and its solution* Various extensions of the basic ICA model* Real-world applications for ICA modelsAuthors Hyvarinen, Karhunen, and Oja are well known for their contributions to the development of ICA and here cover all the relevant theory, new algorithms, and applications in various fields. Researchers, students, and practitioners from a variety of disciplines will find this accessible volume both helpful and informative.
Article
Conditions under which a time varying electromagnetic field problem (such as arises in electrophysiology, electrocardiography, etc.) can be reduced to the conventional quasistatic problem are summarized. These conditions are discussed for typical physiological parameters.
Conference Paper
During the last years, peer to peer (P2P) networks are becoming more and more popular because of their inherent properties that allow users to exchange information and also to collaborate between them with a tremendous easiness.In parallel to this evolution, the use of MANETs has been increased due to the low effort needed to their deployment.Nowadays every user equipped with a mobile device like a laptop or a PDA could create an ad-hoc network just configuring their wireless device. Thus, users without any infrastructure can connect to others. Using P2P platforms upon these kind of network will allow users to collaborate and exchange information. However the mayor application area of these P2P platforms is still the fixed network.This paper presents a new architecture for a collaborative working environment especially designed for MANETs solving the main drawbacks of other P2P platforms designed for static wired networks with the requirement of a high bandwidth. Our approach is based on clustering MANETs allowing efficient communication among nearby nodes without impairment of distant peers connectivity.
Article
Automatic analysis of the frequency content of the fibrillatory baseline on the surface ECG accurately reflects the average rate of atrial fibrillation (AF). This frequency measurement correlates with the behavior of AF and predicts the response to administration of ibutilide, a new antiarrhythmic drug. Neither the temporal pattern of fibrillatory frequency in spontaneous paroxysmal or persistent AF, nor its response to chronic antiarrhythmic medication has been studied so far. Holter ECG recordings were made in 20 patients during AF. One minute ECG segments were selected for analysis. The frequency content of the fibrillatory baseline was then quantified using digital signal processing. After high-pass filtering, the QRST complexes were subtracted using a template matching algorithm. The resulting fibrillatory baseline signal was subjected to Fourier transformation, displayed as a frequency power spectrum and the peak frequency (f) was determined. In 11 patients (7 male, 4 female, age 62 +/- 10 years) 31 paroxysmal AF episodes were analyzed. Duration ranged from 1 min to 665 min (115 +/- 175 min). Initial mean peak f measured 5.1 +/- 0.7 Hz (range 3.9 to 6.9 Hz). There was a positive correlation between f and AF duration (R = 0.53, p = 0.002). AF of less than 15 min duration (n = 13) showed a lower f (4.8 +/- 0.6 Hz) when compared with longer lasting episodes (n = 18, 5.3 +/- 0.7 Hz, p = 0.03). In short AF episodes f was constant, whereas in longer-lasting episodes f increased to 5.8 +/- 0.5 Hz (p < 0.001) within 5 min. In 9 patients (9 male, age 58 +/- 8 years) with persistent AF oral antiarrhythmic drugs (amiodarone n = 5, sotalol n = 3, flecainide n = 1) were given prior to electrical cardioversion for prophylaxis of AF recurrence. Frequency measurements were obtained at baseline and 3 to 5 days after initiation of drug administration. At baseline mean f measured 6.9 +/- 0.4 Hz. Frequency was reduced by antiarrhythmic drugs to 5.8 +/- 0.4 Hz (p < 0.001). (1) The duration of paroxysmal AF episodes can be predicted using spectral analysis of ECG recordings of AF episodes. (2) An increase in fibrillatory frequency is associated with AF persistence. (3) This technique can be used to monitor the response to antiarrhythmic medication.
Article
A computerized method to detect atrial fibrillatory activity on the surface electrocardiogram is presented. After ventricular activity was canceled by creating a remainder electrocardiogram, significant differences were found in the percent power of the remainder electrocardiograms for a group of rhythms with atrial fibrillation (mean +/- SD; lead V1, 47.4 +/- 29.7%; lead II, 39.4 +/- 26.8%) and a control group (irregular rhythms or rhythms without readily detectable P waves; lead V1, 17.6 +/- 14.6%; lead II, 19.2 +/- 13.9%) for both leads (p less than 0.0001). A discrimination algorithm that classified a rhythm as atrial fibrillation if the percent power was greater than 32% and if noncoupled P waves were not present had a specificity of 90.0% and a sensitivity of 69.7% for the training set and a specificity of 87.8% and a sensitivity of 68.3% for the test set. In addition, the algorithm correctly detected all 66 of the 66 sinus rhythms tested. The algorithm produced good results that may be incorporated into arrhythmia interpretation systems to improve their specificity.