Conference PaperPDF Available

Cardiac abnormalities detection from compressed ECG in wireless telemonitoring using principal components analysis (PCA)

Authors:

Abstract and Figures

In Wireless telecardiology applications ECG signal is compressed before transmission to support faster data delivery and reduce consumption of bandwidth. However, most of the ECG analysis and diagnosis algorithms are based on processing of the original ECG signal. Therefore, compressed ECG data needs to be decompressed first before the existing algorithms and tools can be applied to detect cardiovascular abnormalities. Decompression will cause delay on the doctor's mobile device and in wireless nodes that have the responsibilities to detect and prioritize abnormal data for faster processing. This is undesirable in body sensor networks (BSNs) as high processing involved in decompression will waste valuable energy in the resource and power constrained sensor nodes. In this paper, in order to diagnose cardiac abnormality such as Ventricular tachycardia, we applied a novel system to analyse and classify compressed ECG signal by using a PCA for feature extraction and k-mean for clustering of normal and abnormal ECG signals.
Content may be subject to copyright.
Cardiac Abnormalities Detection from Compressed
ECG in Wireless Telemonitoring using Principal
Components Analysis (PCA)
Ayman Ibaida #1 , Ibrahim Khalil #2 and Fahim Sufi #3
Distributed Systems & Networking
School of Computer Science & IT
RMIT University, Melbourne, Australia
1ayman.ibaida@student.rmit.edu.au
2ibrahimk@cs.rmit.edu.au
3fahim.sufi@student.rmit.edu.au
Abstract—In Wireless telecardiology applications ECG signal
is compressed before transmission to support faster data delivery
and reduce consumption of bandwidth. However, most of the
ECG analysis and diagnosis algorithms are based on processing
of the original ECG signal. Therefore, compressed ECG data
needs to be decompressed first before the existing algorithms
and tools can be applied to detect cardiovascular abnormalities.
Decompression will cause delay on the doctor’s mobile device
and in wireless nodes that have the responsibilities to detect and
prioritize abnormal data for faster processing. This is undesirable
in body sensor networks (BSNs) as high processing involved
in decompression will waste valuable energy in the resource
and power constrained sensor nodes. In this paper, in order to
diagnose cardiac abnormality such as Ventricular tachycardia,
we applied a novel system to analyse and classify compressed
ECG signal by using a PCA for feature extraction and k-mean
for clustering of normal and abnormal ECG signals.
I. INTRODUCTION
Cardiac diseases are number one killer in the modern world
as many people die because of sudden heart attack. At the
same time, a large number of people die because of the
delay or errors in diagnosing their cardiac diseases. Electro-
cardiogram (ECG) signal has been intensively used by car-
diac specialists to effectively diagnose cardiovascular diseases
[1].Several researchers have proposed various methods such
as digital signal processing, filtering methods, data mining
tools as well as neural networks for classification of cardiac
anomalies [1]. ECG can also be used for continuous patient
monitoring as well as in biometric authentication techniques
[2], [3], [4].
A typical ECG signal as shown in Fig. 1 contains special
waves such as P, T waves as well as QRS complex. Cardiol-
ogists investigate each of these waves, complexes and other
features such as RR interval, PR interval, PR segment, ST
interval and ST segment etc.[5] to diagnose various types of
abnormal cardiac symptoms. However, accurate extraction of
features using numerous signal processing techniques [6], [7]
can be complex and difficult. Recently, wavelet-based QRS
detectors have been suggested by a variety of researches [8].
Such methods have a post-processing phase in which the
redundant R waves or noise peaks are removed. Other re-
searchers used template matching methods to classify the ECG
signal using neural networks [9]. The techniques mentioned
above are applied to the original enormous sampled ECG
signal. Large sampled ECG signal will make processing both
time and resource consuming.
Time / Samples
A
mplitude
P
Q
R
S
T
Fig. 1. ECG Waves
In a typical wireless telemonitoring scenario as shown in
Fig. 2 a patient wears wireless sensors capable of reading
samples of ECG, possibly compress and diagnose, and send it
wirelessly with the result of diagnosis to a central server and
e-doctors (e.g. doctors who are roaming around with mobile
devices) that can take quick action according to its priority
[10]. However, wireless nodes (e.g. Sensor nodes in body
sensor networks (BSNs) or a roaming doctor’s smartphone) are
power and resource constrained. Therefore, it is obvious that
the existing [11], [12], [13], [14] and above mentioned tech-
niques for diagnosis are suitable for implementation neither
in body sensor networks nor in resource-constrained wireless
environment.
In wireless telemonitoring scenarios digitized ECG data
need to be transferred as fast as possible using the mobile
technologies such as MMS,GPRS, HSDPA or zigbee etc.
However, these technologies can not provide high speed
communication [15] and data must be compressed first to
make the transmission energy efficient. Therefore, in this paper
we have proposed a novel technique to analyse abnormalities
978-1-4244-3518-0/09/$25.00 © 2009 IEEE ISSNIP 2009207
Bluetooth
Link
Bandwidth Constrained
Wireless Link
Gateway
Node
Power Constrained ECG Sensor Nodes
diagnoses from compressed ECGs
Doctors Diagnosing on Resource
Constrained Mobile Nodes
Hospital
Roaming Patients
Roaming Doctors
Wireless Sensor Patch (remote aged
care facility or battlefield)
Diagnosis from
Compressed ECGs
zigbee 802.15
Base Station
ECG
Sensor
Fig. 2. A typical wireless telemonitoring scenario. Compression would save energy on power hungry bluetooth device, resource constrained wireless sensor
nodes and doctor’s smartphone. Compression also helps trasmit faster over bandwidth constrained wireless links. Diagnosis of diseases possible on Mobile
nodes from Compressed data.
from compressed ECG data without decompressing the data.
The abnormal cardiac condition considered in this paper is
Ventricular Tachycardia which is a life-threatening cardiac
desease consisting of a rapid rhtym originating from the lower
chambers of the heart. The rapid rate prevents the heart from
filling adequately with blood, and less blood is able to pump
through the body. To achieve this, we first applied a lossless
compression method as described in [15] before transmission.
We then analyzed the compressed ECG signal directly and
extracted the important features of it from the compressed data
using Principal Component Analysis (PCA). The extracted
features are classified as normal and abnormal using k-means
algorithm [16], [17]. In this research we have made contribu-
tions by answering the following research questions:
How can we classify and detect normal and abnormal
ECG data directly from the compressed ECG signal
without decompression?
How can we implement an attribute selection technique
and apply it to extract compressed ECG features using
principal component analysis (PCA)?
Rest of the paper is organized as follows. Section II briefly
discusses the our previously proposed compression algorithm
that is used in this paper. In section III we discuss the basic
system, present analysis of feature subset selection from com-
pressed ECGs using Principal Component Analysis (PCA).
Next, in section IV we show results of PCA and and simple
k-means algorithm to cluster data into abnormal and normal
segments.Finally, section V concludes the paper.
II. BACKGROUND:THE COMPRESSION ALGORITHM
The ECG encoding algorithm is a symbol substitution based
technique preceded by some mathematical transformations.
According to our previous experiments [15], up to 95% com-
pression (compression rate of 20) was harnessed without any
loss of information (lossless compression) when our encoding
algorithm is jointly applied with existing compression and
encryption algorithms. According to the literature, this was
highest possible compression ratio achieved for compressing
publicly available ECG of MIT BIT Arrhythmia Database. The
character set that is used for substitutions is shown in Fig.
3. Application of the algorithm with this character set would
generate, for example, compressed ECGs shown in Fig. 5.
Apart from providing highest possible compression ratio,
the compression algorithm also preserves features for cardio-
vascular diagnosis directly from the compressed ECG. Ac-
cording to the literature [18] and to the best of our knowledge,
multiple diseases were diagnosed from compressed MIT BIH
ECG for the very first time directly from their compressed
ECG.
The benefit of diagnosis from compressed ECG is immense.
As compressed ECG contains less characters, diagnosis from
compressed ECG can be possible (using the techniques shown
in [18] with fewer reading operations (I/O). Most importantly,
for telecardiology applications, where ECG is transmitted and
stored in compressed format, cardiovascular diagnosis is pos-
sible, without performing decompression, saving processing
power, resource and time. Minimizing delays in diagnosis
entail savings of patients’ lives.
The selected compression algorithm is not only a compres-
sion algorithm, but also an encryption algorithm to make sure
secure transmission of data can be achieved. The compression
algorithm consisting of the following stages:
Normalization stage: to rescale the ECG signal and
convert them to the smallest integer value
208
Fig. 3. Character Set for the compressed ECG signal
Differencing stage: to lower the amplitude of the signal
Value encoding: to encode the unsigned normalized dif-
ference
Sign encoding: to encode the signs of the values
Decimal Values permutation stage as a mapping function
Substitution of ASCII character codes to ASCII charac-
ters
III. THE METHODOLOGY
0 500 1000 1500 2000 2500
−2
1
.5
−1
0
.5
0
0
.5
1
1
.5
2
2
.5
3
(a)
0 500 1000 1500 2000 2500
1
.2
−1
0
.8
0
.6
0
.4
0
.2
0
0
.2
0
.4
0
.6
(b)
Fig. 4. (a) Normal ECG sample for pationt CU01 (b) Abnormal ECG sample
for pationt CU01
Huge amount of ECG data is required to be transmitted
over bandwidth constrained wireless networks as well as
limited power sensor nodes. However, sending large amount
of data is a power consuming and will reduce the lifetime of
body sensor networks. Therefore, compression of ECGs and
diagnosis of abnormalities from compressed ECGs will play
key roles in enhancing the lifetime of body sensor networks. In
this paper, we deployed the compression algorithm proposed
in [15] because it is a lossless algorithm. We analysed the
resulting compressed ECG and used data mining tools to
classify it as normal and abnormal. In Fig. 4(a) we can see a
normal ECG sample for patient CU01 from the CU Ventricular
Tachyarrhythmia Database and Fig. 4(b) shows the abnormal
sample for the same patient.
For the purpose of classification of normal and abnormal
cases, we will only use the compressed ECGs as shown in
Fig. 5. If we observe carefully we may not notice a signifi-
cant difference in normal and abnormal compressed signals.
By using special data mining algorithms we can determine
abnormalities in the compressed ECG data.
A. Analysis of Compressed ECG signal
The compressed ECG data contains the characters set shown
in Fig. 3. This compression is done in an wireless node of
the body sensors carried by a patient. Before transmission of
the compressed ECGs over the wireless networks data mining
module in the nodes needs to be trained with the normal
samples of compressed ECGs.
As shown in Fig. 7 character frequency calculation is
performed for each compressed ECG segment. As a result, we
have the frequency count for each character. Since we have
148 characters, if each character is regarded as an attribute,
then we have 148 attributes. But 148 attributes is large number
for clustering (normal and abnormal). Therefore, we applied
an attribute selection technique called Principle Component
Analysis (PCA) for dimensionality reduction.
ECG
Compression
Frequency
Count
Attribute
Selection (PCA)
Clustering of
Normal/Abnormal
(k-mean)
ECG Signal
Disease
Detection
Fig. 7. Block Diagram for the Proposed ECG detection system
B. Attribute Subset Selection
A preprocessing of data using attribute selection algorithm
is a critical issue in data mining solutions, since the training
will be hard and inaccurate using large number of attributes.
Also, it will make the system more complicated and the
processing time will be large if the number of attributes keep
increasing. In this paper, we adopted PCA which is appropriate
if there is a set of samples with large number of variables
(attributes). The algorithm will generate a new small set of
artificial variables called Principle Components which can be
selected and fed to clustering system. We first prepared the
data set for patient Cu01 as an example. For experimentation
we took 12 samples of which 6 are normal and the rest
abnormal. We then compressed each sample and calculated
character frequency to derive the final data set shown in Fig. 6.
By applying PCA on this data set we first generate the
covariance matrix of the data. Next, we derive eigenvectors and
eigenvalues for the covariance matrix which is then rearranged
as a new matrix starting with the eigenvector that corresponds
to the highest eigenvalue, and so on. As a result, this matrix
will be (n×n)matrix where nis the number of variables
(i.e in this case n= 148). After this, we calculate the scores
matrix which is a (n×m)matrix where nis the number of
samples and mis the number of variables. Equation 1 shows
209
(a) (b)
Fig. 5. Compressed ECG samples for patient CU01 (a) Abnormal ECG of 4(b) in Compressed Format (b) Normal ECG of 4(a) in Compressed Format
0 50 100 150
0
50
00
50
0 50 100 150
0
100
200
0 50 100 150
0
100
200
0 50 100 150
0
00
00
0 50 100 150
0
50
100
150
0 50 100 150
0
50
100
150
0 50 100 150
0
50
00
50
0 50 100 150
0
100
200
0 50 100 150
0
50
100
150
0 50 100 150
0
50
00
50
0 50 100 150
0
50
100
150
0 50 100 150
0
100
200
Fig. 6. data set for patient CU01 the first six plots for abnormal samples and the second six plots for normal samples
the general form to calculate the scores for the first principle
component
C1=b11(X1)+b21 (X2)+...b
p1(Xp)(1)
where
C1= the sample score on the principal component 1
bp1= the regression coefficient (or weight) for observed
variable p,
Xp= the sample value of variable no p
Similarly, other principal components (i.e. PC2, PC3,PC4,
and so on) can also be calculated. The challenge now is how
many Principal components we will keep. A simple calculation
reveals that the first few components represent the high portion
of data, which is clearly shown in Table I with the eigenvalues
and the proportion of each eigenvalue of the total data.
If we look at Table I we can clearly notice that the first and
second eigenvalues represent approximately 70% of the total
data. Proportions of each eigenvalue in this table is derived
by dividing the eigenvalue over the total summation of all
eigenvalues obtained as in equation 2.
Pi=ei
k=m
k=1 ek
(2)
210
−6 −5 −4 −3 −2 −1 0 1 2 3 4
−3
−2
−1
0
1
2
PC1
PC2
CU01
Normal
Abnormal
−4 −3 −2 −1 0 1 2 3 4
−5
−4
−3
−2
−1
0
1
2
PC1
PC2
CU03
Normal
Abnormal
−5 −4 −3 −2 −1 0 1 2 3 4
−3
−2.5
−2
−1.5
−1
−0.5
0
0.5
1
1.5
2
PC1
PC2
CU09
Normal
Abnormal
Fig. 8. Class distribution for Principal Component 1 (i.e. PC1) & Principal Component 2 (i.e. PC2) for patients cu01,cu03 and cu09 respectively
TABLE I
EIGENVALUES FOR VARIOUS PRINCIPAL COMPONENTS OF PATIENT CU01
Principal Components Eigenvalue Proportion
Principal Component 1 (PC1) 82.56614 0.56552
Principal Component 2 (PC2) 20.04823 0.13732
Principal Component 3 (PC3) 7.47957 0.05123
Principal Component 4 (PC4) 6.37829 0.04369
Principal Component 5 (PC5) 5.9081 0.04047
Principal Component 6 (PC6) 5.45423 0.03736
Principal Component 7 (PC7) 5.16522 0.03538
Principal Component 8 (PC8) 3.76667 0.0258
Principal Component 9 (PC9) 3.6941 0.0253
where
Piis proportion of the ith eigenvalue
eiis the ith eigenvalue
mnumber of eigenvalues which is the same number of
variables.
Therefore, in our work we will take just the first two
principal components and their corresponding score will be
used as an input for our clustering part. The same procedure
was repeated for all other patients of Ventricular Tachyarrhyth-
mia Database to determine eignevalues and the principal
components.
TABLE II
SCORES FOR PC1 AND PC2 (COMPONENT 1 &2) OF CU01
Sample Scores for Scores for Class
Number PC1 PC2
1 -1.18113 -2.54187 abnormal
2 -1.26634 -2.64893 abnormal
3 -0.88636 -2.97021 abnormal
4 -0.12991 -1.99849 abnormal
5 3.246313 1.579908 abnormal
6 2.38746 0.798551 abnormal
7 -4.92702 0.419675 normal
8 -5.10988 -0.03882 normal
9 -4.82858 0.275417 normal
10 -4.99787 -0.12245 normal
11 -5.02603 0.221886 normal
12 -4.94025 0.02097 normal
IV. RESULTS AND DISCUSSION
Using the procedure discussed earlier we can now derive Ta-
ble II that shows the first two principal component scores for
every normal and abnormal ECG segment. Since we performed
TABLE III
K-MEAN RESULTS FOR CU01
Sample Distance Distance
Number from class1 from class2 Class
1 21.50359 3.930326 2 (abnormal)
2 21.44839 4.478564 2 (abnormal)
3 26.29707 4.35777 2 (abnormal)
4 27.9701 0.733969 2 (abnormal)
5 69.63799 16.59684 2 (abnormal)
6 54.60353 8.494487 2 (abnormal)
7 0.086221 30.91665 1 (normal)
8 0.047432 31.52044 1 (normal)
9 0.041762 29.41073 1 (normal)
10 0.064142 30.10386 1 (normal)
11 0.011507 31.33386 1 (normal)
12 0.01275 29.84699 1 (normal)
tests on 6 normal and another 6 abnormal ECG segments for
every patient we have 12 sets of values for the principal com-
ponent scores. This particula Table II corresponds to patient
CU01 of CU Ventricular Tachyarrhythmia Database. Similiary,
scores can be derived for other patients in the database. Figure
8 show class distribution for Principal Component1&2for
patients cu01,cu03 and cu09 respectively. It is obvious from
this distribution that abnormal ECGs can be easily separated
from the normal ones. Similar tests were performed on all
other patients from Ventricular Tachyarrhythmia Database, and
they all follow the same trend which confirms that abnormal
ECGs can be distinguished from the normal ECGs when PCA
is applied to compressed ECGs of the patients.
Now we just have two variables (i.e PC1 and PC2) which
can be easily fed to a k-mean clustering to classify abnormal
and normal ECG segments. This will further validate our
earlier observations. Table III shows the results for k-mean
algorithm as it is applied to previous data shown in Table
II. From the results it is clear that the distances of samples
1-6 are small for class 2 (abnormal) and large for class 1
(normal). This is why it is classified as class 2 (abnormal).
Similarly, samples 7-12 have small distance from class 1 and
large distance from class 2. Therefore, it is classified as class
1 (normal). This fact is clearly established in Fig. 8 which
shows plots of scores for three patients.
211
V. CONCLUSION AND FUTURE WORK
Because ECG signal is enormous in size [19],compression
algorithms must be used to make the whole tele-cardiology
faster and efficient. A faster solution is of crucial importance
for diagnoses and treatment of cardiovascular diseases. Al-
though ECG compression enables faster transmission, it also
introduces a delay in the processing phase because of the de-
compression. Since existing methods process the original ECG
signal and not the compressed one, this decompression time
can be enough to threaten patient life. However decompression
in wireless telemonitoring will cause delay on the doctor’s
mobile devices. This (decompression) is also undesirable in
body sensor network as more processing will waste valuable
energy resources. To overcome the decompression delay and
make body sensor network energy efficient, in this paper we
implemented the ECG analysis and data mining solution on
the compressed ECG signal using PCA for feature extraction
and k-mean as a clustering technique. Compressed ECG signal
can be fast in transmission, and now we have clearly shown
that we can classify and analyse the compressed ECG signal
to detect cardiac abnormalities. Encouraged by these results
we intend to develop a neural network model to be trained to
classify more diseases in a node of body sensor networks.
REFERENCES
[1] G. Clifford, F. Azuaje, and P. McSharry, Advanced methods and tools
for ECG data analysis. Artech House.
[2] I. Khalil and F. Sufi, “Legendre Polynomials based biometric authen-
tication using QRS complex of ECG,” in Intelligent Sensors, Sensor
Networks and Information Processing, 2008. ISSNIP 2008. International
Conference on, 2008, pp. 297–302.
[3] F. Sufi and I. Khalil, “An automated patient authentication system for
remote telecardiology, in Intelligent Sensors, Sensor Networks and
Information Processing, 2008. ISSNIP 2008. International Conference
on, 2008, pp. 279–284.
[4] F. Sufi, I. Khalil, and I. Habib, “Polynomial distance measurement
for ECG based biometric authentication.” John Wiley & Sons, Ltd.
Chichester, UK, 2008.
[5] Y. Suzuki, “Self-organizing QRS-wave recognition in ECG using neural
networks,” IEEE Transactions on Neural Networks, vol. 6, no. 6, pp.
1469–1477, 1995.
[6] S. Mahmoodabadi, A. Ahmadian, and M. Abolhasani, “ECG feature
extraction using Daubechies wavelets, in Proceedings of the Fifth
IASTED International Conference, Visualization, Imaging, and Image
Processing, Benidorm, Spain, 2005.
[7] K. Minami, H. Nakajima, and T. Toyoshima, “Real-time discrimination
of ventricular tachyarrhythmia withFourier-transform neural network,
IEEE Transactions on Biomedical Engineering, vol. 46, no. 2, pp. 179–
185, 1999.
[8] A. Ghaffari, H. Golbayani, and M. Ghasemi, “A new mathematical
based QRS detector using continuous wavelet transform, Computers
and Electrical Engineering, vol. 34, no. 2, pp. 81–91, 2008.
[9] S. Barro, M. Fernandez-Delgado, J. Vila-Sobrino, C. Regueiro, and
E. Sanchez, “Classifying multichannel ECG patterns with an adaptive
neuralnetwork,” IEEE Engineering in Medicine and Biology Magazine,
vol. 17, no. 1, pp. 45–55, 1998.
[10] B. Lo, S. Thiemjarus, R. King, and G. Yang, “Body sensor network–a
wireless sensor platform for pervasive healthcare monitoring, in The
3rd International Conference on Pervasive Computing. Citeseer, 2005.
[11] J. De Vos and M. Blanckenberg, “Automated pediatric cardiac auscul-
tation,” IEEE Transactions on Biomedical Engineering, vol. 54, no. 2,
pp. 244–252, 2007.
[12] E. Ubeyli, “Eigenvector Methods for Automated Detection of Electro-
cardiographic Changes in Partial Epileptic Patients.
[13] W. Jiang and S. Kong, “Block-based neural networks for personalized
ECG signal classification,” IEEE Transactions on Neural Networks,
vol. 18, no. 6, 2007.
[14] F. Melgani and Y. Bazi, “Classification of Electrocardiogram Signals
With Support Vector Machines and Particle Swarm Optimization, IEEE
Transactions on Information Technology in Biomedicine, vol. 12, no. 5,
pp. 667–677, 2008.
[15] F. Sufi and I. Khalil, “Enforcing secured ECG transmission for real time
telemonitoring: A joint encoding, compression, encryption mechanism,”
Security and Communication Networks, vol. 1, no. 5, 2008.
[16] I. Jolliffe, Principal component analysis. Springer verlag, 2002.
[17] M. Dunham, Data mining introductory and advanced topics. Prentice
Hall/Pearson Education, 2003.
[18] F. Sufi, Q. Fang, I. Khalil, and S. S. Mahmoud, “Novel methods of faster
cardiovascular diagnosis in wireless telecardiology,” IEEE Journal on
Selected Areas in Communications, vol. 27, no. 4, MAY 2009.
[19] F. Sufi, I. Khalil, Q. Fang, and I. Cosic, “A mobile web grid based
physiological signal monitoring system,” in Technology and Applications
in Biomedicine, 2008. ITAB 2008. International Conference on, 2008,
pp. 252–255.
212
... The principle aim of this paper is to enable patients with various cardiac conditions [3] to stay connected with their family, friends and the community by leveraging advances in communication infrastructure using context-aware service frameworks that facilitate the delivery of a wide range of services using smart sensors and devices. The situational context generated from wearable sensors and ambient devices contains health information of the user [4,5]. ...
... This facilitates the delivery of services among the members a group or between a peer by means of social network. Our cloud-based platform eases the management of context having large variations from vast number of AAL systems and allows simple integration of different social networking services using simple interface [2,3]. ...
Conference Paper
Full-text available
Cardiovascular diseases are major cause of deaths throughout the world. In this work, we develop a context-aware system for wellness monitoring of older adults who leave alone in home and suffers from cardiac disease. The focus here is the integration of social networking services with conventional remote monitoring services by utilizing a scalable cloud platforms. The goal here is to expand patient's social linkage by identifying similarity in his/her cardiac conditions. Here we build a cloud-oriented context-aware model that captures health parameters using modern fit bit device and ECG sensors. The raw data are sent the cloud platforms provided by Amazon Web Service (AWS) where data is converted to high level context. Using social networks this high level context information is send to patient's friends, family and doctors who are interested to know about his/her health condition. The interested parties get notified by Facebook when the context-aware system detects any changes. That is, using this platform a cardiac patient who live alone and need continuous monitoring is always get connected with virtual community by means of his/her health information. This is a new model that utilizes the context data generated by wearable sensors to create interesting social networking services. The system is also designed to promote cardiac patients to interact with their community of interest using various context-aware social services. The results obtained for this innovative model show a new approach of wellness monitoring using social networks.
... Some part of methodology and analysis of this work was described in our previous work [17]. We are also encouraged by our previous studies in biomedical data analysis [2,11,18] and disease diagnosis [8] to develop this probabilistic leaning model for predication of abnormal clinical events in real-time personalized monitoring. ...
... In this paper, we consider Principle Component Analysis (PCA) [18] to capture and validate the variations in multiple vitals and isolate clinical states for unsupervised learning from large number of patients. PCA clusters normal and various abnormalities in multi-parameter data based on their correlated behaviours using principal component features. ...
Article
Full-text available
The objective of this study is to develop a probabilistic model for predicting the future clinical episodes of a patient using observed vital sign values prior to the clinical event. Vital signs (e.g. heart rate, blood pressure) are used to monitor a patient’s physiological functions of health and their simultaneous changes indicate a transition of a patient’s health status. If such changes are abnormal then it may lead to serious physiological deterioration. Chronic patients living alone at home die of various diseases due to the lack of an efficient automated system having prior prediction ability. Our developed system can make probabilistic predictions of future clinical events of an unknown patient in real-time using the learned temporal correlations of multiple vital signs from many similar patients. In this paper, Principal Component Analysis (PCA) is used to separate patients with known medical conditions into multiple categories and then Hidden Markov Model (HMM) is adopted for probabilistic classification and prediction of future clinical states. The advantage of using dynamic probabilistic model over static predictor model for solving our problem is analysed by comparing the results obtained from HMM with a neural network based learning model. Both the learning models are trained and evaluated using six vital signs data of 1023 patient records collected from the MIMIC-II database of MIT physiobank archive. The best HMM models are selected using maximum likelihood probabilities and further used in personalized remote monitoring system to forecast the most probable forthcoming clinical states of a continuously monitored patient. The scalable power of cloud computing is utilized for fast learning of various clinical events from large samples. Our results suggest that the developed technique using multiple physiological parameter trends can significantly enhance the traditional home-based monitoring systems in terms of clinical abnormality detections and predictions.
... As we have in total six signals, so to visualize the partitions in 3-dimensions we took first 3 principal component values after applying PCA. Generally, first 3 principal components contain most of information of all signals [47]. An example of such dynamic pattern discovery for three patients (4, 5 and 6) in Dataset-1 having CHF/Pulmonary Edema is shown in Figure 5. ...
Article
Full-text available
Background and objectivesHealth professionals look for specific patterns by correlating multiple physiological data in the process of deciding treatments to remedy clinical abnormalities. Biomedical data exhibit some common patterns in the event of identical clinical illnesses. The primary interest of this work is automatic discovery of such patterns in vital sign data (e.g. heart rate, blood pressure) using unsupervised learning and utilising them to identify patients with similar clinical conditions. MethodsA patient clustering method is developed that efficiently isolates patients into multiple groups by discovering dynamic patterns in multi-dimensional vital sign data. A dynamic partitioning algorithm and a patient clustering approach is proposed by introducing a measure namely aggregated instance-wise uncertainty (AIU) computed from multi-dimensional physiological time-series data. ResultsThe developed model is evaluated qualitatively using principal component analysis and silhouette value; and quantitatively in terms of its ability of clustering patients associated with different clinical situations. Experiments are conducted using real-world biomedical data of patients having various clinical conditions. Thee observed accuracy was 82.85% and 91.17% on two experimental datasets comprised of 35 and 34 patients data respectively.The comparisons show that the proposed approached outperformed than other methods in state-of-the-art approach. Conclusions The experimental outcomes demonstrate the effectiveness of the proposed approach in discovering distinct patterns with predictive significance.
... Finally, the signals and patient confidential information as well as diagnoses report or any urgent alerts are sent to the central hospital servers via the Internet. Doctors can check those biomedical signals and possibly make a decision in case of an emergency from anywhere using any device 3 . Using Internet as main communication channel introduces new security and privacy threats as well as data integration issues. ...
Article
The Health Insurance Portability and Accountability Act(HIPAA) privacy and security regulations are two crucial provisions in the protection of healthcare privacy. According to HIPAA patient information sent through the internet should be protected and secured. In this paper, to comply with the HIPAA regulations, chaos cryptographic technique is used to encrypt the confident data into unreadable form. After the data decryption process, the encrypted data concealment in ECG signals. A wavelet basedsteganography technique is used to merge the encrypted patient confident data with the host ECG signal. To examine the performance of the proposed technique, Percentage Residual Difference (PRD) a distortion measurement technique is used. The proposed technique brings formidable security protection for bio-medical data with low distortion (less than 1% ) and ECG signal remains diagnosable after watermarking and as well as after watermarks are detach from the watermarked signal.
Article
The main purpose of this paper is to develop and test internet of things (IoT) based physiological parameters monitoring system. This system is implemented using different multilabel classifier (MLC) algorithms and have been used for the health status prediction and classifications. A method has been proposed and developed as medical decision support system (MDSS) based on the outcome of the different MLC algorithms. The developed MDSS enables real time assistance to local and remote supervisor for in time decision better diagnosis plan The performance parameters and results of the different MLC algorithms has been evaluated, in terms of several, Accuracy, Precision, Recall F-measure, and MCC etc. to identify the best algorithm for classification and prediction of health status. Gradient Boost algorithm of classification outperform the other algorithm and achieved approx 94% of accuracy. This study concludes that the developed MDSS will be very useful in remote areas to assist patient and health practitioners to predict heath status and for quick decision making in case of medical emergencies.
Article
The project proposes the enhancement of protection system for secret data communication through encrypted data concealment in ECG signals. The proposed encryption technique used to encrypt the confidential data into unreadable form and not only enhances the safety of secret carrier information by making the information inaccessible to any intruder having a random method. After data encryption, the data hider will conceal the secret data into the ECG signal coefficients. Although encryption achieves certain security effects, they make the secret messages unreadable and unnatural or meaningless. This system is still enhanced with encrypt messages using chaos crypto system. This is the reason a new security approach called reversible data hiding arises. It is the art of hiding the existence of data in another transmission medium to achieve secret communication. It does not replace cryptography but rather boosts the security using its obscurity features. Here the discrete wavelet transformation is used to decompose an ECG signal to different frequency subbands. The data hiding technique uses the LSB replacement algorithm for concealing the secret message bits into the high frequency coefficients. In the data extraction module, the secret data will be extracted by using relevant key for choosing the relevant data to extract the data. By using the decryption keys, extracted text data will be decrypted from encryption to get the original information. Finally the performance of this proposal in encryption and data hiding will be analyzed based on image and data recovery.
Article
With the growing number of aging population and most of the patients are suffering from cardiac diseases, it is conceivable that remote ECG patient monitoring systems are expected to be widely used as Point-of-Care (PoC) applications in hospitals around the world. Therefore, huge amount of ECG signal collected by Body Sensor Networks (BSNs) from remote patients at homes will be transmitted along with other physiological readings such as blood pressure, temperature, glucose level etc. and are then diagnosed by remote patient monitoring systems. It is utterly important that patient confidentiality is protected while data is being transmitted over the public network as well as when they are stored in hospital servers used by remote monitoring systems. In this paper, a wavelet based steganography technique has been introduced which combines encryption and scrambling technique to protect patient confidential data. The proposed method allows ECG signal to hide its corresponding patient confidential data and other physiological information thus guaranteeing the integration between ECG and the rest. It is found that the proposed technique provides high security protection for patients data and ECG data remains diagnosable after watermarking (i.e. hiding patient confidential data) and as well as after watermarks (i.e. hidden data) are removed from the watermarked data.
Book
Full-text available
Introduction Introduction Related Concepts Data Mining Techniques Core Topics Classification Clustering Association Rules Advanced Topics Web Mining Spatial Mining Temporal Mining Appendix Index Salient Features Covers advanced topics such as Web Mining and Spatial/Temporal Mining. Includes succinct coverage of Data Warehousing, OLAP, Multidimensional Data, and Preprocessing. Concise coverage on distributed, parallel, and incremental algorithms. Provides case studies. Offers clearly written algorithms to better understand techniques. Algorithms are presented in a pseudocode. Includes a reference on how to use Prototypes and DM products.
Article
Full-text available
With recent advances in wireless sensor networks and embedded computing technologies, miniaturized pervasive health monitoring devices have become practically feasible. In addition to providing continuous monitoring and analysis of physiological parameters, the recently proposed Body Sensor Networks (BSN) incorporate context aware sensing for increased sensitivity and specificity. To facilitate research and development in BSN and multi-sensor data fusion, a BSN hardware development platform is presented. With its low power, flexible and compact design, the BSN nodes provide a versatile environment for wireless sensing research and development.
Conference Paper
Full-text available
In this work, we have developed and evaluated an electrocardiogram (ECG) feature extraction system based on the multi-resolution wavelet transform. ECG signals from Modified Lead II (MLII) are chosen for processing. The result of applying two wavelet filters (D4 and D6) of different length on the signal is compared. The wavelet filter with scaling function more closely similar to the shape of the ECG signal achieved better detection. In the first step, the ECG signal was de-noised by removing the corresponding wavelet coefficients at higher scales. Then, QRS complexes are detected and each complex is used to locate the peaks of the individual waves, including onsets and offsets of the P and T waves which are present in one cardiac cycle. We evaluated the algorithm on MIT-BIH Database, the manually annotated database, for validation purposes. The proposed QRS detector achieved sensitivity of 75 . 2 % 18 . 99 ± and a positive predictivity of 45 . 4 % 00 . 98 ± over the validation database.
Article
Full-text available
With the rapid development wireless technologies, mobile phones are gaining acceptance to become an effective tool for cardiovascular monitoring. However, existing technologies have limitations in terms of efficient transmission of compressed ECG over text messaging communications like SMS and MMS. In this paper, we first propose an ECG compression algorithm which allows lossless transmission of compressed ECG over bandwidth constrained wireless link. Then, we propose several algorithms for cardiovascular abnormality detection directly from the compressed ECG maintaining end to end security, patient privacy while offering the benefits of faster diagnosis. Next, we show that our mobile phone based cardiovascular monitoring solution is capable of harnessing up to 6.72 times faster diagnosis compared to existing technologies. As the decompression time on a doctor's mobile phone could be significant, our method will be highly advantageous in patient wellness monitoring system where a doctor has to read and diagnose from compressed ECGs of several patients assigned to him. Finally, we successfully implemented the prototype system by establishing mobile phone based cardiovascular patient monitoring.
Conference Paper
Full-text available
In this paper we propose a new Legendre Polynomials based ECG biometric technique that can efficiently be used for person indentification and authentication. we apply high-order Legendre Polynomials on QRS Complex of ECG which is considered one of the most unique signature bearing parts. We show that coefficients generated from various degrees of polynomial matchings are unique for the same person but We show that coefficients generated from various degrees of polynomial matchings are unique for the same person but are significantly different from others. We also show that even with a 4th degree ploynomial fit person authenitication/identification is possible with high degree of accuracy. This is an interesting result as we can achieve significant reduction of key sizes when coefficients generated by these fits are used as unique keys for authentication and verification of subjects.
Chapter
Principal component analysis has often been dealt with in textbooks as a special case of factor analysis, and this tendency has been continued by many computer packages which treat PCA as one option in a program for factor analysis—see Appendix A2. This view is misguided since PCA and factor analysis, as usually defined, are really quite distinct techniques. The confusion may have arisen, in part, because of Hotelling’s (1933) original paper, in which principal components were introduced in the context of providing a small number of ‘more fundamental’ variables which determine the values of the p original variables. This is very much in the spirit of the factor model introduced in Section 7.1, although Girschick (1936) indicates that there were soon criticisms of Hotelling’s method of PCs, as being inappropriate for factor analysis. Further confusion results from the fact that practitioners of ‘factor analysis’ do not always have the same definition of the technique (see Jackson, 1981). The definition adopted in this chapter is, however, fairly standard.
Book
Principal component analysis is central to the study of multivariate data. Although one of the earliest multivariate techniques it continues to be the subject of much research, ranging from new model- based approaches to algorithmic ideas from neural networks. It is extremely versatile with applications in many disciplines. The first edition of this book was the first comprehensive text written solely on principal component analysis. The second edition updates and substantially expands the original version, and is once again the definitive text on the subject. It includes core material, current research and a wide range of applications. Its length is nearly double that of the first edition. Researchers in statistics, or in other fields that use principal component analysis, will find that the book gives an authoritative yet accessible account of the subject. It is also a valuable resource for graduate courses in multivariate analysis. The book requires some knowledge of matrix algebra. Ian Jolliffe is
Conference Paper
Cardiovascular disease (CVD) being the number one killer for many of the developed nations, real-time patient monitoring via the mobile phone network is increasingly becoming popular. The CVD patients, as subscribers for the CVD monitoring service providers, access to the facilities before initiating the dedicated services. However, this authentication must be secured, since the service providers often hold sensitive health information of their subscribers. In this paper, we propose a fully automated and integrated cardiovascular patient authentication system using patients ECG as a biometric entity. The proposed ECG recognition method is up to 12 time faster than existing ECG based biometric algorithms, requires up to 6.5 times less template storage, needs only 2.49 (average) acquisition time with the a high accuracy rate (up to 95%) when experimented a small population size of 15. With this new authentication mechanism in place, the cardiovascular patients no longer need to provide additional details like user name or password for identification purposes to access their health monitoring facility, making the remote tele-cardiology application faster than existing authentication approaches.
Conference Paper
Monitoring of physiological signals of disaster affected patients inherit several challenges. First of all, the care providers of a disaster zone are susceptible to health hazards for themselves. Fast and easy transportation of monitoring equipments is crucial for survival of the injured. Moreover, Enormous physiological data transmission from the disaster zone to the medical server needs to be regulated to prevent network congestion. In this paper, we are proposing a mobile grid based health content delivery service, which can be useful for vital signal monitoring from a remote location. The proposed system is specifically designed for monitoring of a group that is very much mobile and dynamic in nature. Therefore, during a catastrophic event like earth quake, flood, cyclone the whole system can be transported with minimal mobility to the disaster affected patients. Minimally trained people are capable of installing the system within the disaster affected area entirely in ad-hoc manner. Medical experts can monitor the group from a safe location and provide specialist advice for the early recovery of the affected patients. To deal with network congestion, local intelligence is applied within the mobile patient monitoring system. Therefore, only medically urgent information is transmitted to the hospital server or central server. Application of grid network provides additional computational power to analyze raw physiological signal to identify possible health hazards for the monitored patients. In addition, the proposed mobile grid provides load sharing and redundancy of patient data, which are of prime importance for a disaster zone.