ArticlePDF Available

Deep‐learning‐based seizure detection and prediction from electroencephalography signals

Authors:

Abstract

Electroencephalography (EEG) is among the main tools used for analyzing and diagnosing epilepsy. The manual analysis of EEG must be conducted by highly trained clinicians or neuro-physiologists; a process that is considered to have a comparatively low inter-rater agreement. Furthermore, the new data interpretation consumes an excessive amount of time and resources. Hence, an automatic seizure detection and prediction system can improve the quality of patient care in terms of shortening the diagnosis period, reducing manual errors, and automatically detecting debilitating events. Moreover, for patient treatment, it is important to alert the patients of epilepsy seizures prior to seizure occurrence. Various distinguished studies presented good solutions for two-class seizure detection problems with binary classification scenarios. To deal with these challenges, this paper puts forward effective approaches for EEG signal classification for normal, pre-ictal, and ictal activities. Three models are presented for the classification task. Two of them are patient-specific, while the third one is patient non-specific, which makes it better for the general classification tasks. The two-class classification is implemented between normal and pre-ictal activities for seizure prediction and between normal and ictal activities for seizure detection. A more generalized three-class classification framework is considered to identify all EEG signal activities. The first model depends on a Convolutional Neural Network (CNN) with residual blocks. It contains thirteen layers with four residual learning blocks. It works on spectrograms of EEG signal segments. The second model depends on a CNN with three layers. It also works on spectrograms. On the other hand, the third model depends on Phase Space Reconstruction (PSR) to eliminate the limitations of the spectrograms used in the first models. A five-layer CNN is used with this strategy. The advantage of the PSR is the direct projection from the time domain, which keeps the main trend of different signal activities. The third model deals with all signal activities, and it was tested for all patients of the CHB-MIT dataset. It has a superior performance compared to the first models and the state-of-the-art models.
BASIC RESEARCH
Deep-learning-based seizure detection and prediction from
electroencephalography signals
Fatma E. Ibrahim
1
| Heba M. Emara
1
| Walid El-Shafai
1,2
|
Mohamed Elwekeil
1,3
| Mohamed Rihan
1,3
| Ibrahim M. Eldokany
1
|
Taha E. Taha
1
| Adel S. El-Fishawy
1
| El-Sayed M. El-Rabaie
1
|
Essam Abdellatef
4
| Fathi E. Abd El-Samie
1,5
1
Department of Electronics and Electrical
Communications Engineering, Faculty of
Electronic Engineering, Menoufia
University, Menouf, Egypt
2
Security Engineering Lab, Computer
Science Department, Prince Sultan
University, Riyadh, Saudi Arabia
3
Department of Electrical and
Information Engineering (DIEI),
University of Cassino and Southern Lazio,
Cassino, 03043, Italy
4
Delta Higher Institute for Engineering
and Technology (DHIET), Mansoura,
Egypt
5
Department of Information Technology,
College of Computer and Information
sciences, Princess Nourah Bint
Abdulrahman University, Riyadh, Saudi
Arabia
Correspondence
Fatma E. Ibrahim, Department of
Electronics and Electrical
Communications Engineering, Faculty of
Electronic Engineering, Menoufia
University, Menouf 32952, Egypt.
Email: eng.fatmaelsayed@gmail.com
Abstract
Electroencephalography (EEG) is among the main tools used for analyzing
and diagnosing epilepsy. The manual analysis of EEG must be conducted by
highly trained clinicians or neuro-physiologists; a process that is considered to
have a comparatively low inter-rater agreement. Furthermore, the new data
interpretation consumes an excessive amount of time and resources. Hence, an
automatic seizure detection and prediction system can improve the quality of
patient care in terms of shortening the diagnosis period, reducing manual
errors, and automatically detecting debilitating events. Moreover, for patient
treatment, it is important to alert the patients of epilepsy seizures prior to sei-
zure occurrence. Various distinguished studies presented good solutions for
two-class seizure detection problems with binary classification scenarios. To
deal with these challenges, this paper puts forward effective approaches for
EEG signal classification for normal, pre-ictal, and ictal activities. Three
models are presented for the classification task. Two of them are patient-spe-
cific, while the third one is patient non-specific, which makes it better for the
general classification tasks. The two-class classification is implemented
between normal and pre-ictal activities for seizure prediction and between nor-
mal and ictal activities for seizure detection. A more generalized three-class
classification framework is considered to identify all EEG signal activities. The
first model depends on a Convolutional Neural Network (CNN) with residual
blocks. It contains thirteen layers with four residual learning blocks. It works
on spectrograms of EEG signal segments. The second model depends on a
CNN with three layers. It also works on spectrograms. On the other hand, the
third model depends on Phase Space Reconstruction (PSR) to eliminate the
limitations of the spectrograms used in the first models. A five-layer CNN is
used with this strategy. The advantage of the PSR is the direct projection from
the time domain, which keeps the main trend of different signal activities. The
third model deals with all signal activities, and it was tested for all patients of
the CHB-MIT dataset. It has a superior performance compared to the first
models and the state-of-the-art models.
Received: 5 September 2021 Revised: 19 January 2022 Accepted: 19 January 2022
DOI: 10.1002/cnm.3573
Int J Numer Meth Biomed Engng. 2022;e3573. wileyonlinelibrary.com/journal/cnm © 2022 John Wiley & Sons Ltd. 1of30
https://doi.org/10.1002/cnm.3573
KEYWORDS
Convolutional Neural Network (CNN), electroencephalography, epilepsy, Phase Space
Reconstruction (PSR), seizure prediction, spectrogram
1|INTRODUCTION
About 50 million people around the world are affected by epilepsy disorders of different kinds. Any person, of any age,
gender, race, or class, may be affected by epilepsy. In addition, epilepsy seizures can also vary in frequency of occur-
rence. Such seizures sometimes cause cognitive disorders, which may lead to physical injury of the patients.
1
Epilepsy
is recognized by the World Health Organization (WHO) as a public health concern because of its physical and psycho-
logical consequences. Moreover, epilepsy may lead to premature death, loss of work productivity, and increased
healthcare needs and expenditure.
2
For diagnosing epileptic seizures, distinct screening techniques have been devel-
oped; including Electroencephalography (EEG), positron emission tomography, magneto encephalography, and mag-
netic resonance imaging. EEG signals are characterized by being easily acquired with portable devices.
3
EEG can be defined as an electrophysiological exploration method by which electrical activities of the brain are
measured using electrodes fixed on the scalp.
4
These electrodes may be bulky for patients. Utilization of EEG signals
for diagnosing epilepsy is time- and effort-consuming; as epileptologists have to screen EEG signals minute by minute.
Furthermore, human error is inevitable. Hence, a computer-based diagnosis, by which epileptic seizures can be early
detected, is expected to help the patients.
59
Artificial intelligence covers several areas and includes several branches
such as Machine Learning (ML) and Deep Learning (DL).
Conventional ML algorithms, including feature extraction and classification, were formerly used before the appear-
ance of DL. Hand-crafted features limit the performance of the classification algorithms, but deep features are preferred
due to their better representation of signals and images. Such techniques have achieved great progress, when used in
many aspects of medicine, especially in the diagnosis of epileptic seizures. In many fields, such as anomaly detection
from medical signals and images, feature learning, target monitoring, and recognition; DL has achieved great
advances.
1012
In this paper, we propose an efficient strategy for both seizure detection and prediction from medical EEG signals.
Three models are presented for the classification task. Two of them are patient-specific, while the third one is patient
non-specific. EEG signals for epilepsy patients can be divided into three states: normal (inter-ictal), ictal (seizure), and
pre-ictal which represents the period of 3060 min before the ictal state.
13
We assumed in this paper that the pre-ictal
state occurs 30 min before the ictal state. The two-class classification is implemented between normal and pre-ictal
activities for seizure prediction and between normal and ictal activities for seizure detection. A more generalized three-
class classification framework is considered to identify all EEG signal activities. For the first two proposed models, the
spectrogram estimation process is performed on EEG signal segments to convert them to 2D images. Deep features are
extracted from the spectrogram images using a Convolutional Neural Network (CNN). The extracted features are used
for the classification process. Besides, we propose two CNN models that differ in the number of deep layers. The first
model consists of thirteen deep layers and four residual learning blocks, while the second model includes three deep
layers, only. The third model depends on Phase Space Reconstruction (PSR) to eliminate the limitations of the spectro-
grams used in the first two models. A five-layer CNN is used with this strategy. The advantage of the PSR is the direct
projection from the time domain, which keeps the main trend of different signal activities. The third model deals with
all signal activities, and it has been tested for all patients of the CHB-MIT dataset. The main objectives of this study are
as follows:
Proposal of efficient seizure detection and prediction models by generating spectrogram images from EEG signals.
Proposal of CNNs that differ in the number of deep layers, in order to study the effect of varying the number of layers
on the classification performance.
Getting the best possible results from the proposed CNNs by studying their performance with different Optimization
Algorithms (OAs) and Learning Rates (LRs).
Proposal of an efficient patient non-specific approach that depends on 2D PSR images, in order to eliminate the limi-
tations of the spectrograms.
2of30 IBRAHIM ET AL.
Comparison of the proposed CNNs with various state-of-the-art CNNs, in order to ensure the superiority of the pro-
posed models.
In this paper, Section 2covers the main ideas of the related work. Section 3describes the main architecture of the
proposed patient-specific approach. Section 4presents the main architecture of the proposed patient non-specific
approach. Section 5gives the conclusion of this work.
2|RELATED WORK
M. Talha et al.
14
have developed an end-to-end CNN for seizure detection. Their main target was decreasing the num-
ber EEG signal channels to only two channels and comparing the results with those obtained from all channels. EEG
data of 29 pediatric patients, who were admitted to KK Women and Children's Hospital, has been used. The SeizNet
model was used. It yields a sensitivity of 95.8% and 0.17 false alarms per hour on full channel data, while on two-chan-
nel data, SeizNet yields a sensitivity of 93.3% and a 0.58 false alarms per hour. In,
15
before the Nested Long Short-Term
Memory (NLSTM) model, a CNN with three convolution layers is firstly used. Then, the high-level features obtained
from the NLSTM model are fed into the softmax layer to obtain the predicted labels. The accuracy range of this method
is 98.44% to 100% in ten different experiments on Bonn University database. Moreover, the authors built on a larger
EEG database. The average sensitivity was 97.47%, specificity was 96.17%, and the false detection rate was 0.487 per
hour. For the CHB-MIT scalp EEG dataset, this method achieved a segment-level sensitivity of 94.07% with a false
detection rate of 0.66 per hour.
SeizureNet, as a framework of DL, was introduced by using convolutional layers with dense connections on the
TUH EEG seizure corpus dataset. An overall weighted Fscore of up to 0.90 was achieved by this framework.
16
Covert
et al.
17
proposed the Temporal Graph Convolutional Network (TGCN) for automatic seizure detection. For the seizure
detection task on the epilepsy dataset collected at the Boston Children's Hospital, a 2D CNN model was used to obtain
frequency-domain and time-domain features of EEG signals to characterize seizures. This method, for cross-patient
EEG data, gave an overall sensitivity of 90.00%, a specificity of 91.65%, and an overall accuracy of 98.05%, for the whole
dataset of 23 patients.
12
AlexNet has been applied through the transformation of 1D signals to 2D images.
18
B. Bouaziz et al.
19
managed to segment the EEG signals of CHB-MIT dataset into 2-second frames, and then
transformed them with a spatial representation by producing a set of intensity images. These images are fed to a
CNN, which has a total of eight layers, comprising one initial input layer, five hidden layers, one fully-connected
layer, and an output layer. This approach achieved an accuracy of 99.48%. To reduce the dimensionality, and
then allow the Genetic Algorithm (GA) classification, Rajaguru et al.
20
adopted Multi-Layer Auto-Encoders
(MLAEs) and Expectation Maximization merged with Principal Component Analysis (EM-PCA). The perfor-
mance index represented in classification accuracy was 93.78%. Normal and abnormal brain activities were
detected with a great focus by Roy et al.
21
as they used four different deep learning schemes. They developed the
ChronoNet model, which achieved 90.60% and 86.57% for training and testing accuracies, respectively.
A multi-scale 3D-CNN with a bi-directional gated recurrent unit model was introduced by Choi et al.
22
for cross-
patient seizure detection. Short-Time Fourier Transform (STFT) was used to get spectral and temporal features from
EEG signals. The authors evaluated their proposed model on two datasets: the CHB-MIT dataset and the Seoul National
University Hospital (SNUH) scalp EEG dataset. Their approach achieved sensitivities of 89.4% and 97% on the CHB-
MIT and the SNUH datasets, respectively. Truong et al.
23
used the Freiburg Hospital intracranial EEG (iEEG) dataset,
the CHB-MIT dataset, and the American Epilepsy Society (AES) seizure prediction challenge dataset. To construct spec-
trograms, the STFT algorithm has been applied on raw data. This patient-oriented model gave sensitivities of 81.4%,
81.2%, and 75% and FPR values of 0.06/h, 0.16/h, and 0.21/h, for the prementioned datasets, respectively.
In addition, Z. Wang et al.
24
presented an approach in which data is encoded in the form of different types of
images, namely Gramian Angular Fiields (GAFs) and Markov transition fields. S. Barra et al.
25
presented a framework
in which GAF images are used to train an ensemble of CNNs. They also provided a multi-resolution imaging approach
to feed each CNN. A. Emamia et al.
26
applied CNNs on long-term EEG recordings that include epileptic seizures. After
filtering, CNNs are used to classify EEG images as seizure or non-seizure cases. R. Yuvaraj et al.
27
proposed a CNN-
based unsupervised feature learning framework for automated seizure onset detection. M. Zhou et al.
28
designed a
framework to distinguish ictal, pre-ictal, and inter-ictal segments for epileptic seizure detection from raw EEG signals
instead of using manual feature extraction.
IBRAHIM ET AL.3of30
Subasi et al.
29
introduced a hybrid model for epileptic seizure detection using GAs and Particle Swarm Optimization
(PSO) for EEG data classification. The SVM parameters are optimized using GA- and PSO-based approaches. The PSO-
based approach outperforms the GA-based approach in terms of classification accuracy. This hybrid SVM classifier
achieved a classification accuracy of 99.38%. Alickovic et al.
8
proposed an automated seizure detection and prediction
approach for EEG signals based on de-noising of these signals. Multi-Scale Principal Component Analysis (MSPCA) is
applied. After that, the EEG signal is decomposed using Empirical Mode Decomposition (EMD), Discrete Wavelet
Transform (DWT), or wavelet packet decomposition. Then, to extract relevant features, statistical metrics are applied.
Finally, classification is performed using an ML algorithm. For ictal versus inter-ictal classes, 100% accuracy was
achieved. This approach distinguished between inter-ictal, pre-ictal, and ictal EEG states, with an overall accuracy
of 99.77%.
Qaisar et al.
30
employed intelligent event-driven EEG signal acquisition to achieve real-time compression as well as
effective signal processing and data transmission. The experimental results showed that the developed approach
achieves an overall 3.3-fold compression and transmission bandwidth reduction. It guarantees a significant decrease in
post-analysis and classification processing activities. The system performance has been investigated on a standard
three-class EEG epileptic seizure dataset. For the three classes, the best average classification accuracy of 96.4% has
been achieved on the Andrzejak dataset. For EEG signal classification, a feature extraction network based on Local
Graph Structure (LGS) has been introduced in ref. [31]. Logically extended LGS, symmetric LGS, vertical LGS, vertical
symmetric LGS, zigzag horizontal LGS, zigzag horizontal middle LGS, zigzag vertical LGS, and zigzag vertical middle
LGS were exploited. An ensemble feature extraction network has been formed by combining these LGS methods with
the DWT. LGS was used for feature extraction, and 2D-DWT was used for pooling. Two well-known feature reduction
techniques, relief and Neighborhood Component Analysis (NCA) were used in tandem during the feature reduction
phase. Two publicly available EEG datasets were used in the experiments to test the ensemble LGS feature extraction
method.
Tunable-Q factor Wavelet Transform (TQWT) and bootstrap aggregation using EEG signals have been used to
develop an automated epilepsy detection approach in ref. [32]. The TQWT should be used first to decompose the EEG
signal segments into subbands. The TQWT sub-bands are then analyzed for various spectral properties. Statistical
methods and graphical analysis are used to determine whether spectral properties are suitable for the TQWT domain.
The epileptic seizure classification is then done using bagging. The effectiveness of bagging in the detection strategy has
been investigaed. An accuracy of 98.4% has been achieved for the three-class classification.
Hassan et al.
33,34
introduced a Complete Ensemble Empirical Mode Decomposition with Adaptive Noise
(CEEMDAN) calculation to allow useful feature extraction from physiological signals. The CEEMDAN solves the prob-
lem of mode mixing and improves the mode spectrum separation. The CEEMDAN has been used to develop an auto-
matic epileptic seizure detection method to illustrate the feasibility of the CEEMDAN-based features. Various statistical
features are retrieved from the decomposed EEG signal segments by CEEMDAN, and seizure classification is performed
using an artificial neural network. The authors also presented an automated technique for detecting epileptic seizures
based on CEEMDAN. CEEMDAN has been used to break down segments of EEG signals into Intrinsic Mode Functions
(IMFs). Normal Inverse Gaussian (NIG) probability density function parameters were used to model the IMFs. Intui-
tive, graphical, and statistical assessments illustrate the usefulness of the NIG parameters in the CEEMDAN domain.
To perform classification, adaptive boosting, an eminent ensemble learning-based classification model, has been used.
This method has been evaluated on Bonn University dataset. An average accuracy of 97.6% has been obtained for three-
class classification.
35
Sharma et al.
36
provided features based on 2D PSRs of IMFs of EEG signals for automatic epileptic seizure classifica-
tion. For the creation of PSRs of IMFs, fixed values of time lag and embedding dimension have been suggested. For dif-
ferent window sizes, the 2D PSR images of IMFs have been developed as the input feature set for the LS-SVM classifier
for the identification of ictal and normal classes of EEG signals. The performance of various kernels, such as the radial
basis function, Mexican hat wavelet kernel, and Morlet wavelet kernel, was evaluated in a combination with the LS-
SVM classifier for the classification of ictal and normal EEG signals. A maximum accuracy of 98.67% has been obtained
using the Morlet wavelet kernel. Shankar et al.
37
presented the PSR technique to generate 2D input images from EEG
signals for specific brain rhythms. A CNN model with five convolutional layers and three fully-connected layers has
been used. Both Bonn University and CHB-MIT datasets were considered for experimental validation. Accuracies of
93% and 85% have been obtained on Bonn University and CHB-MIT datasets, respectively.
The proposed work in this paper is mainly concerned with EEG signal classification with simple and highly efficient
models based on signal transformations such as spectrogram and PSR transformations.
4of30 IBRAHIM ET AL.
3|THE PROPOSED PATIENT-SPECIFIC APPROACH
Figure 1depicts the main architecture of the proposed approach for seizure detection and prediction. Firstly, the pro-
cess of spectrogram estimation is applied on the EEG signal segments to transform them into image-like format. It is
well known that spectrogram gives a good time-frequency representation of the signals. The spectrogram estimation
produces images from which deep features can be extracted using a CNN. The CNN is responsible for feature extrac-
tion, and hence classification. Different layers including convolutional, pooling, and depth concatenation layers are
used for the process of feature extraction. Finally, the extracted deep features are used for classification to obtain the
detection and prediction results.
3.1 |EEG signals
The proposed models in this approach are evaluated on the publicly available CHB-MIT scalp EEG epilepsy dataset,
which was collected at Children's Hospital Boston (CHB) in collaboration with the Massachusetts Institute of Technol-
ogy (MIT).
38
The data included in this dataset was recorded for 23 epilepsy patients (7 males, and 17 females). Patients
were prevented from taking the anti-seizure medications, and then EEG signals were recorded for a couple of days after
prevention. This dataset was used for analyzing seizures of patients to determine their need for surgery. To record this
dataset, the 10-20 international systems of an EEG electrode montage scheme were used. The sampling frequency of
the signals for all patients is 256 Hz with a 16-bit resolution.
39
In this work, the proposed patient-specific approach is
applied on eight patients that were chosen randomly. The selected patients are chb01, 02, 03, 04, 06, 08, 11, and 14. The
objective is to introduce a proof of concept. Table 1illustrates the gender, age, number of hours, and number of seizures
for each patient.
FIGURE 1 Block diagram for the patient-specific approach
IBRAHIM ET AL.5of30
3.2 |Spectrogram estimation
A spectrogram is a time-frequency representation of time-domain EEG signals. The signal is windowed, and the spec-
trum is estimated for each window. The windowed spectra are used to build the spectrogram as an image. The color in
this image represents the power distributed over the time-frequency axes of the signal.
40
Spectrogram images can be fed
to the CNN as input. Figure 2illustrates the spectrograms of different EEG signal activities. There is a difference in
color between normal, pre-ictal, and ictal activities.
3.3 |CNN-based feature extraction
The CNN depends on relevant filters to extract the temporal and spatial dependencies in an image. The CNNs contain
several layers such as convolutional, pooling, residual learning, and softmax layers. The main essential roles of different
CNN layers are described briefly as follows.
3.3.1 | Convolutional layer
This layer has several learnable filters to compute dot products between the corresponding entries of both the filter and
the input image. The output feature map fC,l
x,y,kfor a particular layer land the input fOp,l1
x,ycan be computed as
39,41
:
fC,l
x,y,k¼wlT
kfOp,l1
x,yþbl
kð1Þ
TABLE 1 Dataset description
Patient No. No. of seizures No. of hours Gender Age
1 7 40.55 F 11
2 3 35.16 M 11
3 7 36 F 14
4 4 150.7 M 22
55 39F7
6 10 68.24 F 1.5
7 3 67.05 F 15.5
8 5 20 M 3.5
9 4 65.02 F 10
10 7 50.02 M 3
11 3 34.62 F 12
12 40 23.67 F 2
13 12 32 F 3
14 8 25.85 F 9
15 40 39.42 M 16
16 10 19 F 7
17 3 22 F 12
18 6 35.63 F 18
19 3 29.93 F 19
20 8 27.59 F 6
21 4 31.81 F 13
22 3 32 F 9
23 7 25.73 F 6
24 16 12 M 12.5
6of30 IBRAHIM ET AL.
where wl
krepresents the shared weights, bl
krepresents the bias and Cdenotes convolution. O
p
represents the input
image, for l=1, while it represents convolution, pooling, or activation, for l>1.
3.3.2 | Maximum pooling layer
Maximum pooling is performed by computing the maximum value in a local spatial neighborhood, and then reducing
the spatial resolution
39,41
:
fP,l
x,y,k¼max
m,nðÞNx,y
fOp,l1
m,n,kð2Þ
where the pooling operation is denoted by Pand the local spatial neighborhood of (x,y) coordinates is denoted by N
x,y
.
3.3.3 | Residual learning
This strategy is used to optimize the loss of CNNs, efficiently. The output of a residual learning block Rcan be expressed
as
39,41
:
fR,l
x,y,k¼fOp,lq
x,yþFf
Op,lq
x,y,wk
 ð3Þ
where fOp,lq
x,yis the input feature map, F(.) is the residual mapping to be learned, and qis the total number of stacked
layers.
FIGURE 2 Various spectrogram images: (A) Normal case, (B) Pre-ictal case, and (C) Ictal case
IBRAHIM ET AL.7of30
3.3.4 | Fully-connected layer
This layer is mainly used for classification purposes. Let layer (l-1) be a fully-connected layer. Layer lexpects ml1ðÞ
1fea-
ture maps having a size of ml1ðÞ
2ml1ðÞ
3as input. The following equation can be used to compute the i
th
unit in layer
l
39,41
:
YlðÞ
i¼fZ
lðÞ
i

,ZlðÞ
i¼X
ml1ðÞ
1
j¼1X
ml1ðÞ
2
r¼1X
ml1ðÞ
3
s¼1
Wl
i,j,r,sYl1
j

r,sð4Þ
where Wl
i,j,r,srepresents the weight used to connect the unit at position (r,s) in the i
th
unit in layer land the j
th
feature
map of layer (l1).
3.3.5 | Softmax layer
This layer is used for computing the loss. The form of softmax loss is given as
39,41
:
Lsoftmax ¼X
N
i¼1
log ewT
yifiþbyi
PK
j¼1ewT
jfiþbjð5Þ
where f
i
denotes features and y
i
is the true class label of the i
th
image. w
j
and b
j
are the weights and bias of the j
th
class,
respectively. Nis the number of training samples and Kis the number of classes.
In this work, we propose two CNNs that are different in the number of deep layers. The first model, namely pro-
posed CNN_1, adopts residual learning and depth concatenation strategies. It includes 13 deep layers and 4 residual
learning blocks. Figure 3presents the main architecture of the proposed CNN_1, which consists of two main residual
learning blocks; ResBL_1 and ResBL_3, with ResBL_2 and ResBL_4 included in ResBL_1 and ResBL_3, respectively.
Moreover, Figure 4shows the structure of ResBL_1, ResBL_3, ResBL_2, and ResBL_4. In the prementioned figures,
conv_1 refers to the first convolutional layer; conv_2 refers to the second convolutional layer, and so on. Each con-
volutional layer is followed by a Rectified Linear Unit (ReLU) activation function. Residual learning blocks allow the
flow of data from initial layers to last layers. Furthermore, these blocks are used to optimize the loss of CNNs through
FIGURE 3 The main architecture of the proposed CNN_1
8of30 IBRAHIM ET AL.
the concatenation of the extracted features from various convolutional layers. Depth concatenation is an essential com-
ponent in the residual learning block. Depth concatenation is used to increase the depth of the feature map. Addition-
ally, Table 2presents the filter specifications of the proposed CNN_1. On the other hand, the second model, namely
proposed CNN_2, is a very simple model which consists of 3 layers. Figure 5shows the main architecture of the pro-
posed CNN_2 that includes 3 convolutional blocks: convolutional block_1, convolutional block_2, and convolutional
block_3. Each convolutional block consists of a convolutional layer, a batch normalization layer, and a ReLU function.
Table 3introduces the filter specifications of the proposed CNN_2 model.
3.4 |Experimental results
In this paper, three different classes were examined for classifying the segments as ictal, normal, or pre-ictal states. The
performance of the proposed spectrogram-based DL approach was tested using standard metrics such as classification
Recall (REC), Specificity (SPE), Accuracy (ACC), Precision (PRE), and F
score
for all binary classes, as shown below.
42
Recall is given by:
REC ¼Tp
TpþFn
100 ð6Þ
Specificity is given by:
SPE ¼Tn
TnþFp
100 ð7Þ
Accuracy is given by:
(A) (B)
(C) (D)
FIGURE 4 Main architecture of the proposed (A) ResBL_1 (B) ResBL_3, (C) ResBL_2 and (D) ResBL_4
IBRAHIM ET AL.9of30
ACC ¼TpþTn
TpþTnþFpþFn
100 ð8Þ
Precision is given as:
PRE ¼Tp
TpþFp
100 ð9Þ
Fscore is given by:
Fscore ¼Tp
Tpþ1
2FpþFn

100 ð10Þ
TABLE 2 Filters Specifications of the Proposed CNN_1
Layer name No. of filters Filter sze
Input layer of size [227,227 3]
Conv_1 192 3 3
ReLU
Conv_2 64 1 1
ReLU
Conv_3 128 3 3
ReLU
Conv_4 64 3 3
ReLU
Conv_5 128 3 3
ReLU
Concatenation of the output of convolutional layers 3 and 5
Conv_6 192 3 3
ReLU
Concatenation of the output of convolutional layers 1 and 6
Conv_7 192 3 3
ReLU
Conv_8 64 1 1
ReLU
Conv_9 128 3 3
ReLU
Conv_10 64 3 3
ReLU
Conv_11 128 3 3
ReLU
Concatenation of the output of convolutional layers 9 and 11
Conv_12 192 3 3
ReLU
Concatenation of the output of convolutional layers 7 and 12
Conv_13 192 3 3
ReLU
Max. Pooling 1 7 7
Fully connected layer
Softmax layer
Output layer
10 of 30 IBRAHIM ET AL.
FIGURE 5 The main architecture of the proposed CNN_2
TABLE 3 Filters Specifications of the Proposed CNN_2
No. Name Type Activations
1 Image input
224 224 3 images
Image Input 224 224 3
2 Convolutional_1
16 3 33 convolutions with stride [1 1] and padding [1 1 1 1]
Convolution 224 224 16
3 Batchnorm_1
Batch normalization with 16 channels
Batch 224 224 16
4 Relu_1 ReLU 224 224 16
5 Maxpool_1
22 max pooling with stride [2 2] and padding [0 0 0 0]
Max pooling 112 112 16
6 Convolutional_2
32 3 316 convolutions with a stride of [1 1] and a padding of [1 1 1 1]
Convolution 112 112 32
7 Batchnorm_1
Batch normalization with 32 channels
Batch 112 112 32
8 Relu_2 ReLU 112 112 32
9 Maxpool_2
22 max pooling with a stride of [2 2] and a padding of [0 0 0 0]
Max pooling 56 56 32
10 Convolutional_3
64 3 332 convolutions with a stride of [1 1] and a padding of [1 1 1 1]
Convolution 56 56 64
11 Batchnorm_1
Batch normalization with 64 channels
Batch 56 56 64
12 Relu_3 ReLU 56 56 64
13 Fc Fully connected
14 softmax Softmax
15 classoutput Output
IBRAHIM ET AL.11 of 30
T
p
denotes the number of images that are truly epileptic and forecasted as epileptic, whereas F
p
denotes the
number of images that are nonepileptic and forecasted as being epileptic. T
n
stands for really nonepileptic clas-
ses predicted to be nonepileptic, whereas F
n
stands for actually epileptic classes predicted to be nonepileptic.
The proposed strategy mainly depends on a CNN to extract deep features from spectrogram images. In this sec-
tion, the experimental results of the proposed approach using various CNNs are provided. The proposed CNNs are
compared with state-of-the-art CNNs such as; VGG-19,
43
ResNet-101,
44
Inception-v3,
45
GoogLeNet,
46
DenseNet-
201,
47
and DarkNet-53.
48
In addition, the performance of the proposed approach is compared with that of another
EEG seizure detection and prediction method for Zhou.
28
He designed a CNN with no more than three layers. The
comparison is held in terms of accuracy, specificity, precision, sensitivity, and F score. The CHB-MIT dataset sig-
nals are divided into three classes: normal, pre-ictal, and ictal. Each class contains 1000 spectrogram images,
which are divided into 700 images for training and 300 images for testing. Experiments are carried out on 8
patients. Moreover, we perform two classification scenarios for each patient: normal versus pre-ictal and normal
versus ictal.
Our target is to reach the optimal performance of the proposed CNNs. To achieve that objective, we have to select
the OAs to be used and adjust the values of various hyperparameters such as weight decay, momentum value, mini-
batch size, maximum number of epochs, and Learning Rates (LRs). We apply L
2
regularization with the weight decay
set to 5 10
4
. In addition, the momentum value is set to 0.9, the mini-batch size is set to 32, and the maximum num-
ber of epochs is adjusted to 5. Tables 4and 5show the performance of the proposed CNN_1 and the proposed CNN_2,
respectively. In Tables 4and 5, three different OAs are tested: Adaptive Moment (Adam), Root Mean Square propaga-
tion (RMS prop), and Stochastic Gradient Descent with Momentum (SGDM). Furthermore, for each algorithm, three
LRs are tested: 0.1, 0.01, and 0.001. We consider the data of patient 1 and the classification scenario of normal versus
pre-ictal.
From Tables 4and 5, it is clear that the optimal performance of the proposed CNN_1 could be achieved by using the
SGDM algorithm for optimization and adjusting the LR to 0.001. Using Rmsprop algorithm for optimization and adjusting
the LR to 0.001 give promising results for CNN_2. It is also clear that the proposed CNN_1 is superior to the proposed
CNN_2, and this is expected due to the utilization of a larger number of deep layers and the application of residual learning
and depth concatenation strategies. On the other hand, the proposed CNN_2 is simpler and less time-consuming than the
proposed CNN_1. For more clarification, Figure 6depicts a graphical comparison between the proposed CNNs.
As mentioned before, experiments have been carried out on 8 patient signals, and we performed two classification
scenarios for each patient: normal versus pre-ictal and normal versus ictal. Tables 613 provide the classification results
of various state-of-the-art CNNs for patients 1, 2, 3, 4, 6, 8, 11, and 14, respectively.
From Tables 613, we can notice that the proposed CNN_1 is superior to the other state-of-the-art CNNs. The pro-
posed model achieves the highest results for all patients except patient 4. The rationale behind these findings is the
hyper-parameter optimization process involved in the proposed model. Finally, to further emphasize the superiority of
the proposed approach, a comparison between the proposed CNN_1 and Zhou's CNN
28
is presented in Table 14. The
comparison between CNNs is performed in terms of Accuracy, Specificity, and Recall. This comparison confirms the
superiority of the proposed model.
From all the previous results, we can say that the proposed approach succeeds in giving higher classification rates than
those of the state-of-the-art methods. These results also ensure that the spectrogram representation of EEG segments reflects
TABLE 4 Performance of the proposed CNN_1 for different optimization algorithms and learning rates
OA LR Accuracy Specificity Recall Precision F
score
SGDM 0.1 0.9573 0.9586 0.9432 0.9469 0.945
0.01 0.9697 0.9713 0.9529 0.9618 0.9573
0.001 0.9788 0.9798 0.9626 0.9721 0.9673
RMSprop 0.1 0.9548 0.956 0.9457 0.9524 0.949
0.01 0.9558 0.9567 0.9437 0.9517 0.9476
0.001 0.9536 0.9542 0.9429 0.9476 0.9452
Adam 0.1 0.9655 0.9667 0.9445 0.9624 0.9533
0.01 0.9682 0.9695 0.9515 0.9637 0.9575
0.001 0.9673 0.9685 0.9523 0.9632 0.9577
12 of 30 IBRAHIM ET AL.
all activities in these segment in a time, frequency and power representation that can be used as a visual pattern for signal clas-
sification. Furthermore, these high percentage results ensure the ability to develop EEG signal classification approaches based
on spectrogram estimation and CNNs.
4|THE PROPOSED PATIENT NON-SPECIFIC APPROACH
Figure 7shows an overview of the proposed patient non-specific approach, in which raw EEG signals are normalized
using a Z-score. The PSR technique is then used to produce 2D images for non-overlapping segments. The PSR images
are assessed using entropy and global statistics. Then, the images are fed into the CNN_3 model to classify epileptic sei-
zures by extracting features, automatically. Signal preprocessing, input image synthesis, CNN classification, and input
image quality evaluation are the primary phases, which are detailed as follows:
4.1 |Signal preprocessing
First, the EEG signals are preprocessed with a low-pass Finite Impulse Response (FIR) filter with 0.3 and 40 Hz cut-off frequen-
cies to remove noise and artifacts. The Z-score is then used to remove the DC components and normalize the EEG data. Feature
scaling is essential in ML, since it ensures that all feature sets are assessed on the same scale. Statistics have brought to light the
notion of feature scaling. Statistical analysis is a method for putting different variablesonthesamescale.Itisfrequentlyused,
when the dataset has various scales. The features of a dataset may have extensive and substantial variances across their ranges
at different times. So, in this scenario, standardization is imposed on the dataset to get all on the same scale. Z-score can be cal-
culated as
49
:
TABLE 5 Performance of the proposed CNN_2 for different optimization algorithms and learning rates
OA LR Accuracy Specificity Recall Precision F
score
SGDM 0.1 0.88 0.91 0.85 0.9043 0.8763
0.01 0.85 0.95 0.75 0.9375 0.8333
0.001 0.8833 0.90 0.8667 0.8966 0.8814
RMSprop 0.1 0.7583 0.9333 0.5833 0.8974 0.7071
0.01 0.7383 0.8333 0.50 0.8498 0.6565
0.001 0.8883 0.9767 0.9433 0.9554 0.8942
Adam 0.1 0.865 0.91 0.82 0.9011 0.8586
0.01 0.8717 0.9067 0.8367 0.8996 0.867
0.001 0.8683 0.96 0.7767 0.951 0.855
FIGURE 6 Graphical comparison between the proposed CNN models
IBRAHIM ET AL.13 of 30
Z¼Xμ
σ
ð11Þ
where X,μ, and σrepresent the input data, the mean, and the standard deviation, respectively. Figure 8shows EEG sig-
nals and their Z-score for normal, ictal, and pre-ictal activities.
TABLE 6 Experimental results for patient 1
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.9035 0.9044 0.8848 0.8961 0.8904
ResNet-101 0.9383 0.9399 0.9196 0.9286 0.924
Inception-v3 0.9407 0.9421 0.9216 0.9322 0.9268
GoogLeNet 0.9183 0.9198 0.8911 0.9016 0.8963
DenseNet-201 0.9507 0.9518 0.9281 0.9077 0.9177
DarkNet-53 0.9234 0.9246 0.9009 0.8852 0.8929
Proposed CNN_1 0.9788 0.9798 0.9626 0.9721 0.9673
Proposed CNN_2 0.8883 0.9767 0.9433 0.9554 0.8942
Normal
versus
Ictal
VGG-19 0.8152 0.8165 0.796 0.806 0.8009
ResNet-101 0.8673 0.8683 0.8476 0.8586 0.853
Inception-v3 0.8847 0.8855 0.8657 0.8764 0.871
GoogLeNet 0.8209 0.8222 0.79 0.8067 0.7983
DenseNet-201 0.8632 0.8643 0.8367 0.84 0.8383
DarkNet-53 0.8266 0.8271 0.7992 0.82 0.8095
Proposed CNN_1 0.9049 0.9061 0.8842 0.8918 0.8879
Proposed CNN_2 0.8415 0.8586 0.8321 0.8542 0.843
TABLE 7 Experimental results for patient 2
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.964 0.9655 0.9482 0.9603 0.9542
ResNet-101 0.9751 0.9768 0.959 0.9698 0.9643
Inception-v3 0.9777 0.9785 0.9612 0.9733 0.9672
GoogLeNet 0.9507 0.9514 0.9243 0.9478 0.9359
DenseNet-201 0.9728 0.9744 0.9524 0.9636 0.9579
DarkNet-53 0.9465 0.9482 0.9061 0.9273 0.9165
Proposed CNN_1 0.9954 0.9967 0.9758 0.9877 0.9817
Proposed CNN_2 0.9657 0.9457 0.9867 0.9433 0.965
Normal
versus
Ictal
VGG-19 0.9157 0.9171 0.8995 0.9093 0.9043
ResNet-101 0.9506 0.9521 0.9346 0.9464 0.9404
Inception-v3 0.9509 0.9517 0.9351 0.9473 0.9411
GoogLeNet 0.9393 0.9417 0.9118 0.9272 0.9194
DenseNet-201 0.9138 0.9148 0.8847 0.903 0.8938
DarkNet-53 0.952 0.9534 0.9253 0.9395 0.9323
Proposed CNN_1 0.9756 0.977 0.9535 0.9626 0.958
Proposed CNN_2 0.9085 0.907 0.91 0.9067 0.9083
14 of 30 IBRAHIM ET AL.
4.2 |Phase Space Reconstruction (PSR)
The PSR of a signal is a visual representation of the signal dynamic behavior as it changes over time. Phase space may
be used to rebuild a univariate time series in a chaotic system. In a dynamic system, the univariate time series contains
all information about variables. Each point in the phase space represents a state of the system, while the phase space
TABLE 8 Experimental results for patient 3
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.6757 0.6769 0.6507 0.6629 0.6567
ResNet-101 0.7008 0.7019 0.6759 0.6891 0.6824
Inception-v3 0.7253 0.7265 0.7011 0.7131 0.707
GoogLeNet 0.7122 0.7142 0.7031 0.7114 0.7072
DenseNet-201 0.7341 0.7354 0.7293 0.739 0.7341
DarkNet-53 0.7162 0.7177 0.7007 0.7047 0.7027
Proposed CNN_1 0.7794 0.7805 0.7635 0.7774 0.7703
Proposed CNN_2 0.6456 0.6445 0.6467 0.6433 0.645
Normal
versus
Ictal
VGG-19 0.7624 0.7641 0.7382 0.7507 0.7443
ResNet-101 0.7893 0.7908 0.7647 0.7769 0.7707
Inception-v3 0.8246 0.8264 0.8002 0.8124 0.8062
GoogLeNet 0.8091 0.8153 0.8111 0.8072 0.8091
DenseNet-201 0.8128 0.8144 0.8026 0.8034 0.8028
DarkNet-53 0.8003 0.8062 0.8081 0.8028 0.8054
Proposed CNN_1 0.8416 0.8427 0.8248 0.835 0.8298
Proposed CNN_2 0.7981 0.7575 0.8433 0.73 0.7867
TABLE 9 Experimental results for patient 4
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.6949 0.6962 0.6685 0.6824 0.6753
ResNet-101 0.7331 0.7344 0.7054 0.7181 0.7116
Inception-v3 0.7564 0.7576 0.7289 0.7436 0.7361
GoogLeNet 0.7683 0.943 0.9675 0.9693 0.9683
DenseNet-201 0.781 0.7892 0.7503 0.7719 0.7609
DarkNet-53 0.7372 0.7425 0.7329 0.7417 0.7373
Proposed CNN_1 0.7931 0.7942 0.7791 0.783 0.781
Proposed CNN_2 0.7143 0.7119 0.7167 0.71 0.7133
Normal
versus
Ictal
VGG-19 0.7116 0.7128 0.6847 0.6978 0.6911
ResNet-101 0.7451 0.7465 0.7174 0.7301 0.7236
Inception-v3 0.7637 0.7649 0.7366 0.7493 0.7428
GoogLeNet 0.7223 0.7365 0.7235 0.7212 0.7223
DenseNet-201 0.804 0.8236 0.8067 0.8014 0.804
DarkNet-53 0.7761 0.7857 0.7412 0.7311 0.7361
Proposed CNN_1 0.7988 0.7997 0.774 0.7812 0.7775
Proposed CNN_2 0.7319 0.7406 0.7233 0.7467 0.735
IBRAHIM ET AL.15 of 30
trajectory reflects the system temporal progress under various initial conditions.
50
A phase space can be generated from
a one-dimensional time series using the taken time-delay embedding theorem.
50
This theorem may be used to analyze
chaotic time series. If a scalar time series T
t
=N
1
,N
2
,N
3
,…….N
n
from a chaotic system is given, then reconstruction is
possible in terms of the phase space vectors X(t) expressed as X(t)=[x(t),x(t+τ),……x(t+[m1]τ)], t=1,2,.M;
M=N(m1)τ. Here, τis the time delay, mis the embedding dimension of the PSR, and Mis the number of phase
TABLE 10 Experimental results for patient 6
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.5414 0.5429 0.5205 0.5307 0.5255
ResNet-101 0.5574 0.5584 0.5374 0.5478 0.5425
Inception-v3 0.5769 0.5776 0.5583 0.5681 0.5631
GoogLeNet 0.6285 0.6404 0.6229 0.6343 0.6285
DenseNet-201 0.6133 0.6252 0.6072 0.6195 0.6133
DarkNet-53 0.6204 0.6323 0.6148 0.6262 0.6204
Proposed CNN_1 0.6377 0.6389 0.6125 0.6224 0.6174
Proposed CNN_2 0.5364 0.5437 0.4567 0.6167 0.4967
Normal
versus
Ictal
VGG-19 0.648 0.6499 0.6287 0.6385 0.6335
ResNet-101 0.6896 0.6908 0.6688 0.6788 0.6737
Inception-v3 0.7114 0.7125 0.6909 0.6997 0.6952
GoogLeNet 0.7395 0.7436 0.7338 0.7454 0.7395
DenseNet-201 0.7283 0.7349 0.7221 0.7346 0.7283
DarkNet-53 0.733 0.7455 0.7279 0.7383 0.733
Proposed CNN_1 0.7636 0.7649 0.7467 0.7553 0.7509
Proposed CNN_2 0.7459 0.6795 0.8267 0.61 0.7183
TABLE 11 Experimental results for patient 8
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.7691 0.7704 0.7505 0.7632 0.7567
ResNet-101 0.8108 0.8117 0.792 0.8025 0.7972
Inception-v3 0.8446 0.8459 0.8254 0.8369 0.8311
GoogLeNet 0.8416 0.8542 0.8356 0.8478 0.8416
DenseNet-201 0.8469 0.8554 0.8414 0.8525 0.8469
DarkNet-53 0.8491 0.8565 0.8437 0.8547 0.8491
Proposed CNN_1 0.9033 0.9046 0.8776 0.8867 0.8821
Proposed CNN_2 0.7483 0.769 0.71 0.7867 0.7383
Normal
versus
Ictal
VGG-19 0.7107 0.7123 0.6914 0.7027 0.697
ResNet-101 0.7526 0.7541 0.7342 0.7458 0.7399
Inception-v3 0.7937 0.7952 0.7748 0.7859 0.7803
GoogLeNet 0.8428 0.8523 0.8374 0.8484 0.8428
DenseNet-201 0.8396 0.8448 0.8338 0.8456 0.8396
DarkNet-53 0.8446 0.8557 0.8386 0.8508 0.8446
Proposed CNN_1 0.8773 0.8782 0.8515 0.8628 0.8571
Proposed CNN_2 0.8126 0.7701 0.86 0.7433 0.8017
16 of 30 IBRAHIM ET AL.
points of the reconstructed phase space. Computation of τand mvalues is very essential in PSR. To estimate the value
of τfor a nonlinear time series, the Average Mutual Information (AMI) method can be used.
51
The equation used to cal-
culate the AMI is given as follows:
IτðÞ¼
X
Nτ
t¼1
PX
t,Xtþτ
ðÞ:log PX
t,Xtþτ
ðÞ
PX
t
ðÞ:PX
tþτ
ðÞ
 ð12Þ
TABLE 13 Experimental results for patient 14
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.6542 0.6559 0.6399 0.6508 0.6453
ResNet-101 0.704 0.7057 0.6913 0.6978 0.6945
Inception-v3 0.7417 0.7432 0.7288 0.7322 0.7304
GoogLeNet 0.6427 0.6538 0.6367 0.6488 0.6427
DenseNet-201 0.6389 0.6497 0.6329 0.645 0.6389
DarkNet-53 0.6445 0.6545 0.6385 0.6507 0.6445
Proposed CNN_1 0.7683 0.7698 0.7362 0.7472 0.7416
Proposed CNN_2 0.6938 0.7057 0.6633 0.7233 0.6833
Normal
versus
Ictal
VGG-19 0.6438 0.6451 0.6291 0.6373 0.6331
ResNet-101 0.6804 0.6815 0.6657 0.6734 0.6695
Inception-v3 0.7163 0.7176 0.7022 0.7101 0.7061
GoogLeNet 0.6363 0.6351 0.6339 0.6389 0.6363
DenseNet-201 0.6282 0.6343 0.6254 0.6312 0.6282
DarkNet-53 0.6431 0.6374 0.6452 0.6412 0.6431
Proposed CNN_1 0.7462 0.7471 0.7134 0.7282 0.7207
Proposed CNN_2 0.7321 0.7081 0.8733 0.64 0.7167
TABLE 12 Experimental results for patient 11
Scenario CNN Accuracy Specificity Precision Recall F
score
Normal
versus
Pre-ictal
VGG-19 0.5136 0.5148 0.4903 0.5007 0.4954
ResNet-101 0.5452 0.5463 0.519 0.5285 0.5237
Inception-v3 0.5723 0.5739 0.5495 0.5594 0.5544
GoogLeNet 0.538 0.5494 0.5324 0.5438 0.538
DenseNet-201 0.5279 0.5381 0.5218 0.5341 0.5279
DarkNet-53 0.5325 0.5437 0.5267 0.5384 0.5325
Proposed CNN_1 0.6013 0.6024 0.5726 0.5845 0.5784
Proposed CNN_2 0.56 0.572 0.4767 0.6433 0.52
Normal
versus
Ictal
VGG-19 0.6469 0.6481 0.6235 0.6354 0.6293
ResNet-101 0.6755 0.677 0.6515 0.6614 0.6564
Inception-v3 0.7067 0.7084 0.6847 0.6946 0.6896
GoogLeNet 0.6418 0.6525 0.6361 0.6477 0.6418
DenseNet-201 0.6489 0.659 0.6429 0.6551 0.6489
DarkNet-53 0.6499 0.6599 0.6438 0.6562 0.6499
Proposed CNN_1 0.7332 0.7344 0.7111 0.7238 0.7173
Proposed CNN_2 0.7159 0.6319 0.91 0.47 0.69
IBRAHIM ET AL.17 of 30
where P(x
t
) is the probability density of X
t
.P(X
t
,X
t+τ
) is the joint probability density of X
t
and X
t+τ
.I(τ) is a measure
of the statistical dependence of the reconstruction variables. For a nonmonotonous decrease of I(τ), the location of the
first local minimum is considered as the suitable value of τ.
52
Kennel et al.
52
introduced a False Nearest Neighbor
(FNN) method to compute the optimum m. The FNN method calculates the distance D(m) to the nearest neighbor and
TABLE 14 Comparison between the proposed CNN_1 and Zhou's CNN
28
Patient ID
Accuracy Specificity Recall
Proposed CNN_1 Zhou's CNN
28
Proposed CNN_1 Zhou's CNN
28
Proposed CNN_1 Zhou's CNN
28
1 0.9788 0.905 0.9798 0.9 0.9721 0.91
2 0.9954 0.623 0.9967 0.607 0.9877 0.64
3 0.7794 0.623 0.7805 0.677 0.7774 0.57
4 0. 7931 0.515 0. 7942 0.573 0. 783 0.457
6 0.6377 0.53 0.6389 0.547 0.6224 0.513
8 0.9033 0.508 0.9046 0.547 0.8867 0.47
11 0.6013 0.548 0.6024 0.607 0.5845 0.49
14 0.7683 0.503 0.7698 0.513 0.7472 0.493
FIGURE 7 Block diagram of the proposed patient non-specific approach
18 of 30 IBRAHIM ET AL.
the distance D(m+1) between two points in m+1 dimensions for each point in the m-dimensional image. The neigh-
bors are said to be false, if the following condition is met
52
:
Dmþ1ðÞ
RA
>Rtol,Atol ð13Þ
whereR
tol
and A
tol
are constant thresholds, while R
A
is the time-series standard deviation. The procedure is continued
for the dimensions, and it is terminated when the proportion of FNN becomes zero or negligible.
4.3 |Evaluation of input quality
In this paper, the PSR images for all EEG segments are created. EEG signals are regarded as time series data with
nonlinear and nonstationary properties. The time series incorporates recurrent properties such as nonlinearity and cha-
otic behavior of the time series. The PSR is a good candidate for creating 2D input images. It also has the capability of
graphically analyzing hidden patterns, pattern correlations, predictive characteristics, stability, fundamental similarity,
and structural changes in time series data across time. The entropy and global statistical values of produced PSR images
are used to assess quality. Entropy (E
t
) is a statistical measure of randomness that can be exploited to depict the texture
of an image and to compare image information. E
t
is calculated for images as:
FIGURE 8 EEG signals and their Z-scores for (A) Normal EEG signals, (B) Ictal EEG signals, and (C) Pre-ictal EEG signals
IBRAHIM ET AL.19 of 30
FIGURE 9 The main architecture of the proposed CNN_3
FIGURE 10 Samples of (A) ictal, (B) pre-ictal and (C) normal PSR images for τ=1 and m=2
20 of 30 IBRAHIM ET AL.
Et¼X
N
i¼1
hilog hi
ðÞ ð14Þ
where h
i
is the count of the normalized histogram and Nis the number of levels in the image. Furthermore, image
global statistics such as mean, variance, and standard deviation are calculated and used.
4.4 |CNN-based feature extraction
A simple CNN model, namely proposed CNN_3 has been used to classify EEG signal segments. The 2D PSR images
serve as input to the CNN_3 model. Figure 9shows the proposed CNN_3 architecture for ictal, normal, and pre-ictal
activity classification based on PSR images. The CNN_3 model has five convolution blocks (each with a convolutional
TABLE 15 Experimental results of the two-class classification with the CNN_3 model from spectrogram images
Scenario CNN Accuracy Recall Specificity Precision Fscore Fpr Mcc Kappa
Normal
versus
Pre-ictal
VGG-19 0.8335 0.8761 0.8244 0.844 0.8604 0.12 0.844 0.8744
ResNet-101 0.893 0.8886 0.8765 0.8675 0.8863 0.057 0.89 0.8678
Inception-v3 0.8409 0.8342 0.8451 0.8276 0.8298 0.089 0.825 0.862
GoogLeNet 0.813 0.805 0.8104 0.7931 0.7923 0.175 0.798 0.7998
DenseNet-201 0.8506 0.8071 0.851 0.821 0.857 0.0856 0.834 0.8653
Inception-ResNet-v2 0.829 0.7872 0.8696 0.8001 0.793 0.134 0.729 0.8234
Proposed CNN _3 0.9128 0.911 0.9058 0.892 0.9256 0.0589 0.820 0.9034
Normal
versus
Ictal
VGG-19 0.87 0.856 0.878 0.8234 0.8409 0.154 0.842 0.876
ResNet-101 0.8973 0.8886 0.8983 0.8876 0.873 0.0899 0.797 0.9053
Inception-v3 0.8847 0.8764 0.8855 0.8657 0.871 0.098 0.797 0.8971
GoogLeNet 0.8909 0.8967 0.8822 0.89 0.8983 0.075 0.877 0.884
DenseNet-201 0.9032 0.94 0.943 0.9367 0.9383 0.063 0.8756 0.9150
Inception-ResNet-v2 0.9266 0.92 0.9271 0.8992 0.9095 0.0143 0.813 0.935
Proposed CNN_3 0.9249 0.9318 0.9061 0.9042 0.9079 0.0112 0.8126 0.0937
FIGURE 11 Confusion matrix and ROC curve for two-class classification with the proposed CNN_3 model from spectrogram images, 1
refers to pre-ictal class and 2 refers to normal class
IBRAHIM ET AL.21 of 30
layer, batch normalization, and a max-pooling layer) and three Fully-Connected (FC) layers. A 2 2 kernel is used to
perform convolution with the 128 128 input images. In the convolution and FC layers, the ReLU is exploited as an
activation function, but the final output layer employs a softmax function.
The ReLU activation function is used in the CNN_3 model, because it reduces the computing complexity. The
CNN_3 extracts features automatically by using multiple kernel functions to execute convolution at multiple levels.
53
The outputs of several convolution blocks produce smooth and detailed feature vectors from lower to higher levels of
input images, which helps in the classification and detection operations. Furthermore, batch normalization reduces the
FIGURE 12 Training progress for CNN_3 model for two-class classification from spectrogram images
TABLE 16 Experimental results for two-class classification with CNN_3 model from PSR images
Scenario CNN Accuracy Recall Specificity Precision Fscore Fpr Mcc Kappa
Normal
versus
Pre-ictal
VGG-19 0.964 0.9603 0.9655 0.9482 0.9542 0.013 0.82 0.956
ResNet-101 0.9751 0.9698 0.9768 0.959 0.9643 0.011 0.816 0.9612
Inception-v3 0.9777 0.9733 0.9785 0.9612 0.9672 0.008 0.807 0.9681
GoogLeNet 0.9507 0.9478 0.9514 0.9243 0.9359 0.0091 0.858 0.936
DenseNet-201 0.9728 0.9636 0.9744 0.9524 0.9579 0.006 0.817 0.959
Inception-ResNet-v2 0.9865 0.979 0.9873 0.9921 0.9865 0.0053 0.854 0.983
Proposed CNN _3 1 1 1 1 1 0 1 1
Normal
versus
Ictal
VGG-19 0.9757 0.9693 0.9871 0.9795 0.9743 0.007 0.867 0.9748
ResNet-101 0.9806 0.9864 0.9921 0.9846 0.9804 0.0034 0.856 0.9861
Inception-v3 0.9809 0.9873 0.9917 0.9751 0.9611 0.0067 0.952 0.9773
GoogLeNet 0.9793 0.9772 0.9717 0.9618 0.9694 0.0082 0.987 0.9718
DenseNet-201 0.9738 0.973 0.9848 0.9647 0.9538 0.00865 0.978 0.9898
Inception-ResNet-v2 0.992 0.9895 0.9934 0.9753 0.9723 0.002 0.912 0.9921
Proposed CNN_3 1 1 1 1 1 0 1 1
22 of 30 IBRAHIM ET AL.
internal covariance shift in the convolution layer output.
54
Additionally, to minimize complexity, spatial size, and fea-
ture variances, the max-pooling of size 2 2 is considered.
55
Overfitting has been addressed by setting a predetermined
dropout rate before each FC layer. Finally, in the output layer, the classification is performed by using the softmax acti-
vation function.
4.5 |Experimental results
In this approach, the segments are classified as ictal, normal, or pre-ictal states. The performance of the proposed
PSR-based DL approach has been investigated on the signals of all patients of the CHB-MIT dataset.
38
The proposed
PSR-based approach has been tested using the same metrics adopted with the patient-specific approach in addition to
Matthews Correlation Coefficient (MCC), False Positive Rate (F
pr
), and Kappa for all binary and ternary classes, as
shown below.
42
The MCC is defined as:
MCC ¼TpTnFpFn
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
TpþFp

TpþFn

TnþFp

TnþFn
ðÞ
q100 ð15Þ
False-Positive Rate is given by:
Fpr ¼Fp
TnþFp
ð16Þ
Kappa coefficient is defined as:
Kappa ¼2TpTnFnFp

TpþFp

FpþTn

TpþFn

FnþTn
ðÞ
100 ð17Þ
To create the PSR images for different segments, three essential parameters are determined: ϵ,τand m. First, the
value of τis calculated using the AMI function. The optimum value is chosen based on the point of rising of AMI at ini-
tial local minima. AMI values often decrease with an increase of τ. The optimum values for normal, ictal, and pre-ictal
FIGURE 13 Confusion matrix and ROC curve for the two-class classification with CNN_3 model on PSR images, 1 refers to pre-ictal class and 2
refers to normal class
IBRAHIM ET AL.23 of 30
states are 112, 125, and 115, respectively. The EEG signal points are mapped to a 2D space by setting mto 2. This
value has been considered for generating 2D representations of EEG signal segments for epileptic seizure detec-
tion.
36,56,57
The PSR images of EEG signals for the ictal state are more regular than those for the normal state, as seen
in Figure 10. Visual evaluation of the PSR images of the EEG signals reveals that the dispersion of points in the PSR
images of normal intervals is significantly greater than that of PSR images of ictal intervals. In other words, as com-
pared to normal states, the majority of data points in ictal states are around the coordinate center. Furthermore, the
deviation of data points from the center of the coordinate plane is greater in normal states than in ictal states. The diag-
onal plots represent the histograms of x(t) and x(t+1).
FIGURE 14 Training progress for CNN_3 model for two-class classification from PSR images
TABLE 17 Experimental results of three-class classification with with CNN_3 model from spectrogram and PSR images
Preprocessing CNN Accuracy Recall Specificity Precision Fscore Fpr Mcc Kappa
Spectrogram VGG-19 0.8904 0.8961 0.8848 0.9044 0.9035 0.173 0.821 0.856
ResNet-101 0.924 0.9286 0.9196 0.9399 0.9383 0.121 0.909 0.9112
Inception-v3 0.9068 0.9322 0.9216 0.9421 0.9407 0.089 0.9012 0.8981
GoogLeNet 0.8963 0.9016 0.8911 0.9198 0.9183 0.191 0.812 0.826
DenseNet-201 0.9077 0.9077 0.9281 0.9518 0.9507 0.016 0.872 0.909
Inception-ResNet-v2 0.9065 0.939 0.9373 0.9521 0.9565 0.0083 0.907 0.893
Proposed CNN_3 0.9121 0.9446 0.9524 0.9447 0.9446 0.0076 0.9370 0.9256
Phase Space Reconstruction VGG-19 0.9257 0.913 0.871 0.9195 0.8943 0.125 0.8562 0.9048
ResNet-101 0.9406 0.9164 0.9021 0.9146 0.9104 0.0134 0.883 0.9061
Inception-v3 0.9309 0.943 0.9317 0.9451 0.9311 0.0067 0.9352 0.9273
GoogLeNet 0.9193 0.9172 0.9017 0.9218 0.9394 0.0082 0.907 0.9118
DenseNet-201 0.9438 0.953 0.9348 0.9547 0.9238 0.00865 0.9056 0.9298
Inception-ResNet-v2 0.952 0.9695 0.9534 0.9453 0.9423 0.002 0.9134 0.921
Proposed CNN_3 0.9989 0.9989 0.9995 0.9989 0.9989 0.00005 0.9984 0.9975
24 of 30 IBRAHIM ET AL.
4.5.1 | Results of two-class classification
Six pre-trained models, namely VGG-19, ResNet-101, Inception-v3, GoogLeNet, DenseNet-201 and Inception-ResNet-v2
are used in this task. Moreover, the proposed CNN_3 trained from scratch is also used. Table 15 presents the detection
results obtained for different pre-trained models and the proposed CNN_3 based on spectrograms. It is clear that the
proposed CNN_3 outperforms the pre-trained models for normal-pre-ictal and normal-ictal classification with
FIGURE 15 Confusion matrix and ROC curve for three-class classification with the CNN_3 model from spectrogram images, 1 refers to
ictal class, 2 refers to pre-ictal class and 3 refers to normal class
FIGURE 16 Training progress of the CNN_3 model for three-class classification from spectrogram images
IBRAHIM ET AL.25 of 30
accuracies that reach 90.28% and 92.49%, respectively. The corresponding confusion matrix and ROC curve of the pro-
posed CNN_3 model on spectrogram images for per-ictal-normal classification are shown in Figure 11. Figure 12 dis-
plays the performance of the proposed CNN_3 model in terms of both accuracy and loss. There is a coincidence in
performance between validation and training accuracies as well as validation and training loss. The Minimal Square
Error (MSE) has been chosen as the loss function. Based on the MSE, a distance minimization approach is used.
FIGURE 17 Confusion matrix and ROC curve for three-class classification with CNN_3 model from PSR images, 1 refers to ictal class,
2 refers to pre-ictal class, and 3 refers to normal class
FIGURE 18 Training progress for the CNN_3 model for three-class classification from PSR images
26 of 30 IBRAHIM ET AL.
Table 16 presents the detection results obtained from different pre-trained and the proposed CNN_3 model from
PSR images. It is clear that the proposed CNN_3 model outperforms the pre-trained models with accuracies that reach
100% for normal-pre-ictal and normal-ictal classification. The corresponding confusion matrix and ROC curve of the
proposed CNN_3 model on PSR images for per-ictal-normal classification are shown in Figure 13. Figure 14 displays
the performance of the CNN_3 model in terms of both accuracy and loss. There is a coincidence in performance
between validation and training accuracies as well as validation and training loss.
4.5.2 | Results for three-class classification
Table 17 presents the detection results obtained from spectrogram and PSR images for three-class classification. It is
clear that the proposed CNN_3 model performs better on PSR images. Moreover, the proposed CNN_3 model outper-
forms the other pre-trained models on PSR images. It achieves an average accuracy of 99.89%. The corresponding
TABLE 18 Comparison of the proposed CNN_3 model with the state-of-the-art models
Authors Technique Dataset Classes Accuracy %
M. Talha et al.
14
CNN KK women and children's hospital Normal-ictal 93.3
Yang et al.
15
NLSTM Boon University Normal-ictal 98.44
CHB-MIT 97.47
Covert et al.
17
TGCN Boston children's hospital Normal-ictal 98.05
B. Bouaziz et al.
19
CNN CHB-MIT Normal-ictal 99.48
Rajaguru et al.
20
MAE +EM-PCA CHB-MIT Normal-ictal 93.78
Roy et al.
21
ChronoNet CHB-MIT Normal-ictal 86.57
Choi et al.
22
3D-CNN +GRU CHB-MIT Normal-ictal 89.4
SNUH 97
Truong et al.
23
Spectrograms +STFT Freiburg hospital intra-cranial EEG Normal-ictal 81.4
CHB-MIT 81.2
Subasi et al.
29
GA +PSO Boon University Normal-ictal 99.38
Alickovic et al.
8
MSPCA +DWT CHB-MIT Normal-ictal 100
Normal-ictal-pre-ictal 99.77
M. Zhou et al.
28
CNN CHB-MIT Ictal-pre-ictal
Normal-ictal
Normal-ictal-pre-ictal
95.6
97.5
93
Freiburg Ictal-pre-ictal
Normal-ictal
Normal-ictal-pre-ictal
96.7
95.4
92.3
Qaisar et al.
30
CNN Andrzejak Normal-ictal-pre-ictal 96.4
Hassan et al.
34
TQWT Boon University Normal-ictal-pre-ictal 98.4
Hassan et al.
35
CEEMDAN +NIG Boon University Normal-ictal-pre-ictal 97.6
Sharma et al.
36
2D PSRs +EMD +LS-SVM CHB-MIT Normal-ictal 98.67
Shankar et al.
37
PSR +CNN Boon University Normal-ictal 93
CHB-MIT Normal-ictal 85
Proposed CNN_3 model Spectrogram images CHB-MIT Normal-ictal
Normal-pre-ictal
Normal-ictal-pre-ictal
91.28
92.49
90.21
PSR images Normal-ictal
Normal-pre-ictal
Normal-ictal-pre-ictal
100
100
99.89
IBRAHIM ET AL.27 of 30
confusion matrices and ROC curves of the proposed CNN_3 model on spectrogram and PSR images are shown in Fig-
ures 15, 16 and 17. The performance of the CNN_3 model on spectrogram and PSR images in terms of both accuracy
and loss is shown in Figures 16 and 18, respectively. It is clear from Table 17 that the utilization of PSR images
improves the accuracy by 5.7%, recall by 5.8%, kappa with 7.77% compared to the results obtained with spectrogram
images. In addition, with PSR images the CNN_3 model achieves a reduction of the F
pr
by about 99% of the value
obtained on spectrogram images.
4.6 |Discussion and comparison with the-state-of-the-art methods
Asymptotically, chaotic signals like EEG signals may evolve towards lower-dimensional attractors. The PSR can be used
to visualize these attractors, indirectly. We can observe from Figure 10 that such lower-dimensional attractors exist for
EEG signals, especially for the ictal class. This was also shown in previous research,
36,37
where the PSR was utilized as
a starting point for manually extracting features, and subsequently classifying them. The choice of time delay and
embedding dimension are essential aspects in PSR and its chaotic features. If the range of τis too small, the attractor
will be condensed into the area of the coordinate system diagonal line. On the other hand, if the range of τis too large,
the data point trajectory will twist and fold, making obvious projection correlations difficult to create. The experimental
results reveal that working on spectrogram images gives good results, but the phase information is ignored which limits
the accuracy levels to 91.28%, 92.49%, and 90.21% for ictal-pre-ictal, normal-ictal, and normal-pre-ictal-ictal classifica-
tion, respectively. Hence, we thought to avoid this limitation, by working with PSR, which gives accuracy levels up to
100%, 100%, and 99.89% for ictal-preictal, normal-ictal, and normal-pre-ictal-ictal classification, respectively, due to the
kept phase information.
The CNN_3 takes 1125 s for the two-class classification and 2015 s for the three-class classification from PSR
images.
The accuracy level with the CNN_3 model reaches 100%, which is higher than the levels of the traditional methods
given in Table 18, for two-class classification. Moreover, an accuracy level of 99.89% has been obtained for the three-
class classification. These findings confirm the efficiency of the CNN_3 model on PSR images for the classification of
EEG signals.
5|CONCLUSIONS
This paper dealt with the problem of EEG signal classification for two purposes: seizure detection and seizure predic-
tion. Two patient-specific CNN models were presented for seizure detection and seizure prediction. They work on spec-
trogram images of EEG signal segments. The third CNN model is patient non-specific. It has the ability to classify two
and three states of EEG signals. Moreover, it is valid for operation on spectrogram as well as PSR images of EEG signal
segments. Experimental results revealed the best classification performance with the patient non-specific CNN model
from PSR images of EEG signal segments. This is attributed to the elegant EEG signal representation with PSR. On the
other hand, the first two models are acceptable for patient-specific applications, but the dependence on spectrogram
images limits their performance to some extent. In future work, the proposed models will be tested on different datasets
and different signal representations.
DATA AVAILABILITY STATEMENT
The data that support the findings of this study are openly available in CHB-MIT dataset at https://archive.physionet.
org/pn6/chbmit/, reference number 56.
ORCID
Walid El-Shafai https://orcid.org/0000-0001-7509-2120
Mohamed Rihan https://orcid.org/0000-0003-4030-2559
REFERENCES
1. WHO, Programmes and Projects.http://www.who.int/mediacentre/factsheets/fs999/en/, 2017.
2. Thurman DJ, Logroscino G, Beghi E, et al. The burden of premature mortality of epilepsy in high-income countries: a systematic review
from the mortality task force of the international league against epilepsy. Epilepsia. 2017;58(1):17-26.
28 of 30 IBRAHIM ET AL.
3. Shoeibi A, Khodatars M, Ghassemi N, et al. Epileptic seizures detection using deep learning techniques: a review. Int J Environ Res
Public Health. 2021;18(11):5780.
4. Blinowska KJ, Zygierewicz J. Practical Biomedical Signal Analysis Using MATLAB R. CRC Press; 2011.
5. Acharya UR, Hagiwara Y, Adeli H. Automated seizure prediction. Epilepsy Behav. 2018;88:251-261.
6. Sharma M, Bhurane AA, Acharya UR. Mmsfl-owfb: a novel class of orthogonal wavelet filters for epileptic seizure detection. Knowledge-
Based Sys. 2018;160:265-277.
7. Acharya UR, Hagiwara Y, Deshpande SN, et al. Characterization of focal eeg signals: a review. Future Generation Comput Sys. 2019;91:
290-299.
8. Alickovic E, Kevric J, Subasi A. Performance evaluation of empirical mode decomposition, discrete wavelet transform, and wavelet
packed decomposition for automated epileptic seizure detection and prediction. Biomed Signal Process Control. 2018;39:94-102.
9. Emara HM, Elwekeil M, Taha TE, et al. Hilbert transform and statistical analysis for channel selection and epileptic seizure prediction.
Wireless Personal Commun. 2021;116(4):3371-3395.
10. Emara HM, Shoaib MR, Elwekeil M, et al. Deep convolutional neural networks for covid-19 automatic diagnosis. Microsc Res Tech.
2021;84:2504-2516.
11. Emara HM, Elwekeil M, Taha TE, et al. Efficient frameworks for EEG epileptic seizure detection and prediction. Annals of Data Science.
Springer; 2021:1-36.
12. Hossain MS, Amin SU, Alsulaiman M, Muhammad G. Applying deep learning for epilepsy seizure detection and brain mapping
visualization. ACM Transac Multimedia Comput Commun Appl (TOMM). 2019;15(1s):1-17.
13. Parvez MZ, Paul M, Antolovich M. Detection of pre-stage of epileptic seizure by exploiting temporal correlation of emd decomposed
EEG signals. J Med Bioeng. 2015;4(2):110-116.
14. Avcu MT, Zhang Z, Chan DWS. Seizure detection using least eeg channels by deep convolutional neural network. ICASSP 20192019
IEEE IntConf Acoustics Speech Signal Process (ICASSP). 2019;7:1120-1124.
15. Li Y, Yu Z, Chen Y, et al. Automatic seizure detection using fully convolutional nested lstm. Int J Neural Syst. 2020;30(04):2050019.
16. Asif U, Roy S, Tang J, Harrer S. Seizurenet: multi-spectral deep feature learning for seizure type classification. Machine Learning in
Clinical Neuroimaging and Radiogenomics in Neuro-Oncology. Springer; 2020:77-87.
17. Covert IC, Krishnan B, Najm I, et al. Temporal graph convolutional networks for automatic seizure detection. Machine Learn Healthcare
Confe. 2019;5:160-180.
18. Bizopoulos P, Lambrou GI, Koutsouris D. Signal2image modules in deep neural networks for EEG classification. 2019 41st Ann Int Conf
IEEE Eng Med Biol Soc (EMBC). 2019;6:702-705.
19. Bouaziz B, Chaari L, Batatia H, Quintero-Rincon A. Epileptic seizure detection using a convolutional neural network. Digital Health
Approach for Predictive, Preventive, Personalised and Participatory Medicine. Springer; 2019:79-86.
20. Rajaguru H, Prabhakar SK. Multilayer autoencoders and empca with genetic algorithm for epilepsy classification from EEG. 2018
Second Int Conf Electron Commun Aerospace Technol (ICECA). IEEE; 2018:353-358.
21. Roy S, Kiral-Kornek I, Harrer S. Chrononet: a deep recurrent neural network for abnormal EEG identification. Conf Artif Intelligence
Med Eur. 2019;11:47-56.
22. Choi G, Park C, Kim J, et al. A novel multi-scale 3d CNN with deep neural network for epileptic seizure detection. 2019 IEEE Int Conf
Consum Electron (ICCE). 2019;3:1-2.
23. Truong ND, Nguyen AD, Kuhlmann L, et al. Convolutional neural networks for seizure prediction using intracranial and scalp
electroencephalogram. Neural Netw. 2018;105:104-111.
24. Wang Z, Oates T. Encoding time series as images for visual inspection and classification using tiled convolutional neural networks,in
Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015.
25. Barra S, Carta SM, Corriga A, Podda AS, Recupero DR. Deep learning and time series-to-image encoding for financial forecasting.
IEEE/CAA J Automatica Sinica. 2020;7(3):683-692.
26. Emami A, Kunii N, Matsuo T, Shinozaki T, Kawai K, Takahashi H. Seizure detection by convolutional neural network-based analysis of
scalp electroencephalography plot images. NeuroImage: Clinical. 2019;22:101684.
27. Yuvaraj R, Thomas J, Kluge T, Dauwels J. A deep learning scheme for automatic seizure detection from long-term scalp EEG. 2018 52nd
Asilomar ConfSig Sys Comput. 2018;9:368-372.
28. Zhou M, Tian C, Cao R, et al. Epileptic seizure detection based on EEG signals and CNN. Front Neuroinform. 2018;12:95.
29. Subasi A, Kevric J, Canbaz MA. Epileptic seizure detection using hybrid machine learning methods. Neural Comput Appl. 2019;31(1):
317-325.
30. Qaisar SM, Subasi A. Effective epileptic seizure detection based on the event-driven processing and machine learning for mobile
healthcare. J Ambient Intelligence Human Comput Secur. 2020;8:1-13.
31. Tuncer T, Dogan S, Ertam F, Subasi A. A novel ensemble local graph structure-based feature extraction network for EEG signal analysis.
Biomed Signal Process Control. 2020;61:102006.
32. Hassan AR, Siuly S, Zhang Y. Epileptic seizure detection in EEG signals using tunable-q factor wavelet transform and bootstrap
aggregating. Comput Methods Programs Biomed. 2016;137:247-259.
33. Hassan AR, Haque MA. Epilepsy and seizure detection using statistical features in the complete ensemble empirical mode decomposi-
tion domain,in TENCON 20152015 IEEE region 10 conferences, pp. 16, 2015.
34. Hassan AR, Subasi A. Automatic identification of epileptic seizures from EEG signals using linear programming boosting. Computer
Methods and Programs in Biomedicine. Vol 136. Elsevier; 2016:65-77.
IBRAHIM ET AL.29 of 30
35. Hassan AR, Subasi A, Zhang Y. Epilepsy seizure detection using complete ensemble empirical mode decomposition with adaptive noise.
Knowledge-Based Sys. 2020;191:105333.
36. Sharma R, Pachori RB. Classification of epileptic seizures in EEG signals based on phase space representation of intrinsic mode
functions. Expert Sys Appl. 2015;42(3):1106-1117.
37. Shankar A, Khaing HK, Dandapat S, Barma S. Analysis of epileptic seizures based on EEG using recurrence plot images and deep
learning. Biomed Signal Process Control. 2021;69:102854.
38. PhysioNet, CHB-MIT Scalp EEG Database.https://www.physionet.org/pn6/chbmit/, 2000.
39. Hasnat A, Bohn'e J, Milgram J, Gentric S, Chen L. Deepvisage: making face recognition simple yet with powerful generalization skills.
Proc IEEE Int Conf Comput Vision Workshops. 2017;3:1682-1691.
40. Pham L, Phan H, Nguyen T, Palaniappan R, Mertins A, McLoughlin I. Robust acoustic scene classification using a multispectrogram
encoder-decoder framework. Digital Signal Process. 2021;110:102943.
41. Abdellatef E, Omran EM, Soliman RF, et al. Fusion of deep-learned and hand-crafted features for cancelable recognition systems. Soft
Comput. 2020;24(20):15189-15208.
42. Sokolova M, Lapalme G. A systematic analysis of performance measures for classification tasks. Inf Process Manag. 2009;45(4):427-437.
43. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv. 2014;12:1409.1556.
44. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. Proc IEEE Conf Comput Vision Pattern Recog. 2016;6:
770-778.
45. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer vision. Proc IEEE Conf Comput
Vision Pattern Recog. 2016;5:2818-2826.
46. Szegedy C, Liu W, Jia Y, et al. Going deeper with convolutions. Proc IEEE Conf Comput Vision Pattern Recog. 2015;7:1-9.
47. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. Proc IEEE Conf Comput Vision Pattern
Recog. 2017;11:4700-4708.
48. Redmon J, Farhadi A. Yolov3: an incremental improvement. arXiv. 2018;8:1804.02767.
49. Polat K, Nour M. Epileptic seizure detection based on new hybrid models with electroencephalogram signals. IRBM. 2020;41(6):331-353.
50. Salini GA, P'erez P. A study of the dynamic behavior of fine particulate matter in Santiago, Chile. Aerosol Air Qual Res. 2015;15(1):
154-165.
51. Fraser AM, Swinney HL. Independent coordinates for strange attractors from mutual information. Phys Rev A. 1986;33(2):1134-1140.
52. Kennel MB, Brown R, Abarbanel HD. Determining embedding dimension for phase-space reconstruction using a geometrical construc-
tion. Phys Rev A. 1992;45(6):3403-3411.
53. Madhavan S, Tripathy RK, Pachori RB. Time-frequency domain deep convolutional neural network for the classification of focal and
non-focal EEG signals. IEEE Sensors J. 2019;20(6):3078-3086.
54. Zhang Y, Guo Y, Yang P, Chen W, Lo B. Epilepsy seizure prediction on EEG using common spatial pattern and convolutional neural
network. IEEE J Biomed Health Inform. 2019;24(2):465-474.
55. Acharya UR, Oh SL, Hagiwara Y, Tan JH, Adeli H. Deep convolutional neural network for the automated detection and diagnosis of
seizure using EEG signals. Comput Biol Med. 2018;100:270-278.
56. Akbari H, Saraf Esmaili S, Farzollah Zadeh S. Detection of seizure EEG signals based on reconstructed phase space of rhythms in ewt
domain and genetic algorithm. Signal Process Renewable Energy. 2020;4(2):23-36.
57. Yakopcic C, Alom MZ, Taha TM. Extremely parallel memristor crossbar architecture for convolutional neural network implementation.
2017 Int Joint Conf Neural Networks (IJCNN). 2017;8:1696-1703.
How to cite this article: Ibrahim FE, Emara HM, El-Shafai W, et al. Deep-learning-based seizure detection and
prediction from electroencephalography signals. Int J Numer Meth Biomed Engng. 2022;e3573.
doi:10.1002/cnm.3573
30 of 30 IBRAHIM ET AL.
... An SVM is used for classification. Ibrahim et al. [18] introduced three models for the classification task of EEG signals. Two of the models are patient-specific and designed for the classification of NS versus PS activities for seizure prediction, and NS versus S activities for seizure detection. ...
... In this work, we propose a CNN model in which the residual learning and depth concatenation strategies have been adopted as illustrated in Fig. 7. The size of the input layer is 227 × 227 × 3. The proposed model contains thirteen convolutional layers, each of which produces an output feature map f c,l x,y,k for a particular layer l and an input f O p ,l−1 x,y [18]: ...
Article
Full-text available
Automatic seizure detection and prediction using clinical Electroencephalograms (EEGs) are challenging tasks due to factors such as low Signal-to-Noise Ratios (SNRs), high variance in epileptic seizures among patients, and limited clinical data constraints. To overcome these challenges, this paper presents two approaches for EEG signal classification. One of these approaches depends on Machine Learning (ML) tools. The used features are different types of entropy, higher-order statistics, and sub-band energies in the Hilbert Marginal Spectrum (HMS) domain. The classification is performed using Support Vector Machine (SVM), Logistic Regression (LR), and K-Nearest Neighbor (KNN) classifiers. Both seizure detection and prediction scenarios are considered. The second approach depends on spectrograms of EEG signal segments and a Convolutional Neural Network (CNN)-based residual learning model. We use 10000 spectrogram images for each class. In this approach, it is possible to perform both seizure detection and prediction in addition to a 3-state classification scenario. Both approaches are evaluated on the Children's Hospital Boston and the Massachusetts Institute of Technology (CHB-MIT) dataset, which contains 24 EEG recordings for 6 males and 18 females. The results obtained for the HMS-based model showed an accuracy of 100%. The CNN-based model achieved accuracies of 97.66%, 95.59%, and 94.51% for Seizure (S) versus Pre-Seizure (PS), Non-Seizure (NS) versus S, and NS versus S versus PS classes, respectively. These results demonstrate that the proposed approaches can be effectively used for seizure detection and prediction. They outperform the state-of-the-art techniques for automatic seizure detection and prediction. Block diagram of proposed epileptic seizure detection method using HMS with different classifiers.
... Given that classification and prediction tasks based on EEG signals are popular multivariable time-series tasks, the automatic identification of EP from EEG signals has long been a research topic of interest to clinical physicians. The advent of machine learning in computing has enhanced the automated analysis of EP [8,9], demonstrating promising classification capabilities across time [10][11][12], frequency [13,14], and time-frequency domains [15], as well as measures of complexity and synchrony [16][17][18][19][20][21][22]. ...
Article
Full-text available
Background Over 60% of epilepsy patients globally are children, whose early diagnosis and treatment are critical for their development and can substantially reduce the disease’s burden on both families and society. Numerous algorithms for automated epilepsy detection from EEGs have been proposed. Yet, the occurrence of epileptic seizures during an EEG exam cannot always be guaranteed in clinical practice. Models that exclusively use seizure EEGs for detection risk artificially enhanced performance metrics. Therefore, there is a pressing need for a universally applicable model that can perform automatic epilepsy detection in a variety of complex real-world scenarios. Method To address this problem, we have devised a novel technique employing a temporal convolutional neural network with self-attention (TCN-SA). Our model comprises two primary components: a TCN for extracting time-variant features from EEG signals, followed by a self-attention (SA) layer that assigns importance to these features. By focusing on key features, our model achieves heightened classification accuracy for epilepsy detection. Results The efficacy of our model was validated on a pediatric epilepsy dataset we collected and on the Bonn dataset, attaining accuracies of 95.50% on our dataset, and 97.37% (A v. E), and 93.50% (B vs E), respectively. When compared with other deep learning architectures (temporal convolutional neural network, self-attention network, and standardized convolutional neural network) using the same datasets, our TCN-SA model demonstrated superior performance in the automated detection of epilepsy. Conclusion The proven effectiveness of the TCN-SA approach substantiates its potential as a valuable tool for the automated detection of epilepsy, offering significant benefits in diverse and complex real-world clinical settings.
... Inceptionv3 [30] is a CNN architecture that builds upon the successful GoogLeNet [31] model, which has shown high accuracy in classifying biomedical data using transfer learning [32][33][34]. In addition, inception-v3 introduces an inception module that combines convolutional filters of various sizes into a single filter, similar to GoogLeNet. ...
Article
Diabetic retinopathy (DR) is a significant cause of vision impairment, emphasizing the critical need for early detection and timely intervention to avert visual deterioration. Diagnosing DR is inherently complex, as it necessitates the meticulous examination of intricate retinal images by experienced specialists. This makes the early diagnosis of DR essential for effective treatment and prevention of eventual blindness. Traditional diagnostic methods, relying on human interpretation of these medical images, face challenges in terms of accuracy and efficiency. In the present research, we introduce a novel method that offers superior precision in DR diagnosis, compared to traditional methods, by employing advanced deep learning techniques. Central to this approach is the concept of transfer learning. This entails the utilization of pre-existing, well-established models, specifically InceptionResNetv2 and Inceptionv3, to extract features and fine-tune selected layers to cater to the unique requirements of this specific diagnostic task. Concurrently, we also present a newly devised model, DiaCNN, which is tailored for the classification of eye diseases. To prove the efficacy of the proposed methodology, we leveraged the Ocular Disease Intelligent Recognition (ODIR) dataset, which comprises eight different eye disease categories. The results are promising. The InceptionResNetv2 model, incorporating transfer learning, registered an impressive 97.5% accuracy in both the training and testing phases. Its counterpart, the Inceptionv3 model, achieved an even more commendable 99.7% accuracy during training, and 97.5% during testing. Remarkably, the DiaCNN model showcased unparalleled precision, achieving 100% accuracy in training and 98.3% in testing. These figures represent a significant leap in classification accuracy when juxtaposed with existing state-of-the-art diagnostic methods. Such advancements hold immense promise for the future, emphasizing the potential of our proposed technique to revolutionize the accuracy of DR and other eye disease diagnoses. By facilitating earlier detection and more timely interventions, this approach stands poised to significantly reduce the incidence of blindness associated with DR, thus heralding a new era of improved patient outcomes. Therefore, this work, through its novel approach and stellar results, not only pushes the boundaries of DR diagnostic accuracy but also promises a transformative impact in early detection and intervention, aiming to substantially diminish DR-induced blindness and champion enhanced patient care.
... However, some methods of DL obtain the particular output for the locations of pathological that neglects the interpretability problems in the practices of medicine because of the characteristics of the black box for the neural networks. Furthermore, recent works on epileptic seizure prediction utilised enhanced DL algorithms which tend to overcome the class imbalance problems by extracting the geometric and handcrafted features [14,15]. Widely, General Adversarial Networks and their enhanced versions were adopted; however, they are limited with stability issues [16,17]. ...
Article
Full-text available
Epilepsy is considered as a serious brain disorder in which patients frequently experience seizures. The seizures are defined as the unexpected electrical changes in brain neural activity, which leads to unconsciousness. Existing researches made an intense effort for predicting the epileptic seizures using brain signal data. However, they faced difficulty in obtaining the patients' characteristics because the model's distribution turned to fake predictions, affecting the model's reliability. In addition, the existing prediction models have severe issues, such as overfitting and false positive rates. To overcome these existing issues, we propose a deep learning approach known as Deep dual‐patch attention mechanism (D²PAM) for classifying the pre‐ictal signals of people with Epilepsy based on the brain signals. Deep neural network is integrated with D²PAM, and it lowers the effect of differences between patients to predict ES. The multi‐network design enhances the trained model's generalisability and stability efficiently. Also, the proposed model for processing the brain signal is designed to transform the signals into data blocks, which is appropriate for pre‐ictal classification. The earlier warning of epilepsy with the proposed model obtains the auxiliary diagnosis. The data of real patients for the experiments provides the improved accuracy by D²PAM approximation compared to the existing techniques. To be more distinctive, the authors have analysed the performance of their work with five patients, and the accuracy comes out to be 95%, 97%, 99%, 99%, and 99% respectively. Overall, the numerical results unveil that the proposed work outperforms the existing models.
... In addition, we deployed machine learning and deep learning, which have been employed in various applications, including emotion recognition [37], speech recognition [38], image reconstruction [39], and medical diagnosis [40][41][42][43]. The present study suggests a deep learning approach based on 2D and 3D CNNs for identifying fatigue in images and videos. ...
Article
Full-text available
Due to the widespread issue of road accidents, researchers have been drawn to investigate strategies to prevent them. One major contributing factor to these accidents is driver fatigue resulting from exhaustion. Various approaches have been explored to address this issue, with machine and deep learning proving to be effective in processing images and videos to detect asymmetric signs of fatigue, such as yawning, facial characteristics, and eye closure. This study proposes a multistage system utilizing machine and deep learning techniques. The first stage is designed to detect asymmetric states, including tiredness and non-vigilance as well as yawning. The second stage is focused on detecting eye closure. The machine learning approach employs several algorithms, including Support Vector Machine (SVM), k-Nearest Neighbor (KNN), Multi-layer Perceptron (MLP), Decision Tree (DT), Logistic Regression (LR), and Random Forest (RF). Meanwhile, the deep learning approach utilizes 2D and 3D Convolutional Neural Networks (CNNs). The architectures of proposed deep learning models are designed after several trials, and their parameters have been selected to achieve optimal performance. The effectiveness of the proposed methods is evaluated using video and image datasets, where the video dataset is classified into three states: alert, tired, and non-vigilant, while the image dataset is classified based on four facial symptoms, including open or closed eyes and yawning. A more robust system is achieved by combining the image and video datasets, resulting in multiple classes for detection. Simulation results demonstrate that the 3D CNN proposed in this study outperforms the other methods, with detection accuracies of 99 percent, 99 percent, and 98 percent for the image, video, and mixed datasets, respectively. Notably, this achievement surpasses the highest accuracy of 97 percent found in the literature, suggesting that the proposed methods for detecting drowsiness are indeed effective solutions.
... As regards machine learning models, it should be considered that diagnosing brain diseases require to intervene human generally, but by machine learning techniques this issue would be removed, where the machine learns how to perform with the scripts [8,9]. Many works recently have been done by utilizing deep learning algorithms, for example, in the study of the entire part of the process implemented by Xception networks with an accuracy percentage of 98.47%, which is compared to the current study this number has reached 100% diagnosis of the epilepsy disease. ...
Chapter
Full-text available
Even though, epilepsy is recognized as a component of brain disease, and seizures in these diseases vary from person to person. This study aimed to implement some procedures based on machine learning and deep learning on EEG signal frequency, where they were compared with each other. Accordingly, to extract more accurate statistical features such as the Short-Time Fourier Transform (STFT) environment in terms of which artificial intelligent algorithms are the most effective applications for early diagnosis in epilepsy disease, is the significant point of the research. After removing the noise from the EEG signal, it entered the STFT environment, where its spectrogram was determined. In the STFT environment, the statistical features of the signal have been extracted, which are segmented at a frequency sampling rate of 173.61 Hz in the form of triangular windows. Finally, seven segmentations have been obtained, in each of which four features have been extracted, for a total of 28 statistical features. Technically, in terms of the deep learning algorithm, feature extraction is done by performing CNN models such as Xception, EfficientNetB0, ResNet50, and VGG16, which are given as inputs to a simple model with the label of being healthy or an epileptic patient, and then the classification of two classes has been applied. In addition, in the machine learning algorithm, basically, the EEG signals were transferred from the time domain to the STFT domain while they were placed next to each other, where we applied an ANOVA feature selection, and extracted 10 main features, which are given to the model as inputs. Furthermore, K-Fold Validation was performed in the dataset evaluation and division section, and eventually, the results were computed in the Support Vector Machine (SVM) model, which equaled 95%, Linear Discriminant Analysis (LDA) equaled 97.5%, and K-nearest neighbors (KNN) equaled 90%. As a conclusion, the study has been figured out that the deep learning algorithm has better performance than the machine learning algorithm in order to early diagnose epilepsy disease.KeywordsEpilepsy diseaseMachine learningDeep learningEEG signals
Article
Full-text available
Epilepsy is a widespread neurological disorder characterized by recurring seizures that have a significant impact on individuals' lives. Accurately recognizing epileptic seizures is crucial for proper diagnosis and treatment. Deep learning models have shown promise in improving seizure recognition accuracy. However, optimizing their performance for this task remains challenging. This study presents a new approach to optimize epileptic seizure recognition using deep learning models. The study employed a dataset of Electroencephalography (EEG) recordings from multiple subjects and trained nine deep learning architectures with different preprocessing techniques. By combining a 1D convolutional neural network (Conv1D) with a Long Short-Term Memory (LSTM) network, we developed the Conv1D ? LSTM architecture. This architecture, augmented with dropout layers, achieved an effective test accuracy of 0.993. The LSTM architecture alone achieved a slightly lower accuracy of 0.986. Additionally, the Bidirectional LSTM (BiLSTM) and Gated Recurrent Unit (GRU) architectures performed exceptionally well, with accuracies of 0.983 and 0.984, respectively. Notably, standard scaling proved to be advantageous, significantly improving the accuracy of both BiLSTM and GRU compared to MinMax scaling. These models consistently achieved high test accuracies across different percentages of Principal Component Analysis (PCA), with the best results obtained when retaining 50% and 90% of the features. Chi-square feature selection also enhanced the classification performance of BiLSTM and GRU models. The study reveals that different deep learning architectures respond differently to feature scaling, PCA, and feature selection methods. Understanding these nuances can lead to optimized models for epileptic seizure recognition, ultimately improving patient outcomes and quality of life.
Article
Objective: In this study, we present a novel biomimetic deep learning network for epileptic spasms and seizure prediction and compare its performance with state-of-the-art conventional machine learning models. Methods: Our proposed model incorporates modular Volterra kernel convolutional networks and bidirectional recurrent networks in combination with the phase amplitude cross-frequency coupling features derived from scalp EEG. They are applied to the standard CHB-MIT dataset containing focal epilepsy episodes as well as two other datasets from the Montefiore Medical Center and the University of California Los Angeles that provide data of patients experiencing infantile spasm (IS) syndrome. Results: Overall, in this study, the networks can produce accurate predictions (100%) and significant detection latencies (10 min). Furthermore, the biomimetic network outperforms conventional ones by producing no false positives. Significance: Biomimetic neural networks utilize extensive knowledge about processing and learning in the electrical networks of the brain. Predicting seizures in adults can improve their quality of life. Epileptic spasms in infants are part of a particular seizure type that needs identifying when suspicious behaviors are noticed in babies. Predicting epileptic spasms within a given time frame (the prediction horizon) suggests their existence and allows an epileptologist to flag an EEG trace for future review.
Article
Gene sequence classification can be seen as a challenging task due to the nonstationary, noisy and nonlinear characteristics of sequential data. The primary goal of this research is to develop a general solution approach for supervised DNA chromatogram (DNAC) classification in the absence of sufficient training data. Today, deep learning comes to the fore with its achievements, however this requires a lot of training data. Finding enough training data can be exceedingly challenging, particularly in the medical area and for rare disorders. In this paper, a novel supervised DNAC classification method is proposed, which combines three techniques to classify hepatitis virus DNA trace files as HBV and HCV. The features that are capable of reflecting the complex-structured sequential data are extracted based on both embedding and spectral entropies. After the supervised dimension reduction step, not only global behavior of the entropy features but also local behavior of the entropy features is taken into account for classification purpose. A memory-based learning, which cannot lose any information coming from training data as its nature, is being used as a classifier. Experimental results show that the proposed method achieves good results that although 19% training data is used, a performance of 92% is obtained.
Article
Full-text available
Epilepsy is a brain disorder which stems from the abnormal activity of neurons and recording of seizures has primary interest in the evaluation of epileptic patients. A seizure is the phenomenon of rhythmicity discharge from either a local area or the whole brain and the individual behavior usually lasts from seconds to minutes. In this work, empirical wavelet transform (EWT) is applied to decompose signals into Electroencephalography (EEG) rhythms. EEG signals are separated into the delta, theta, alpha, beta and gamma rhythms using EWT. The proposed method has been evaluated by the benchmark dataset which is freely downloadable from the Bonn University website. Ellipse area (A) and shortest distance to 45 and 135-degree lines are computed from the 2D projection of reconstructed phase space (RPS) of rhythms as features. After that, the genetic algorithm is used as feature selection. Finally, selected features are fed to the K-nearest neighbor (KNN) classifier for the detection of the seizure (S) and seizure-free (SF) EEG signals. Our proposed method archived 98.33% accuracy in the classification of S and SF EEG signals with a tenfold cross-validation strategy that is higher than previous techniques.
Article
Full-text available
A variety of screening approaches have been proposed to diagnose epileptic seizures, using electroencephalography (EEG) and magnetic resonance imaging (MRI) modalities. Artificial intelligence encompasses a variety of areas, and one of its branches is deep learning (DL). Before the rise of DL, conventional machine learning algorithms involving feature extraction were performed. This limited their performance to the ability of those handcrafting the features. However, in DL, the extraction of features and classification are entirely automated. The advent of these techniques in many areas of medicine, such as in the diagnosis of epileptic seizures, has made significant advances. In this study, a comprehensive overview of works focused on automated epileptic seizure detection using DL techniques and neuroimaging modalities is presented. Various methods proposed to diagnose epileptic seizures automatically using EEG and MRI modalities are described. In addition, rehabilitation systems developed for epileptic seizures using DL have been analyzed, and a summary is provided. The rehabilitation tools include cloud computing techniques and hardware required for implementation of DL algorithms. The important challenges in accurate detection of automated epileptic seizures using DL with EEG and MRI modalities are discussed. The advantages and limitations in employing DL-based techniques for epileptic seizures diagnosis are presented. Finally, the most promising DL models proposed and possible future works on automated epileptic seizure detection are delineated.
Article
Full-text available
Seizure detection and seizure prediction are a very hot topics in medical signal processing due to their importance in automatic medical diagnosis. This paper presents three efficient frameworks for applications related to electroencephalogram signal processing. The first one presents an automatic seizure detection framework based on the utilization of scale-invariant feature transform as an extraction tool. The second one depends on the utilization of the fast Fourier transform (FFT) and an artificial neural network for epileptic seizure prediction. Finally, an automated patient specific framework for channel selection and seizure prediction is presented based on FFT. The simulation results show the success of the proposed frameworks for automated medical diagnosis.
Article
Full-text available
This paper is concerned with Electroencephalography (EEG) seizure prediction, which means the detection of the pre-ictal state prior to ictal activity occurrence. The basic idea of the proposed approach for EEG seizure prediction is to work on the signals in the Hilbert domain. The operation in the Hilbert domain guarantees working on the low-pass spectra of EEG signal segments to avoid artifacts. Signal attributes in the Hilbert domain including amplitude, derivative, local mean, local variance, and median are analyzed statistically to perform the channel selection and seizure prediction tasks. Pre-defined prediction and false-alarm probabilities are set to select the channels, the attributes, and bins of probability density functions (PDFs) that can be useful for seizure prediction. Due to the multi-channel nature of this process, there is a need for a majority voting strategy to take a decision for each signal segment. Simulation results reveal an average prediction rate of 96.46%, an average false-alarm rate of 0.028077/h and an average prediction time of 60.1595 min for a 90-min prediction horizon.
Article
Full-text available
This article is mainly concerned with COVID-19 diagnosis from X-ray images. The number of cases infected with COVID-19 is increasing daily, and there is a limitation in the number of test kits needed in hospitals. Therefore, there is an imperative need to implement an efficient automatic diagnosis system to alleviate COVID-19 spreading among people. This article presents a discussion of the utilization of convolutional neural network (CNN) models with different learning strategies for automatic COVID-19 diagnosis. First, we consider the CNN-based transfer learning approach for automatic diagnosis of COVID-19 from X-ray images with different training and testing ratios. Different pre-trained deep learning models in addition to a transfer learning model are considered and compared for the task of COVID-19 detection from X-ray images. Confusion matrices of these studied models are presented and analyzed. Considering the performance results obtained, ResNet models (ResNet18, ResNet50, and ResNet101) provide the highest classification accuracy on the two considered datasets with different training and testing ratios, namely 80/20, 70/30, 60/40, and 50/50. The accuracies obtained using the first dataset with 70/30 training and testing ratio are 97.67%, 98.81%, and 100% for ResNet18, ResNet50, and ResNet101, respectively. For the second dataset, the reported accuracies are 99%, 99.12%, and 99.29% for ResNet18, ResNet50, and ResNet101, respectively. The second approach is the training of a proposed CNN model from scratch. The results confirm that training of the CNN from scratch can lead to the identification of the signs of COVID-19 disease.
Article
This work proposes deep learning (DL) based epileptic seizure detection by generating 2D recurrence plot (RP) images of EEG signals for specific brain rhythms. The DL bypasses hand-crafted feature engineering, but extracts feature automatically from input images has displayed significant performance in various domain classification tasks. However, generating 2D images from 1D EEG signals and its quality assessment for DL pipeline has not been addressed properly, which is very crucial as the performance of the DL highly relies on input quality. Besides, suitable brain rhythm for seizure analysis has not been explored properly. Therefore, in this work, 2D input images have been generated by the RP technique from EEG signals for specific brain rhythms by preserving the nonlinear characteristics of EEG and employed a well-known DL, called convolution neural network (CNN). For, experimental validation, two well recognized EEG databases for seizure analysis from Bonn University and CHB-MIT (PhysioNet) have been considered. Eventually, three major parameters — recurrence threshold, time delay, and embedding dimension for an RP image generation have been evaluated and detailed. The results show that the proposed method can achieve classification accuracy up to 93%, which is significantly higher and the δ rhythm has been found suitable for seizure detection. The entropy of RP has been found as a suitable parameter for image quality assessment along with two global statistical parameters such as skewness of root mean square and standard of RP images. In performance evaluation, the proposed method demonstrates its competency by displaying the best classification accuracy compared to related works.
Chapter
Automatic classification of epileptic seizure types in electroencephalograms (EEGs) data can enable more precise diagnosis and efficient management of the disease. This task is challenging due to factors such as low signal-to-noise ratios, signal artefacts, high variance in seizure semiology among epileptic patients, and limited availability of clinical data. To overcome these challenges, in this paper, we present SeizureNet, a deep learning framework which learns multi-spectral feature embeddings using an ensemble architecture for cross-patient seizure type classification. We used the recently released TUH EEG Seizure Corpus (V1.4.0 and V1.5.2) to evaluate the performance of SeizureNet. Experiments show that SeizureNet can reach a weighted F1 score of up to 0.95 for seizure-wise cross validation and 0.62 for patient-wise cross validation for scalp EEG based multi-class seizure type classification. We also show that the high-level feature embeddings learnt by SeizureNet considerably improve the accuracy of smaller networks through knowledge distillation for applications with low-memory constraints.
Article
This article proposes an encoder-decoder network model for Acoustic Scene Classification (ASC), the task of identifying the scene of an audio recording from its acoustic signature. We make use of multiple low-level spectrogram features at the front-end, transformed into higher level features through a well-trained CNN-DNN front-end encoder. The high-level features and their combination (via a trained feature combiner) are then fed into different decoder models comprising random forest regression, DNNs and a mixture of experts, for back-end classification. We conduct extensive experiments to evaluate the performance of this framework on various ASC datasets, including LITIS Rouen and IEEE AASP Challenge on Detection and Classification of Acoustic Scenes and Events (DCASE) 2016 Task 1, 2017 Task 1, 2018 Tasks 1A & 1B and 2019 Tasks 1A & 1B. The experimental results highlight two main contributions; the first is an effective method for high-level feature extraction from multi-spectrogram input via the novel CNN-DNN architecture encoder network, and the second is the proposed decoder which enables the framework to achieve competitive results on various datasets. The fact that a single framework is highly competitive for several different challenges is an indicator of its robustness for performing general ASC tasks.
Article
Objectives: Epileptic seizures are one of the most common diseases in society and difficult to detect. In this study, a new method was proposed to automatically detect and classify epileptic seizures from EEG (Electroencephalography) signals. Methods: In the proposed method, EEG signals classification five-classes including the cases of eyes open, eyes closed, healthy, from the tumor region, an epileptic seizure, has been carried out by using the support vector machine (SVM) and the normalization methods comprising the z-score, minimum-maximum, and MAD normalizations. To classify the EEG signals, the support vector machine classifiers having different kernel functions, including Linear, Cubic, and Medium Gaussian, have been used. In order to evaluate the performance of the proposed hybrid models, the confusion matrix, ROC curves, and classification accuracy have been used. The used SVM models are Linear SVM, Cubic SVM, and Medium Gaussian SVM. Results: Without the normalizations, the obtained classification accuracies are 76.90%, 82.40%, and 81.70% using Linear SVM, Cubic SVM, and Medium Gaussian SVM, respectively. After applying the z-score normalization to the multi-class EEG signals dataset, the obtained classification accuracies are 77.10%, 82.30%, and 81.70% using Linear SVM, Cubic SVM, and Medium Gaussian SVM, respectively. With the minimum-maximum normalization, the obtained classification accuracies are 77.20%, 82.40%, and 81.50% using Linear SVM, Cubic SVM, and Medium Gaussian SVM, respectively. Moreover, finally, after applying the MAD normalization to the multi-class EEG signals dataset, the obtained classification accuracies are 76.70%, 82.50%, and 81.40% using Linear SVM, Cubic SVM, and Medium Gaussian SVM, respectively. Conclusion: The obtained results have shown that the best hybrid model is the combination of cubic SVM and MAD normalization in the classification of EEG signals classification five-classes.