ChapterPDF Available

Classifying the Human Activities of Sensor Data Using Deep Neural Network

Authors:
  • Al-Mustaqbal University

Abstract and Figures

Today sensors represent one of the most important applications for generating data stream. This data has a number of unique characteristics, including fast data access, huge volume, as well as the most prominent feature, the concept drift. Machine learning in general and deep learning technique in particular is among the predominant and successful selections to classify the human activities. This is due to several reasons such as results quality and processing time. The recognition of human activities that produced from sensors considers is an effective and vital task in the healthcare field, meanwhile, it is an attractive to researchers. This paper presents a DNN model to classify the human activities of the HuGaDB sensor dataset by implementing multilayer perceptron (MLP) structure. The current model achieved results, 91.7% of accuracy, 92.5% precision, 92.0% recall, and 92.0% of F1-score, using a tiny time. The model results were compared with the previous models and it has proven its efficiency by outperforming those models.
Content may be subject to copyright.
Classifying the Human Activities
of Sensor Data Using Deep Neural
Network
Hussein A. A. Al-Khamees(B
), Nabeel Al-A’araji ,
and Eman S. Al-Shamery
Babylon University, Babylon - Hilla, Iraq
Hussein.alkhamees7@gmail.com,
{nhkaghed,emanalshamery}@itnet.uobabylon.edu.iq
Abstract. Today sensors represent one of the most important appli-
cations for generating data stream. This data has a number of unique
characteristics, including fast data access, huge volume, as well as the
most prominent feature, the concept drift. Machine learning in general
and deep learning technique in particular is among the predominant and
successful selections to classify the human activities. This is due to sev-
eral reasons such as results quality and processing time. The recognition
of human activities that produced from sensors considers is an effec-
tive and vital task in the healthcare field, meanwhile, it is an attractive
to researchers. This paper presents a DNN model to classify the human
activities of the HuGaDB sensor dataset by implementing multilayer per-
ceptron (MLP) structure. The current model achieved results, 91.7% of
accuracy, 92.5% precision, 92.0% recall, and 92.0% of F1-score, using a
tiny time. The model results were compared with the previous models
and it has proven its efficiency by outperforming those models.
Keywords: Deep neural network ·MultiLayer Perceptron (MLP) ·
Human activities classification ·Sensor data stream ·HuGaDB dataset
1 Introduction
Many real-world applications in different domains can generate a massive amount
of data; It is known as a data stream which has various unique properties that
traditional data do not have. Some of these characteristics are unlimited data
size, fast-access data from the source, limited memory, processing time and the
evolving in its nature this causes the concept drift [1]. Most traditional data algo-
rithms fail when dealing with a data stream, this is due to newly characteristics
of the data stream [2].
Machine learning is a sub-field of Artificial Intelligence (AI). In reality,
machine learning consists of many techniques that can be used on data stream
Supported by Babylon University.
c
Springer Nature Switzerland AG 2022
A. Bennour et al. (Eds.): ISPR 2022, CCIS 1589, pp. 107–118, 2022.
https://doi.org/10.1007/978-3-031-08277-1_9
108 H. A. A. Al-Khamees et al.
such as classification, clustering, regression, ..., etc. [3]. Despite these techniques,
neural networks (NN) are just as important as those techniques and that can
be also implemented on the data stream. Neural networks have two types, shal-
low or deep. Recently, deep learning techniques that use deep neural networks
(DNN) are a major area of interest and increasingly being applied [4]. Therefore,
DNN is applied in various fields such as healthcare [5].
Deep learning depends mainly on Artificial Neural Networks (ANN), which
are originally inspired by neurons in the human brain [6]. However, the DNNs
structure involves three layers (input, hidden and output), where every layer
has several neurons and the neuron numbers differ from a layer to another [4].
Multilayer Perceptron (MLP) is an important and widely used architectural type
of deep learning [7].
The recognition task of human activities that produced from sensors consid-
ers is an effective and vital task in the healthcare field. Indeed, the recognition
models are either wearable or external sensor-based models [8].
This paper presents a deep learning model based on MLP and the back-
propagation algorithm to train MLP for classifying the human activities. This
model consists of four hidden layers that able to implement the classification
task in a short period of time. For evaluating the proposed model, the HuGaBD
sensor dataset was used. More specifically, five sub-datasets of the main HuGaDB
dataset were selected for the current model.
The proposed model achieved results as follow, 91.7% of accuracy, 92.5% pre-
cision, 92.0% recall, and 92.0 % of F1-score, using a tiny time. Accordingly, this
model outperforms many previous models that used the same dataset (HuGaDB
dataset) to classify the human activities. Furthermore, our evaluation demon-
strates how this DNN model proved the enhancing of results by implementing it
into different numbers of both hidden layers and also neurons for every hidden
layer.
The current paper organizes as follows. Section 2discusses related works
which related to deep neural network that implemented on HuGaDB dataset.
Section 3explains the DNN structure. The methodology of the proposed model
is presented in Sect. 4. While dataset description is introduces in Sect. 5. Section
6illustrates the evaluation metrics and Sect. 7dedicates to the results of the
model and finally, the conclusion of the current paper summarizes.
2 Related Work
This section covers the studies based on NNs as a machine learning technique
that applied to the HuGaDB dataset to classify the human activities.
In [9], the authors presented model aims to classify different activities of the
human. The model depended on ANN to estimate many parameters such as
IMUs displacements, velocity, and angle. The study focuses on three body area
that are shin, thigh and waist that resulted in accurate results of the lower limbs
of the human body. Moreover, the proposed model aims to solve an important
issue, which is the contradictions that occur (while capturing the motion signal)
Classifying the Human Activities of Sensor Data 109
to the movements of body parts such as the hand and the leg. In general, the
model consists of two phases that are training and application. In the first phase,
the ANN is trained to estimate the received signals while in the second phase,
the ANN that was trained is implemented to estimate the signals related to the
lower extremities (during real time). This model achieved an accuracy of 88.0%.
Accordingto[10], the authors applied feature vector length reduction and
how it affects deep learning networks besides other techniques of machine learn-
ing. The key idea behind the model is to apply Long Short-term Memory (LSTM)
as a deep learning classifier to extract different high dimensional features. The
model has several phases which are data pre-processing, feature extraction,
feature selection, training and finally the testing phase. The proposed model
attained an accuracy of 91.1%.
B. Fang et al. [11] suggested a gait neural network (GNN) model which
depended on a temporal convolutional neural network. The model aims to predict
a human activity in the lower limbs. In general, the structure of the proposed
model consists of gait prediction and gait recognition where it focuses on the
gait data that received from the right leg. The accuracy achieved by the model
based on GNN is 79.24%, which is considered the highest accuracy among the
techniques used in the same study.
3 DNN Structure
The DNN structure consists of three layer types that are, input layer, hidden
layers and output layer. The data are received from the external source through
the input layer, therefore there isn’t any processing (computations). Most of the
processing steps that implemented in the hidden layers are nonlinear computa-
tions, whereas the processing in the output layer either linear or nonlinear [12].
The nonlinear transforming which starts from the input to the hidden layers till
the output layer, is called as the forward propagation.
The number of hidden layers and the number of neurons in each layer has an
effective effect on the final results of the deep neural network model. Therefore,
it must be carefully selected (after testing) [13].
Each layer contains several neurons, take into consideration that the neuron
number are differs from a layer to another. In a specific layer, every neuron
is connected to their counterparts in adjacent layers. This connection can be
indicated by weights which reflecting both strength and direction. Every neuron
can transform data through computation of weighted sum (of the output neurons
in past layer) and then passes it by a nonlinear function (activation functions)
for deriving the neuron outputs [14].
3.1 Multilayer Perceptron (MLP)
It’s a feed-forward neural network with multi hidden layers. MLP doesn’t require
any prior assumptions about the distribution of data. In MLP, the neurons are
connected by weights and also the signals of output that represented as a function
of the sum of the inputs to the neuron modified by an activation function [15].
110 H. A. A. Al-Khamees et al.
3.2 MLP Training
Usually, the training of the deep neural network is more difficult and complex
than the classic neural network [16].
The training of DNN contains many sequential steps for adjusting the weights
between the neurons in the network, in a similar way to the learning of the human
brain. But before the adjustment step, the model must initialize these weights.
This initialization is done randomly [17], where the resulting weights have the
ability to [18]:
Maximize the relationship strength between network input and its output.
Minimize a difference of the neural networks (such as an error) between a
specific task and its real target (i.e. between the network prediction and its
associated target). Usually, a neural network technique aims to minimize this
error value.
More specifically, the back propagation (BP) is the most successful and widely
used algorithm for MLP training [15]. BP repeatedly can analyze the errors and
optimize every value of weight depending on the errors that generated by the
next layer [18]. Accordingly, this algorithm was used in the current model.
To simplify the weight computation, suppose a neural network contains (m)
neuron, this neuron is driven by input vector Xn, where n indicates to the time
step of the iterative process contains the adjusting step of the input weights
w(mi). Therefore, each sample of data passes through the training step of a
DNN containing X(n)and its output denoting by d(n).
Then the processing step to X(n), of a neuron (m) is generating an output
which is referred by ym(n), and computed by:
ym(n)=f
j
i=1
x.w(mi)(1)
where f indicates to activation function. This output is compared with the target
output dm(n)which normally is given in a sample. The error em(n)can compute
by:
em(n)=(dm(n)ym(n)) (2)
Because its capacity of the back propagation, it is a very appropriate method
to problems that don’t have any relation between the input and output [19].
4 Methodology
The proposed DNN model consists of:
1. Prepare the dataset that will be used in DNN model. In this model, HuGaDB
is used.
Classifying the Human Activities of Sensor Data 111
2. Apply the data pre-processing step by implementing an appropriate tech-
nique. Normalization is a major step in most problems. The normalization
technique has several methods, including the Min-max that applied to this
model. Mathematically, if there is a set of matching scores (Ms) where, s =
1, 2, ..., n, the normalized scores (Ms’) calculate as:
Ms=(Msmin)/(max min) (3)
3. Divide the dataset into training and testing data by applying a suitable tech-
nique. In this model, the cross validation is used, 80% as a training data and
20% as a testing data.
4. Determine the number of hidden layers that required to build the MLP model.
For further analysis, two and four hidden layers were applied.
5. Determine the number of neurons in each hidden layer. In the case of two
hidden layers, the number of neuron is set to (10, 10, 12) while in the case of
four hidden layers, the number of neuron is set to (10, 12, 14, 26, 30).
6. Determine the training algorithm. In this model, the back propagation (BP) is
used for training MLP. In addition to the number of hidden layers and neurons
for every layer, setting another parameters such as, (a) the weights that can
be computed according to equation (1); (b)the error based on equation (2);
and (c) the learning rate that set to 0.001.
7. Start the training phase using the training data (step 3) and parameters
(steps 4 and 5) by the back-propagation (BP) training method (step 6).
8. Start the testing phase by using the test data (step 3). However, in this phase,
the ability of the proposed model is tested if it has been trained to accurately
classify data samples.
9. After completing the training and testing phases, the evaluation step is imple-
mented, as it is the last step in this model. The model uses four different mea-
sures, namely, accuracy, precision, recall and F1-score to evaluate the current
model.
Figure 1shows the model methodology.
5 Data Set Description
Human Gait Database (HuGaDB) to activity recognition from six inertial sen-
sor networks was presented in 2017 [20], these sensors can be shown in Fig. 2(a).
HuGaDB dataset contains 12 behaviors actions which are: walking, sitting, sit-
ting down, sitting in a car, going up, going down, standing, standing up, up
by elevator, down by elevator, bicycling, and running. However, some of these
actions are displayed in Fig.2(b).
According to these behaviors actions, the dataset contains static and dynamic
activities. These several activities are implemented and recorded at various times
like recording the running behavior over about 20 min. Additionally, all the
behaviors actions are gathered by 18 participants.
Furthermore, the main HuGaDB dataset consists of 637 data files and all of
them has the same number of features that are 39 features. Also, all these files
112 H. A. A. Al-Khamees et al.
Fig. 1. Methodology of proposed DNN model.
contain the sentence (various) in their titles to indicate the various activities it
contains. This dataset is a publicly available1.
In the current model, five sub-datasets from the main HuGaDB dataset are
used therefore, it 10 of the 12 activities have been covered through this study.
The activities covered are all activities above except sitting in a car and bicycling.
These sub-datasets are:
1. HuGaDB-v2-various-01-01: consists of 2435 records and it has four classes
that are, ‘sitting’, ‘sitting-down’, ‘standing’, and ‘standing-up’. This dataset
denotes by DS1.
2. HuGaDB-v2-various-05-12: it has 4393 records and it has three classes that
are, ‘going-down’, ‘standing’, and ‘walking’. DS2 is the symbol of this dataset.
3. HuGaDB-v2-various-13-10: it contains 4850 records and it has three classes
that are, ‘down-by-elevator’, ‘standing’, and ‘up-by-elevator’. HuGaDB-v2-
various-13-10 has the symbol DS3.
4. HuGaDB-v2-various-14-05: this dataset has 2392 records. Two classes for this
dataset which are ‘running’ and ‘walking’ and denotes by DS4.
5. HuGaDB-v2-various-17-07: it consists of 2930 records and it has three classes
that are, ’going-up’, ‘standing’, and ‘walking’. DS5 is the symbol for this
dataset.
6 Evaluation Metrics
The performance of the proposed model is evaluated by four different measure-
ments that are [10]:
1https://www.kaggle.com/romanchereshnev/hugadb-human-gait-database.
Classifying the Human Activities of Sensor Data 113
1. Accuracy (refers to the ratio of all true cases divided by the overall dataset
cases).
Accuracy = TP + TN/(TP + TN + FP + FN)
2. Precision (determine the number of true cases predictions which really belong
to the true cases).
Precision = TP/(TP + FP)
3. Recall (determines the number of true cases predictions that implemented
over all true cases).
Recall = TP/(TP + FN)
4. F1-score (indicates the harmonic mean measure to both the precision and
recall).
F1-score = 2 ×(P r ecision ×Recall)/(P recision +Recall)
7 Results
The model implements with two hidden layers, four hidden layers respectively.
Figure 2shows the comparison of accuracy between the two implementations.
Fig. 2. The accuracy for all sub-datasets of HuGaDB dataset with two and four hidden
layers.
Furthermore, Table 1describes the measurements with two hidden layers,
while these measurements with four hidden layers, detail in Table2, and the
best results are highlighted in bold font. While Figs. 3and 4visualize these
measurement values.
114 H. A. A. Al-Khamees et al.
Table 1. The measurements of the model (two hidden layers).
Dataset name Accuracy Precision Recall F1-score
DS1 95.0 96.3 92.5 94.4
DS2 92.3 61.5 63.2 62.3
DS3 74.0 75.9 74.0 74.9
DS4 97.2 97.7 96.4 97.0
DS5 92.4 93.8 93.9 93.8
AVE 90.1 85.0 84.0 84.4
Table 2. The measurements of the model (four hidden layers).
Dataset name Accuracy Precision Recall F1-score
DS1 97.2 97.3 96.4 96.9
DS2 94.0 95.8 96.0 95.9
DS3 76.8 78.7 77.1 77.9
DS4 98.1 98.1 97.8 97.5
DS5 92.4 92.6 92.9 91.9
AVE 91.7 92.5 92.0 92.0
Fig. 3. The measurements of implementation with two hidden layers for every dataset.
Classifying the Human Activities of Sensor Data 115
Fig. 4. The measurements of implementation with four hidden layers for every dataset.
After all these implementations, we notice that the proposed model with
four hidden layers achieves higher accuracy (in all five sub-datasets) than its
counterpart when implementing with two hidden layers.
In the same context, it achieves the highest results in terms of other mea-
surements (precision, recall, and also F1-score) as shown in Table2and Fig. 4.
In fact, the overall accuracy of the proposed model is 91.7 %, which is supe-
rior to many other methods that implemented on the same dataset (HuGaDB
dataset). Table 3and Fig. 5indicate the comparison between the previous models
and our model that implemented for HuGaDB dataset.
Additionally, in term of processing time, the proposed DNN model needs
1.71 s to classify the first sub-dataset (DS1) and 1.66 s to the second sub-dataset
(DS2). While it needs 1.84 s to implement (DS3) and 1.93 s for (DS4). Finally,
it requires 1.85 s to do the classification of the last sub-datasets (DS5). Figure 6
indicates these time details.
Table 3. The accuracy comparisons between previous models and our model.
No Study, publication year Accuracy %
1 [9], 2018 88.0
2 [10], 2020 91.1
3 [11], 2020 79.2
4Our model 91.7
116 H. A. A. Al-Khamees et al.
Fig. 5. The accuracy comparisons between previous models and our model.
Fig. 6. The time needed to implement every dataset.
Classifying the Human Activities of Sensor Data 117
8 Conclusion
The past decade witnessed a prominent development in sensors to generate the
data stream in various fields includes the health field. In this field, the classifica-
tion of the patient’s activities has become the focus of many researchers because
it provides knowledge of the current state of a patient.
Deep Neural Networks (DNNs) are the latest and most preferred machine
learning techniques, especially when processing the data stream. DNN includes
many architectures, the Multi-Layer Perceptron (MLP) is a significant architec-
ture.
This paper presents a deep neural network model based MLP architecture to
classify the human activity during a tiny time. The proposed model was tested
by HuGaDB dataset and evaluated its performance by four measurements which
are accuracy, precision, recall and F1-score. The results proved the superiority
of the proposed model over the previous works, as it achieved an accuracy of
91.7 %, precision of 92.5 %, recall of 92.0% recall, and F1-score as 92.0%.
References
1. Al-Khamees, H.A.A., Al-A’ara ji, N., Al-Shamery, E.S.: Survey: clustering tech-
niques of data stream. In: 1st Babylon International Conference on Information
Technology and Science (BICITS), pp. 113–119. IEEE, Babil (2021). https://doi.
org/10.1109/BICITS51482.2021.9509923
2. Bahri, M., Bifet, A.: Incremental k-nearest neighbors using reservoir sampling
for data streams. In: Soares, C., Torgo, L. (eds.) DS 2021. LNCS (LNAI), vol.
12986, pp. 122–137. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-
88942-5 10
3. Al-Khamees, H.A.A., Al-A’araji, N., Al-Shamery, E.S.: Data stream clustering
using fuzzy-based evolving Cauchy algorithm. Int. J. Intell. Eng. Syst. 14(5), 348–
358 (2021). https://doi.org/10.22266/ijies2021.1031.31
4. Saikiaa, P., Baruaha, R.D., Singhb, S.K., Chaudhurib, P.K.: Artificial neural net-
works in the domain of reservoir characterization: a review from shallow to deep
models. Comput. Geosci. 135, 104357 (2020). https://doi.org/10.1016/j.cageo.
2019.104357
5. Al-Khamees, H.A.A., Al-Jwaid, W.R.H., Al-Shamery, E.S.: The impact of using
convolutional neural networks in COVID-19 tasks: a survey. Int. J. Comput. Digit.
Syst. 11(1), 189–197 (2022). https://doi.org/10.12785/ijcds/110194
6. Lee, J., Chang, C., Kao, T., Wang, J.: Age estimation using correlation-refined
features of convolutional neural network. J. Inf. Sci. Eng. 37(6), 1435–1448 (2021).
https://doi.org/10.6688/JISE.202111-37(6).0014
7. Zhang, S., Yao, L., Sun, A., Tay, Y.: Deep learning based recommender system:
a survey and new perspectives. ACM Comput. Surv. (CSUR) 52(1), 1–38 (2019).
https://doi.org/10.1145/3285029
8. Jansi, R., Amutha, R.: A novel chaotic map based compressive classification scheme
for human activity recognition using a tri-axial accelerometer. Multimed. Tools
Appl. 77(23), 31261–31280 (2018). https://doi.org/10.1007/s11042-018-6117-z
118 H. A. A. Al-Khamees et al.
9. Sun, Y., Yang, G., Lo, B.: An artificial neural network framework for lower limb
motion signal estimation with foot-mounted inertial sensors. In: 15th International
Conference on Wearable and Implantable Body Sensor Networks (BSN), pp. 132–
135. IEEE, Las Vegas (2018). https://doi.org/10.1109/BSN.2018.8329676
10. Kumari, G., Chakraborty, J., Nandy, A.: Effect of reduced dimensionality on deep
learning for human activity recognition. In: 11th International Conference on Com-
puting. Communication and Networking Technologies (ICCCNT), pp. 1–7. IEEE,
Kharagpur (2020). https://doi.org/10.1109/ICCCNT49239.2020.9225419
11. Fang, B., et al.: Gait neural network for human-exoskeleton interaction. Front.
Neurorobot. 14, 1–9 (2020). https://doi.org/10.3389/fnbot.2020.00058
12. Sarker, I.H.: Deep learning: a comprehensive overview on techniques, taxonomy,
applications and research directions. SN Comput. Sci. 2(6), 1–20 (2021). https://
doi.org/10.1007/s42979-021-00815-1
13. Madhiarasan, M., Deepa, S.N.: Comparative analysis on hidden neurons estimation
in multi layer perceptron neural networks for wind speed forecasting. Artif. Intell.
Rev. 23, 1–23 (2016). https://doi.org/10.1007/s10462-016-9506-6
14. Goli, P.: A new perceptually weighted cost function in deep neural network based
speech enhancement systems. Hear. Balance Commun. 17(3), 191–196 (2019).
https://doi.org/10.1080/21695717.2019.1603948
15. Gardner, M.W., Dorling, S.R.: Artificial neural networks (the multilayer percep-
tron) a review of applications in the atmospheric sciences. Atmos. Environ. 32(14–
15), 2627–2636 (1998). https://doi.org/10.1016/S1352-2310(97)00447-0
16. Xu, Z.-Q.J., Zhang, Y., Xiao, Y.: Training behavior of deep neural network in
frequency domain. In: Gedeon, T., Wong, K.W., Lee, M. (eds.) ICONIP 2019.
LNCS, vol. 11953, pp. 264–274. Springer, Cham (2019). https://doi.org/10.1007/
978-3-030-36708- 4 22
17. Larochelle, H., Bengio, Y., Louradour, J., Lamblin, P.: Exploring strategies for
training deep neural networks. J. Mach. Learn. Res. 10(1), 1–40 (2009)
18. Vieira, S., Pinaya, W.H.L., Garcia-Dias, R., Mechelli, A.: Machine Learning Meth-
ods and Applications to Brain Disorders, 1st edn. Academic Press, San Diego
(2019)
19. Nawi, N.M., Khan, A., Rehman, M.Z.: A new back-propagation neural network
optimized with Cuckoo search algorithm. In: Murgante, B., et al. (eds.) ICCSA
2013. LNCS, vol. 7971, pp. 413–426. Springer, Heidelberg (2013). https://doi.org/
10.1007/978-3-642-39637- 3 33
20. Chereshnev, R., Kert´esz-Farkas, A.: HuGaDB: human gait database for activity
recognition from wearable inertial sensor networks. In: van der Aalst, W.M.P.,
et al. (eds.) AIST 2017. LNCS, vol. 10716, pp. 131–141. Springer, Cham (2018).
https://doi.org/10.1007/978-3-319- 73013-4 12
... For the accuracy of the NSL-KDD dataset which is 98.67%, the proposed model outperforms the model presented by [25] that attained an accuracy of 97.05% and also the model proposed by [26] which achieved an accuracy of 97.97%. In terms of the accuracy of HuGaDB dataset which is 97.63%, the proposed model outperforms previous models such as [6] that attained an accuracy of 92.5%, [12] which achieved an accuracy of 88.0%, and [27] that obtained 91.7% as an accuracy. After implementing our proposed idea to a previous MLP model which has two hidden layers and a constant learning rate (0.001) that attained an accuracy rate of 50.2% [28], the accuracy rate increased to 85.80%. ...
Article
Full-text available
The data stream is considered the backbone of many real-world applications. These applications are most effective when using modern techniques of machine learning like deep neural networks (DNNs). DNNs are very sensitive to set parameters, the most prominent one is the learning rate. Choosing an appropriate learning rate value is critical because it is able to control the overall network performance. This paper presents a new developing DNN model using a multi-layer perceptron (MLP) structure that includes network training based on the optimal learning rate. Thereupon, this model consists of three hidden layers and does not adopt the stability of the learning rate but has a non-constant value (varying over time) to obtain the optimal learning rate which is able to reduce the error in each iteration and increase the model accuracy. This is done by deriving a new parameter that is added to and subtracted from the learning rate. The proposed model is evaluated by three streaming datasets: electricity, network security layer-knowledge discovery in database (NSL-KDD), and human gait database (HuGaDB) datasets. The results proved that the proposed model achieves better results than the constant model and outperforms previous models in terms of accuracy, where it achieved 88.16%, 98.67%, and 97.63% respectively.
... Based on the confusion matrix, which provides good details about the classifier, we use accuracy and purity [21]. Accuracy is computed according to ...
Article
Full-text available
Data streams are a modern type of data that differ from traditional data in various characteristics: their indefinite size, high access, and concept drift due to their origin in non-stationary environments. Data stream clustering aims to split these data samples into significant clusters, depending on their similarity. The main drawback of data stream clustering algorithms is the large number of clusters they produce. Therefore, determining an optimal number of clusters is an important challenge for these algorithms. In practice, evolving models can change their general structure by implementing different mechanisms. This paper presents a fuzzy model that mainly consists of an evolving Cauchy clustering algorithm which is updated through a specific membership function and determines the optimal number of clusters by implementing two evolving mechanisms: adding and splitting clusters. The proposed model was tested on six different streaming datasets, namely, power supply, sensor, HuGaDB, UCI-HAR, Luxem-bourg, and keystrokes. The results demonstrated that the efficiency of the proposed model in producing an optimal number of clusters for each dataset outperforms that of previous models.
Article
Full-text available
computer tasks. Machine Learning (ML) as an essential type of AI and deep learning (DL) is merely a branch of (ML). DL can mainly be helping to fast analysis of the medical images, especially the complex images, and this can speed up an early diagnosis of diseases. The Covid-19 pandemic has spread rapidly within societies, creating real panic for all people. Convolutional Neural Network (CNN) is a sub-class of DL which is used to classify medical images. Researchers have exploited the merits of CNNs to deal with COVID-19. This merits and diversity enabled researchers and workers in this field to devise new methods used to detect early cases, predict patients, diagnose patients, design vaccines and drugs and others. This paper aims to conduct a comprehensive survey of the previous works that used CNNs to implement.
Chapter
Full-text available
The online and potentially infinite nature of data streams leads to the inability to store the flow in its entirety and thus restricts the storage to a part of – and/or synopsis information from – the stream. To process these evolving data, we need efficient and accurate methodologies and systems, such as window models (e.g., sliding windows) and summarization techniques (e.g., sampling, sketching, dimensionality reduction). In this paper, we propose, RW-kNN, a k-Nearest Neighbors (kNN) algorithm that employs a practical way to store information about past instances using the biased reservoir sampling to sample the input instances along with a sliding window to maintain the most recent instances from the stream. We evaluate our proposal on a diverse set of synthetic and real datasets and compare against state-of-the-art algorithms in a traditional test-then-train evaluation. Results show how our proposed RW-kNN approach produces high-predictive performance for both real and synthetic datasets while using a feasible amount of resources.
Article
Full-text available
Many different applications in the real world can generate huge amount of data, that has unconventional features including massive size, fast access, besides the evolving in its nature; this is data stream. Data stream clustering algorithms began to grow at breakneck speed. evolving Cauchy (eCauchy) is a significant algorithm of density-based data stream clustering. The major limitation of eCauchy is the high number of clusters generated in dynamic environments. This paper presents an evolving model for data stream by optimizing e-Cauchy algorithm to decrease the number of clusters and reach to an ideal number by implementing evolving mechanisms (adding, merging, splitting clusters) based on a specific membership function. Model is tested by two real datasets NSL-KDD99 and keystroke. Proposed model outperforms two other algorithms, e-Cauchy and FEAC-Stream. Model constructs five and four clusters with less time to implement 1.30 and 2.30 minutes respectively for each dataset.
Article
Full-text available
Deep learning (DL), a branch of machine learning (ML) and artificial intelligence (AI) is nowadays considered as a core technology of today’s Fourth Industrial Revolution (4IR or Industry 4.0). Due to its learning capabilities from data, DL technology originated from artificial neural network (ANN), has become a hot topic in the context of computing, and is widely applied in various application areas like healthcare, visual recognition, text analytics, cybersecurity, and many more. However, building an appropriate DL model is a challenging task, due to the dynamic nature and variations in real-world problems and data. Moreover, the lack of core understanding turns DL methods into black-box machines that hamper development at the standard level. This article presents a structured and comprehensive view on DL techniques including a taxonomy considering various types of real-world tasks like supervised or unsupervised. In our taxonomy, we take into account deep networks for supervised or discriminative learning, unsupervised or generative learning as well as hybrid learning and relevant others. We also summarize real-world application areas where deep learning techniques can be used. Finally, we point out ten potential aspects for future generation DL modeling with research directions. Overall, this article aims to draw a big picture on DL modeling that can be used as a reference guide for both academia and industry professionals.
Article
Full-text available
Robotic exoskeletons are developed with the aim of enhancing convenience and physical possibilities in daily life. However, at present, these devices lack sufficient synchronization with human movements. To optimize human-exoskeleton interaction, this article proposes a gait recognition and prediction model, called the gait neural network (GNN), which is based on the temporal convolutional network. It consists of an intermediate network, a target network, and a recognition and prediction model. The novel structure of the algorithm can make full use of the historical information from sensors. The performance of the GNN is evaluated based on the publicly available HuGaDB dataset, as well as on data collected by an inertial-based wearable motion capture device. The results show that the proposed approach is highly effective and achieves superior performance compared with existing methods.
Article
Full-text available
Nowadays Machine Learning approaches are getting popular in almost all the domains of Engineering Applications. One such widely used approach is Artificial Neural Networks (ANN), that has been successfully applied in many disciplines and becoming popular in the domain of Reservoir Characterization too. A considerable number of neural network papers have been published till now in this domain, and its application is still on the way. The main motive of application of ANN in this domain is to use acquired data from different geological and geophysical sources in determining the characteristics of a reservoir by analyzing the correlation of various data sources. When properly trained, ANN can predict the reservoir properties, by dentifying the complex nonlinear relationship associated with the input data. Different ANN models have been used in the domain of reservoir characterization starting from shallow to deep models with the progress over time. In some scenarios, ANN is combined with other soft computing methodologies resulting in hybrid models. Recently deep learning models of ANN are gaining popularity in many fields including oil exploration. The popularity of deep models is due to its automatic feature extraction capability, ability to handle high dimensional data and above all the ability to solve a problem like our human brain does, learning with multiple levels of abstractions. In this survey, we focus on different evolution of ANN in Reservoir Characterization over time. The evolution is in terms of architecture, learning, as well as to combine with other models of machine learning to improve its modeling capability, that now extends towards recent advanced techniques of ANN called deep learning. From the survey, it is apparent that the application of ANN is very vital in this field and its application will continue even in the future in making intelligent interpretation of oil reservoir.
Chapter
Why deep neural networks (DNNs) capable of overfitting often generalize well in practice is a mystery [24]. To find a potential mechanism, we focus on the study of implicit biases underlying the training process of DNNs. In this work, for both real and synthetic datasets, we empirically find that a DNN with common settings first quickly captures the dominant low-frequency components, and then relatively slowly captures the high-frequency ones. We call this phenomenon Frequency Principle (F-Principle). The F-Principle can be observed over DNNs of various structures, activation functions, and training algorithms in our experiments. We also illustrate how the F-Principle helps understand the effect of early-stopping as well as the generalization of DNNs. This F-Principle potentially provides insight into a general principle underlying DNN optimization and generalization. KeywordsDeep Neural NetworkDeep learningFourier analysisGeneralization
Article
Speech intelligibility improvement is an important task to increase human perception in telecommunication systems and hearing aids when the speech is degraded by the background noises. Although, deep neural network (DNN) based learning architectures which use mean square error (MSE) as the cost function has been found to be very successful in speech enhancement areas, they typically attempt to enhance the speech quality by uniformly optimizing the separation of a target speech signal from a noisy observation over all frequency bands. In this work, we propose a new cost function which further focuses on speech intelligibility improvement based on a psychoacoustic model. The band-importance function, which is a principal component of speech intelligibility index (SII), has been used to determine the relative contribution to speech intelligibility provided by each frequency band in learning algorithm. In addition, we augment a signal to noise ratio (SNR) estimation to the network to improve the generalization of the method to unseen noisy conditions. The performance of the proposed MSE cost function is compared with the conventional MSE cost function in the same conditions. Our approach shows better performance in objective speech intelligibility measures such as coherence SII (CSII) and short-time objective intelligibility (STOI), while mitigating quality scores in perceptual evaluation of speech quality (PESQ) and speech distortion (SD) measure.