Conference PaperPDF Available

Human action recognition to human behavior analysis

Authors:

Abstract and Figures

Human machine interaction becomes one of the most research topics in multimedia processing, traditional techniques for communication are developed in order to tackle technology advances and allow disable person to communicate easily with the machine, and to understand their activity using computer computing. In this paper we are focused on human behavior analysis from video scene and it is worth noticed that many information are hidden behind gesture, sudden motion and walking speed, many research works tried to model and then recognize human behavior through motion analysis. In our work we will explain the human action recognition By K Nearest Neighbors approach.
No caption available
… 
Content may be subject to copyright.
Human Action Recognition to Human Behavior
Analysis
Neziha JAOUEDI
National Engineering School of
Gabes, Tunisia
R.U. SETIT.
Higher Institute
of Biotechnology of sfax,
Tunisia
neziha_jaouedi@yahoo.fr
Noureddine BOUJNAH
Faculty of Sciences Gabes,
Tunisia
boujnah_noureddine@yahoo.fr
Oumayma HTIWICH
Higher Institute of Computer
Science and Multimedia of
Gabes, Tunisia
htiwichoumayma@gmail.com
Med Salim BOUHLEL
National Engineering School of
Sfax , Tunisia
R.U. SETIT.
Higher Institute
of Biotechnology of sfax,
Tunisia
medsalim.bouhlel@enis.rnu.tn
Abstract
Human machine interaction becomes one of the most
research topics in multimedia processing, traditional techniques
for communication are developed in order to tackle technology
advances and allow disable person to communicate easily with
the machine, and to understand their activity using computer
computing. In this paper we are focused on human behavior
analysis from video scene and it is worth noticed that many
information are hidden behind gesture, sudden motion and
walking speed, many research works tried to model and then
recognize human behavior through motion analysis. In our
work we will explain the human action recognition By K
Nearest Neighbors approach.
Keywords
Background Subtraction; Motion tracking; Human
action recognition; K Nearest Neighbors
I. INTRODUCTION
The analysis of human behavior is an important area of
research in computer vision dedicated to the detection,
monitoring and understanding of the physical behavior of
people.
The applications that consider behavior analysis are
embedded in many systems such as smart video surveillance
[1] for automatic control inputs and outputs of certain
objects, the identification and recognition of persons,
detection of unusual behavior, virtual reality systems [2],
human-machine interaction (HMI), augmented reality
systems as virtual systems taking into account the behavior
of people are the most used [3] [4][5].
Steps of human behavior analysis from video are depicted in
fig.1
Fig. 1. Human Behavior Analysis Phases in a Video Sequence
In this article we will explain the third stage of human
behavior analysis process it human action recognition, the
first and the second stages are covered in another article that
I used the Gaussian Mixture Model algorithm [6] [7] [8]for
human detection and Kalman filter for moving human
tracking , by K Nearest Neighbors method
II. BACKGROUND SUBTRACTION BY GMM
METHOD
The background subtraction is widely used in video
processing. It simplifies subsequent processing by locating
regions of interest in the image.
Many methods[9] [10] are developed in this context such as
ViBe methods ("Background VIsual Extractor")[6] KDE
(Kernel Density Estimation) [7] and the temporal averaging
filter [8], those methods are noise sensitive (non-stationary
medium, climate change), motion changes ( camera non
stability).
We propose in our work, a more sophisticated statistical
model, which is Gaussian Mixture Model (GMM).
GMM is a mixture of K Gaussian distributions which
determines the change of state of the corresponding pixels of
an image to another. The algorithm is applied to each image
and transforms the images once stained binary images. By
assigning a value 0 (black) for the background, and the value
1 (white) for the foreground
Fig. 2. Flowchart of the GMM method applied to a video
human
detection
human
tracking
human action
recognition
Video sequence
III. MOVING HUMAN TRACKING BY KALMAN
FILTER METHOD
Objects tracking task is frequently encountered by a
computer vision, and the people tracking literature is
abundant. The objectives of tracking objects are to determine
the trajectories of these in the image plan and assign each
object in the scene a consistent label timely. The monitoring
of human beings is a difficult task for several reasons.
The people followed may have complex motions difficult to
predict. The human body is very articulate; many eclipses
can occur (the person Itself, from other moving objects or by
objects in the background). The illumination of the scene
changes may result in a non-consistency of pixel values
representing a person [11] [12].
The human tracking method depends directly on the
representation of the person[13].Some constraints can be
found here likes distance, the regularity of speed, rigidity (the
movements of the closest points should be similar),
smoothness of motion of the points of the same object.
Kalman filter [13] is the method that reduces those
constraints.
Kalman filter progresses cyclically in two phases: prediction
and correction. The prediction phase is to produce an
estimate of the current state using the previous state. Our goal
is to get a more accurate estimate.
The state and observation equations are given by the
following system:
The human motion tracking is explained in the following
figure
Fig. 3. Kalman Filtering Flowchart
IV. CLASSIFICATION OF MOVING HUMAN
USING KNN (K Nearest Neighbors)
To recognize human actions we must determine the feature
vector of each video. The feature vector is the trajectory
tracking by a person, it is determined, in the first and second
step of human action analysis, by the Kalman filter after the
background subtraction by GMM method.
In this step we will explain the K Nearest Neighbors
[14][15]. The k-NN method is a lazy algorithm unlike many
other machines learning methods such as artificial neural
networks, kernel methods, wavelet networks etc, it has not a
phase of parameters determining of a function by bias of a
mathematical optimization.
K-NN classifier determines the class of a new object by
assigning the majority class of the k objects most like him in
the learning base.
The use of K-NN method requires a learning base, a test
Base, an integer K and a metric for the proximity. The k-NN
algorithm is detailed in the following
Algorithm of K Nearest Neighbors
Input : =

…


…

are
the training data which each row represents
a feature vector of a video

…
are the training data classes (actions)
whose every class present a human action

…
feature vector of test video
For ←1 to M do
For ←1 to N do
/* create a vector D and calculate the
Euclidean distance */
D
i
sum (sqrt(pow((Y-X
i
),2))
end
end
- Sort the vector d ascending
- Take the indexes ,in the training data, of the K
first minimum distance of a vector D
- Recover the K class (action) of the K indices
and put them in a vector T
/* Count the number of occurrence of each
class in T */
/* the video will be classified in
the class that has the max number of occurrence
*/
For ←1 to q do
For ←1 to K do
If (Z
i
= T
j
)
/* nb initialized to 0 */
nbnb+1
End
/* create vector P */
P
i
nb
End
/* recover the max and its index in P */
[a,i]←Max (P)
The action of test video Y is Z
i
• initialise the position of the object
to x0.
• initialise the error tolerance
weight ( 0= 1)
1-Initialisation
(k = 0)
• Use the Kalman filter to predict
the object position such
is considered research center to
find the object
2. Prediction
(k> 0)
• locate the object (which is in the
poit predicted in the previous
step of ) (neighborhood) and
its actual position is used
(measurement) for making the
correction of the state using the
Kalman filter to find this way
3. Correction
(k> 0)
V. EXPERIMENTAL RESULTS
A. Overview of the base video
During the realization of our work we used the basis of KTH
human action videos. It contains 600 movies. The basic data
KTH gathers six different types of individual human actions
(running, jogging, walking, boxing, hand clapping and
handwaving) made repeatedly by 25 people in four different
scenarios: outside , outside with variation of scale, outside
with different clothes and inside. This database also contains
videos in two different environments such as external
environment and another inside. Videos footage of the KTH
base were taken with a homogeneous background and static
camera 25fps. Base KTH people wear different outfits.
Fig. 4. Exemples of the actions KTH basis
B. Detection of moving objects by GMM
Background subtraction by GMM is a popular method for
segmentation object in the scenes video. A Gaussian mixture
treats the multimodality background caused by the shadow,
repetitive movement of objects such as moving leaves.
Fig. 5. Detection running human
C. Tracking by Kalman filter
Tracking moving object is the step that precedes the
background subtraction plan. In this step the trajectory
treated represents the movement object. We chose the
jogging action as an example and try to act on the state
transition matrix.
The trajectory of human motion is presented by the two
curves: the red color is the measured path and the green color
is the estimated trajectory.
Fig. 6. Tracking running human
D. Classification by KNN
After finding the human moving tracking which represent the
vector features of any scene of human action we will analyze
the human behavior by KNN method.
In order to evaluate the performance of our method, we focus
on good classification rate much as using the algorithm
of K Nearest Neighbor (KNN) with K = 10.
Rate of recognition
=


100
The results of the classification are presented in the fig. 7.
Fig. 7. The rate of human action recognition by KNN method
Recognition of human action by KNN algorithm achieved a
rate of recognition 71.1%. To evaluate our result we will
compare it with other methods most used in the literature
such as the method of SVM (Support vector Machine) [13]
and naïve bayes[14], all these methods use the same base of
the human action the KTH basis. The measurement of
performance of each method is expressed by the rate of good
classification .in effect the SVM rate of recognition in 66%
by naïve bayes rate of recognition is 61%.
The fig.8. shows the results of recognition rate.
Fig. 8. Comparaison KNN , SVM and Naïve Bayes
-50 050 100 150 200
-40
-20
0
20
40
60
80
100
0
20
40
60
80
1
KNN 71.1
71,1
The Rate of Recognition
KNN
55
60
65
70
75
KNN SVM Naive
Bayes
71.1
66
61
The Rate of Recognition
KNN
SVM
Naive Bayes
Running Walking Jogging
HandClapping Boxing HandWaving
VI. CONCLUSION
In this paper, we describe the steps to analyze the behavior of
a human individual in a video scene by the recognition of
actions. Especially we have focused on the classification of
videos, in this part we used the classification by KNN metho-
d on the KTH basis of videos. We have obtained a higher
KNN classification than SVM and Naïve Bayes[16][17]
methods.
REFERENCES
[1] Marco Cristani ,R.Raghavendra , Alessio Del Bue Vittorio
Murino“Human behavior analysis in videosurveillance”: A Social
Signal Processing perspective 2012.Neurocomputing 100 (2013)
86–97
[2] Dimitris Metaxas,Shaoting Zhang. “A review of motion analysis
methods for human Non verbal Communication Computing.Image
and Vision Computing 31 (2013) 421–433
[3] Hao-Cheng Mo, Jin-Jang Leou, and Cheng-Shian Lin“Human
Behavior Analysis Using Multiple 2D Features and Multicategory
Support Vector Machine.MVA 2009 IAPR Conference on
Machine Vision Applications, May 20-22, 2009, Yokohama,
JAPAN.
[4] Md. Atiqur Rahman Ahad, JooKooi Tan, Hyoung Seop Kim and Seiji
Ishikawa. “Analysis of Motion Self-Occlusion Problem Due to
Motion Overwriting for Human Activity Recognition”. Journal of
Multimedia, vol. 5, no. 1, February 2010.
[5] Nesrine TRIKI, Mohamed KALLEL, Mohamed Salim
BOUHLEL“Imaging and HMI :Fondations and
complementarities”SETIT ( Sciences of Electronics, Technologies
of Information and Telecommunications) march 2012 Tunisia
[6] Z. Zivkovic and F. vander Heijden. “Efficient adaptive density
estimation per image pixel for the task of background subtraction.
Pattern recognition letters ”,27 :773_780, 2006.
[7] C. Wren, A. Azarbayejani, T. Darrell, and A.Pentland.Pfinder :”Real-
time tracking of the human body. Pattern Analysis and Machine
Intelligence”, 1997.
[8] Benjamin Langmann, Seyed E. Ghobadi, Klaus Hartmann, Otmar
Loffeld“Multi-Modal Background Subtraction Using Gaussian
Mixture Models”In: Paparoditis N., Pierrot-Deseilligny M., Mallet
C., Tournaire O. (Eds), IAPRS, Vol. XXXVIII, Part 3A – Saint-
Mandé, France, September 1-3, 2010
[9] Adil CHERGUI, Wafae SABBAR, Abdelkrim BEKKHOUCHA Video
Scene Segmentation Using the Shot Transition Detection by Local
Characterization of the Points of Interest SETIT ( Sciences of
Electronics, Technologies of Information and Telecommunications)
march 2012 Tunisia
[10] Jin MA, Fuqing DUAN, Ping GUO Improvement of Texture Image
Segmentation Based on Visual ModelSETIT( Sciences of
Electronics, Technologies of Information and Telecommunications)
march 2012 Tunisia
[11] Houcine ESSID, Ali BEN ABBES, Imed Riadh FARAH, Vincent
BARRA Spatio-temporal modeling based on Hidden Markov Model
for Object Tracking in Satellite Imagery SETIT( Sciences of
Electronics, Technologies of Information and Telecommunications)
march 2012 Tunisia
[12] D.Harihara Santosh,P.Venkatesh,P. Poornesh, L. Narayana Rao, N.
Arun Kumar may«Tracking Multiple Moving Objects Using
Gaussian Mixture Model»International Journal of Soft Computing
and Engineering(IJSCE) ISSN: 2231-2307, Volume-3, Issue-2, May
2013
[13] Houcine ESSID, Ali BEN ABBES, Imed Riadh FARAH, Vincent
BARRA Spatio-temporal modeling based on Hidden Markov Model
for Object Tracking in Satellite Imagery SETIT( Sciences of
Electronics, Technologies of Information and Telecommunications)
march 2012 Tunisia
[14] Maxime Devanne, Hazem Wannous, Stefano Berretti “3-D Human
Action Recognition by Shape Analysis of Motion Trajectories on
Riemannian Manifold” IEEE TRANSACTIONS ON
CYBERNETICS, VOL. 45, NO. 7, JULY 2015
[15] Salim Al-Ali, Mariofanna Milanova, Agata Manolova, and Victoria
Fox“Human action recognition using combined contour-based and
silhouette-based features and employing KNN or SVM classifier”
International JOURNAL OF COMPUTERS Volume 9, 2015
[16] Vladimir N. Vapnik. “Statistical learning theory” Septembre 1998.
[17] David D. Lewis. “Naive (bayes) at forty” The independence
assumption in information retrieval. pages 4–15, 1998.
... Action detection is one of the fields of research that has recently attracted a lot of attention. 36 Hand movement is an example of an action level. One of the challenges in this field is the action detection with a high degree of ambiguity in belonging to a particular category. ...
... 37 Human-computer interaction has become one of the research topics in video processing. 36,38,39 The HBA phases of video include human detection, human tracking, and human action detection, respectively. 38 A review of HBA levels is presented in Table 1 and their challenges are presented in Table 2. ...
Article
Full-text available
In recent years, understanding human behavior has become one of the most important topics in the field of computer vision research. The reason for this growing attention is the wide range of applications that can benefit from the results of this field of research. The human behavior analysis (HBA) includes a wide range of research areas from the detection of human motion and action. Datasets created through actions and detection of human activities make it possible to compare different detection methods with the same input data. Data mining and big data approaches are very popular for analyzing data related to human behavior and can be used to address the challenges of fast processing. This article provides a systematic survey of data mining and big data in the HBA. We focus on current datasets and models related to the detection of human behavior patterns in the literature. The purpose of this survey is to assist researchers in select appropriate datasets and models to evaluate algorithms, as well as to identify research gaps for future work. To achieve this goal, articles published between 2010 and 2021 have been reviewed. These articles fall into five general categories in terms of dataset focus: object detection, motion, action, activity, and behavior. This article provides a summary of data mining and big data models in the HBA, as well as related datasets based on these categories, to highlight promising research avenues for future work.
... Recent development in action recognition have opened up a wide range of real-world applications: visual security systems [19,51,57], behavioral studies [35], sports analytics [47], elderly person fall detection systems [10,50,86], etc. Most of these developments are mainly courtesy of large-scale curated datasets like Kinetics [12], HVU [21], and HACS [88]. ...
Preprint
Full-text available
Semi-Supervised Learning can be more beneficial for the video domain compared to images because of its higher annotation cost and dimensionality. Besides, any video understanding task requires reasoning over both spatial and temporal dimensions. In order to learn both the static and motion related features for the semi-supervised action recognition task, existing methods rely on hard input inductive biases like using two-modalities (RGB and Optical-flow) or two-stream of different playback rates. Instead of utilizing unlabeled videos through diverse input streams, we rely on self-supervised video representations, particularly, we utilize temporally-invariant and temporally-distinctive representations. We observe that these representations complement each other depending on the nature of the action. Based on this observation, we propose a student-teacher semi-supervised learning framework, TimeBalance, where we distill the knowledge from a temporally-invariant and a temporally-distinctive teacher. Depending on the nature of the unlabeled video, we dynamically combine the knowledge of these two teachers based on a novel temporal similarity-based reweighting scheme. Our method achieves state-of-the-art performance on three action recognition benchmarks: UCF101, HMDB51, and Kinetics400. Code: https://github.com/DAVEISHAN/TimeBalance
... Penelitian (Shahroudy, Liu, Ng & Wang, 2016) berhasil memperoleh dataset yang terdiri dari 56.880 sampel video RGB + D. Data tersebut diperoleh dari 40 subjek manusia yang berbeda dengan bantuan Microsoft Kinect v2 untuk proses pengambilan data. Selain itu juga terdapat penelitian (Jaouedi, Boujnah, Htiwich & Bouhlel, 2016) yang menggunakan video dari dataset KTH (www.csc.kth.se/cvap/actions/, 2020) yang berjumlah 600 video. ...
Article
Full-text available
p>Pengenalan aktivitas manusia atau Human Activity Recognition (HAR) merupakan salah satu topik yang populer karena besarnya peluang untuk diterapkan di kehidupan sehari-hari. Tujuan dari pengenalan ini adalah untuk mengenali, mendeteksi, dan mengklasifikasikan aktivitas yang dilakukan manusia. Pengenalan aktivitas manusia adalah salah satu teknologi penting untuk memantau dinamisme seseorang sehingga dapat bermanfaat di berbagai hal. Selain untuk menjaga kesehatan, pencegahan penyakit, dan membantu menentukan jenis olah raga, HAR dapat dimanfaatkan juga untuk diterapkan pada bidang keamanan dan pengembangan teknologi. Penelitian ini menggunakan smartphone sebagai teknologi utama dalam memperoleh data dengan memanfaatkan sensor akselerometer dan giroskop yang telah tertanam di dalamnya. Terdapat 8 macam aktivitas yang diteliti dengan kombinasi lama waktu eksperimen 5, 10, dan 15 detik serta posisi smartphone dipegang bebas maupun di dalam saku celana kanan. Data yang diperoleh terdiri dari 3 sumbu (x, y, dan z) pada setiap sensor yang digunakan. Data tersebut kemudian melalui proses pengolahan dan klasifikasi menggunakan algoritme k - Nearest Neighbor ( k -NN). Hasil akurasi yang didapat dalam penelitian ini mencapai 79,56%. Hasil yang diperoleh melalui penelitian ini menunjukkan bahwa perbedaan peletakan smartphone mempengaruhi hasil pengenalan aktivitas manusia secara stabil. Selain itu, perbedaan jumlah data akibat perbedaan lamanya waktu eksperimen dapat mengakibatkan perbedaan lamanya waktu komputasi. Penelitian ini menjadi penting karena hasilnya dapat menjadi batu loncatan bagi penelitian selanjutnya. Beberapa peluang pengembangan juga dilampirkan pada bagian akhir. Abstract Human activity recognition (HAR) is one of the most popular topics because of the large opportunities for its application in life. The purpose of HAR is to recognize, detect and classify human activities. Human activity recognition is one of the important technologies for monitoring a person's dynamism so that it can be utilized in various ways. Apart from maintaining health, preventing disease, and helping determine the type of exercise, HAR can also be used to be applied in the field of security and technological developments. This study uses smartphones as the main technology in obtaining data by utilizing the built-in accelerometer and gyroscope sensors. There are 8 types of activities studied with a combination of 5, 10, and 15 seconds of experimental time and the position of the smartphone is carried freely or in the right trouser pocket. The data obtained consists of 3 axes (x, y, and z) on each sensor used. The data then processed and classified using the k-Nearest Neighbor (k-NN) algorithm. The accuracy results obtained in this study reaches 79.56%. The results obtained through this study indicate that differences in smartphone placement affect the results of human activity recognition stably. In addition, differences in the amount of data due to differences in the length of the experiment period can result in differences in the length of computation time. This research is important because the results can be used as material for further research assistance. Some development opportunities are also attached at the end. </p
... The K nearest neighbors (K-NN), [24] , is one of the simplest methods for classification. K-NN is a supervised learning algorithm in which the result of a new object is classified according to the majority of the categories K Nearest neighbors in the sense of a distance chosen on the basis of sets of the training data. ...
... Meanwhile, the KNN algorithm may also be utilized in observing human behavior. Recognition of human behavior through motion analysis has shown that the KNN algorithm can achieve a higher percentage compared to SVM and Naive Bayes (Jaouedi et al, 2016). Despite this, KNN is not scalable because its execution requires considerable processing time and the size of the databases continually expands (Bobadilla et al., 2013). ...
Article
The Bataan Nuclear Power Plant (BNPP) is one of the many decommissioned Nuclear Power Plant (NPP) globally and its reopening has led to different perceptions among Filipinos. It was established in 1984 but was not utilized due to political liability and safety reasons. This study aimed to predict factors affecting the acceptance of the BNPP by utilizing Machine Learning Algorithms (MLA). The MLAs utilized in this study were Decision Tree, Random Forest Classifier (RFC), and Artificial Neural Network (ANN) as a highlight to predict human behavior. 1,252 Filipinos voluntarily answered an online questionnaire which consist of 37 questions, leading to 46,324 datasets. MLA showed that Filipinos are knowledgeable about the benefits of NPPs, leading to the acceptance of the reopening of the BNPP. In addition, MLA indicated that perceived benefits (PB) was found to be the highest factor that affect the Filipino’s acceptance of the reopening of BNPP. Job opportunities, economic growth, lower and clean energy consumption, and sustainability were the indicators for the acceptance of the reopening of BNPP. Interestingly, the result showed that PB relatively outweighed the perceived risk of the BNPP. ANN and RFC proved to be effective with accuracy rates of 93.44% and 97.00%, respectively. Finally, the MLA approach in this study can be applied and extended in predicting the acceptance of NPPs worldwide.
Chapter
In recent years, interest in real-time video sequences for human action recognition has risen rapidly, and discrimination is required to identify which type of human action appears in the videos. The use of video for human action recognition is becoming widespread in many applications like monitoring of traffic, pedestrian detection, anomalous behavior identification of the human being, health monitoring, human–computer interaction, robotics, etc. Various combinations of feature representation are used to implement different feature detectors for the identification of features and to classify human actions. For testing the given data, Machine learning-based techniques such as unsupervised and supervised learning are utilized to estimate the task of identifying human behaviors. For classification, unsupervised learning techniques like data clustering, hierarchical learning, etc., are implemented, whereas to handle action classification, supervised learning techniques like SVM, random forest, KNN, Naive–Bayes, and artificial neural networks are used. In this paper, real-time applications for suspicious action detection, tracking, and recognition are surveyed and classified. We propose a new technology like MediaPipe Holistic which provides pose, face, and hand mark detection models where the suspicious movement of the human being is being detected by the system from a video that may be stored in memory or from real-time monitoring.KeywordsFeature detectorsMachine learningDeep learningSVMCNNKalman filteringMediaPipe
Article
Full-text available
Human communication entails subtle non-verbal modes of expression, which can be analyzed quantitatively using computational approaches and thus support human sciences. In this paper we present huSync, a computational framework and system that utilizes trajectory information extracted using pose estimation algorithms from video sequences to quantify synchronization between individuals in small groups. The system is exploited to study interpersonal coordination in musical ensembles. Musicians communicate with each other through sounds and gestures, providing nonverbal cues that regulate interpersonal coordination. huSync was applied to recordings of concert performances by a professional instrumental ensemble playing two musical pieces. We examined effects of different aspects of musical structure (texture and phrase position) on interpersonal synchronization, which was quantified by computing phase locking values of head motion for all possible within-group pairs. Results indicate that interpersonal coupling was stronger for polyphonic textures (ambiguous leadership) than homophonic textures (clear melodic leader), and this difference was greater in early portions of phrases than endings (where coordination demands are highest). Results were cross-validated against an analysis of audio features, showing links between phase locking values and event density. This research produced a system, huSync, that can quantify synchronization in small groups and is sensitive to dynamic modulations of interpersonal coupling related to ambiguity in leadership and coordination demands, in standard video recordings of naturalistic human group interaction. huSync enabled a better understanding of the relationship between interpersonal coupling and musical structure, thus enhancing collaborations between human and computer scientists.
Article
In today's world on technological know- How AI is becoming an engulfing quantity in our lives. Human-machine interaction has grown significantly throughout the course of history. As a reaction, users began to express themselves on social media. Along with it, human-computer connection has sprung rapid answers to help enhance our society. With it proposed tool between that demand bills we bear added modified course of the usage of word embedding in imitation of convert textual content between numeric structures by using keeping its context. To study behaviour sample on individual person including Word2vec mannequin mixed including Machine instruction algorithm in conformity with assist observe modifications between conduct over a character over duration regarding time. This lookup mission choice be over immense advantage to Psychologist/Psychiatrist or assist governance as like it analyzes yet monitor a person's mental fitness yet choice additionally help among classifying the regular customers beside the non-normal users at early degrees in imitation of shortlist the victims about psychological trouble in accordance with help focus regimen agency because of their treatment. These enterprise do reach abroad these human beings whoever do not want after visit a psychiatrist due to the fact it advise it as much half sort concerning a taboo.
Article
Full-text available
This paper presents a new algorithm for human action recognition in videos. This algorithm is based on a combination of two different feature types extracted from Aligned Motion Images (AMIs). The AMI is a method for capturing the motion of all frames in a human action video in one image. The first feature is a contour-based type and is employed to grasp boundary details of the AMI. It relies on the 1st and 2nd discrete time differential of the chord-distance signature feature, so it is called Derivatives of Chord-Distance Signature (DCDS). The second feature is a silhouette-based type that is used to capture regional appearance details. It catches most of the visual components for the AMI using a Histogram of Oriented Gradients (HOG) feature. Combining both features creates a complementary feature vector that makes it possible to obtain an optimal correct recognition rate of 100%. For the classification, the algorithm is utilized two different classifiers: K-Nearest-Neighbor (KNN) and Support Vector Machine (SVM). The KNN is based on the 1st norm distance and achieves slightly better results than this obtained by SVM. The performance of the algorithm is tested through six experiments. Three experiments for the KNN classifier and others for the SVM. For each classifier, three experiments conducted to determine the effectiveness of each feature separately and when combined. The experimental results demonstrate the potential power of this algorithm and its promising success in human action recognition in videos.
Article
Full-text available
Recognizing human actions in 3D video sequences is an important open problem that is currently at the heart of many research domains including surveillance, natural interfaces and rehabilitation. However, the design and development of models for action recognition that are both accurate and efficient is a challenging task due to the variability of the human pose, clothing and appearance. In this paper, we propose a new framework to extract a compact representation of a human action captured through a depth sensor, and enable accurate action recognition. The proposed solution develops on fitting a human skeleton model to acquired data so as to represent the 3D coordinates of the joints and their change over time as a trajectory in a suitable \textit{action space}. Thanks to such a 3D joint-based framework, the proposed solution is capable to capture both the shape and the dynamics of the human body simultaneously. The action recognition problem is then formulated as the problem of computing the similarity between the shape of trajectories in a Riemannian manifold. Classification using $k$NN is finally performed on this manifold taking advantage of Riemannian geometry in the open curve shape space. Experiments are carried out on four representative benchmarks to demonstrate the potential of the proposed solution in terms of accuracy/latency for a low-latency action recognition. Comparative results with state-of-the-art methods are reported.
Article
Full-text available
In this study, a human behavior analysis system using multiple 2D (two-dimensional) features and a multicate-gory support vector machine is proposed. In the proposed system, three kinds of features, namely, human star skeleton, angles of six sticks in the star skeleton, and object motion vectors, are employed to train the human posture classifier and recognize human postures. Based on the recognized human postures, a backward search strategy is proposed to recognize human actions. Based on the experimental results obtained in this study, in terms of recall and precision rates, the proposed system has good performance and is superior to the comparison system.
Article
Background subtraction is a common first step in the field of video processing and it is used to reduce the effective image size in subsequent processing steps by segmenting the mostly static background from the moving or changing foreground. In this paper previous approaches towards background modeling are extended to handle videos accompanied by information gained from a novel 2D/3D camera. This camera contains a color and a PMD chip which operates on the Time-of-Flight operating principle. The background is estimated using the widely spread Gaussian mixture model in color as well as in depth and amplitude modulation. A new matching function is presented that allows for better treatment of shadows and noise and reduces block artifacts. Problems and limitations to overcome the problem of fusing high resolution color information with low resolution depth data are addressed and the approach is tested with different parameters on several scenes and the results are compared to common and widely accepted methods.
Conference Paper
Nowadays, the new technologies of image processing and Human Machine Interaction (HMI) are mostly linked to the scientific domains of research (medicine, transport, etc). Indeed, the emergence of the tactile interfaces as well as the gestural interfaces allows increasing the physical world which surrounds us with the digital information. It also allows us to use natural hand gestures to interact with this information which is a digital image. In this paper, we describe the basic concepts of the images processing as well as the HMI foundations. Then, we identify the gestural human interaction as well as its relation with the imaging.
Conference Paper
The time series of satellite images are an important source of information for monitoring spatiotemporal changes of land surfaces. Furthermore, the number of satellite images is increasing constantly, for taking full advantage, tools dedicated to the automatic processing of information content is developed. However these techniques do not completely satisfy the geographers who exploit more currently, the data extracted from the images in their studies to predict the future. In this research we propose a generic methodology based on a hidden Markov model for analyzing and predicting changes in a sequence of satellite images. The methodology that is proposed presents two modules: a processing module which incorporating descriptors and algorithms conventionally used in image 'interpretation and a learning module based on hidden Markov models. The performance of the approach is evaluated by trials of interpretation of spatiotemporal events conducted in several study sites. Results obtained allow us to analyze and to predict changes from various time series of SPOT images for observation of spatiotemporal events such as urban development. It is thus quite reasonable to use this methodology to follow the evolution of other phenomena and to predict their future states.
Article
Human Nonverbal Communication Computing aims to investigate how people exploit nonverbal aspects of their communication to coordinate their activities and social relationships. Nonverbal behavior plays important roles in message production and processing, relational communication, social interaction and networks, deception and impression management, and emotional expression. This is a fundamental yet challenging research topic. To effectively analyze Nonverbal Communication Computing, motion analysis methods have been widely investigated and employed. In this paper, we introduce the concept and applications of Nonverbal Communication Computing and also review some of the motion analysis methods employed in this area. They include face tracking, expression recognition, body reconstruction, and group activity analysis. In addition, we also discuss some open problems and the future directions of this area.
Article
The analysis of human activities is one of the most intriguing and important open issues for the automated video surveillance community. Since few years ago, it has been handled following a mere Computer Vision and Pattern Recognition perspective, where an activity corresponded to a temporal sequence of explicit actions (run, stop, sit, walk, etc.). Even under this simplistic assumption, the issue is hard, due to the strong diversity of the people appearance, the number of individuals considered (we may monitor single individuals, groups, crowd), the variability of the environmental conditions (indoor/outdoor, different weather conditions), and the kinds of sensors employed. More recently, the automated surveillance of human activities has been faced considering a new perspective, that brings in notions and principles from the social, affective, and psychological literature, and that is called Social Signal Processing (SSP). SSP employs primarily nonverbal cues, most of them are outside of conscious awareness, like face expressions and gazing, body posture and gestures, vocal characteristics, relative distances in the space and the like. This paper is the first review analyzing this new trend, proposing a structured snapshot of the state of the art and envisaging novel challenges in the surveillance domain where the cross-pollination of Computer Science technologies and Sociology theories may offer valid investigation strategies.