Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Actual Pathogen Detection: Sensors and Algorithms - a Review
Next Article in Special Issue
Combining Multiple Algorithms for Road Network Tracking from Multiple Source Remotely Sensed Imagery: a Practical System and Performance Evaluation
Previous Article in Journal / Special Issue
Design of Sensor Networks for Chemical Plants Based on Meta-Heuristics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recognizing Human Activities from Sensors Using Hidden Markov Models Constructed by Feature Selection Techniques

Computer Science Department, Universidad Carlos III de Madrid, Avda. de la Universidad Carlos III, 22, Colmenarejo, Spain
*
Author to whom correspondence should be addressed.
Algorithms 2009, 2(1), 282-300; https://doi.org/10.3390/a2010282
Submission received: 28 November 2008 / Revised: 2 February 2009 / Accepted: 16 February 2009 / Published: 21 February 2009
(This article belongs to the Special Issue Sensor Algorithms)

Abstract

:
In this paper a method for selecting features for Human Activity Recognition from sensors is presented. Using a large feature set that contains features that may describe the activities to recognize, Best First Search and Genetic Algorithms are employed to select the feature subset that maximizes the accuracy of a Hidden Markov Model generated from the subset. A comparative of the proposed techniques is presented to demonstrate their performance building Hidden Markov Models to classify different human activities using video sensors

1. Introduction

Sensors allow computers to perceive the world that surrounds them. By the use of sensors, like thermostats or photocells, computers can meassure the temperature or lighting conditions of a room. In the last years,the deployment of multisensor networks has become popular due to the cut down on sensor prizes. This networks can include different kind of sensors, maybe more complex, like cameras, indoor location systems (ILS), microphones, etc …By the use of sensor networks computer systems can take more accurate decissions due to the richer information about the environment that they have.
A common task in sensor processing is the recognition of situations in the environment perceived. If these situations to recognize are known a priori, i.e. there is a set of labels that describe each situation, the problem can be tackled as a classification task. Then, the problem to solve is to find a model relating the data from sensor readings with situation labels. Some sensor data may be too noisy, and other not provide any information for label prediction. Relevant sensor data has to be identified to build the model from them, not including irrelevant data.
An application where the use of sensor networks can be useful is Human Activity Recognition, i.e., the understanding by the computer of what humans are doing. This field has received increasing attention in the last years, due to their promising applications that have in surveillance, human computer interaction or ambient intelligence, and the interests that governments and commercial organizations have placed in the area. Human Activity Recognition systems can be integrated with existing systems as the proposed by Corchado et al. [1] to monitor alzehimer patients. If a patient falls to the floor, the system can alert a nurse to attend the accident. Human Activity Recognition systems can also be integrated with the system proposed by Pavón et al. [2], detecting forbidden activities being performed and alerting security staff.
Human activity recognition may be considered as a classification task, and human activity recognizers can be created using supervised learning. A set of activities to recognized has to be defined a priori. Different observations about the activities to recognize are extracted using different sensors. Then, the problem to solve is to find the function that best relates observations to activity labels. Data from sensors is not free from noise, so relevant attributes have to be identified before training the activity recognizer. Better results are expected to be obtained when this previous step is performed.
The sensors that provides the most information for Human Activity Recognition are video cameras. Works in activity recognition from video could be divided in two groups [3]: (1) those that are centered in small duration activities (i.e. walking, running,…); and (2) those that deal with large duration activities (i.e. leaving place, going to living room,…). The former are centered in choosing good features for activity recognition, whereas the latter usually tackle the temporal structure in the classifier.
Regarding small duration activities, Neural networks, Neuro-fuzzy systems and C4.5 have been succesfully employed [4]. Also, time delay neural networks have been used to recognize the case where the activities are hand gestures [5]. In [6] is shown how to perform feature extraction, feature selection and classification using a simple bayesian classifier to solve the small duration activity recognition problem. In their approach, they use ground truth data from CAVIAR* dataset of people bounding boxes instead of a blob tracker output. Perez et al. [7] use different time averaged speed measures to classify the activities present in the CAVIAR dataset using HMM.
Robertson et al. [3] uses trajectory and velocity concatenated data for five frames, and blob optical flow, to decide what is the current small duration activity. This small duration activity is then introduced in a Hidden Markov Model (HMM) to decide which is the small duration activity performed.
To classify large duration activities, Brand et al. [8] have used HMMs for modelling activities performed in an office. They use different spatial meassures taken from the blob bounding ellipse.The HMM approach has been extended in [9] to model activities involving two agents, using a Coupled-HMM. They use different velocity meassures as features, being all of them invariant with respect to camera rotations and translations, providing camera independent recognizers.
HMMs are recognized as one effective technique for activity classification, because they offer dynamic time warping, have clear bayesian semantics and well-understood training algorithms. In this paper a method to adapt them, selecting the best features in the observation space, is defined, trying to maximize its accuracy for classifying short durative activities. HMMs will be incrementally built using both heuristic search and genetic algorithms. A comparative of the performance obtained using these algorithm for HMM construction will be shown. Our approach differs from the proposed by Ribeiro et al. in some aspects: (1) blob tracker output is used instead of ground truth data; (2) the foreground mask of the blob tracker is used to extract features; (3) the classifier used for the recognition of activities is an HMM; (4) we use different temporal window sizes; and (4) we use different search methods for feature selection.
This paper is organized as follows: on section 2, an overview of the activity recognition system where the feature selection is going to be performed is given; on section 3, the feature selection algorithms that are used to build HMMs are presented; on section 4, experiments selecting good features for human activity recognition are performed and results are discussed; finally, on section 5 conclusions of this work are presented and future lines are discused.

2. Definition of Activity Recognition problem

2.1. Functional description

Figure 1. Overview of the activity recognition process from multiple sensors
Figure 1. Overview of the activity recognition process from multiple sensors
Algorithms 02 00282 g001
In the proposed human activity recognition architecture (see Figure 1), the objective is to select from a set of activity labels A = { a 1 , , a N } the one that is the most likely according to the signal p t retrieved by the sensors at each time step. Sensors could be video cameras, thermal cameras or indoor localization systems, depending of the available resources and the conditions of the environment. In case of imaging sensors, they get a frame on each time step, p t I t , x , y . Using p t , the system needs to extract the position and size b i t of each human i in the environment. When using imaging sensors, b i t may be obtained using a blob tracker [10]. Blob tracker takes an image sequence and provides the location of each object i moving in the scene, mantaining a temporal coherence of objects between frames. Human tracking is a hard task and, despite of the advances in the last years, tracking methods still provide a noisy output, caused in part by scene illumination changes or by the complexity of object movements [11]. Using p t and b i t , a set of additional features ζ i t may have to be extracted to be able to differenciate between the activities to be recognized. Finally, using ζ i t , a function Λ ζ i t , Λ : ζ A is used to decide wich is the activity being performed.
Activity recognition solves the correspondence:
q i t = arg max a k P ζ i t , ζ i t 1 , ζ i 1 a k
where q i t is the activity that maximizes the probability of the observations ζ i t , ζ i t 1 ,…, ζ i 0 at instant t by individual i.
Figure 2. The three steps to create an activity recognizer
Figure 2. The three steps to create an activity recognizer
Algorithms 02 00282 g002
To build an activity recognition system, three processes have to be performed (see Figure 2). First, an extensive feature set Z i t has to be extracted from b i t and I t , x , y . On a second step, the subset ζ i t has to be selected from Z t , selecting the subset with the most predictive power for A. The last problem to solve is to select the best classifier architecture for activity recognition.

2.2. Feature extraction for activity representation using video sensors

The features that can be used for activity representation depend of the type of sensor inputs to be processed. The most common sensor used for Human Activity recognition are video cameras, so all the features presented here are computed from image sequences.
The objective of this section is to present a large set of features to characterize the activities to classify as complete as possible. The input to the feature extraction process includes the original frame I t , x , y , its foreground mask F g t , x , y , and the blob bounding box b t given by a blob tracker, that represents the human (see Figure 3). Using these information, different type of features are going to be derived. Most of the features presented here have been proposed in [6]. The first group of extracted features comes from the blob bounding box properties and its time evolution, taking first order derivatives and time averaged derivatives, using a temporal window of T frames, as shown on Table 1
Figure 3. Inputs for the feature extraction process
Figure 3. Inputs for the feature extraction process
Algorithms 02 00282 g003
Table 1. First feature group: blob bounding box properties
Table 1. First feature group: blob bounding box properties
#Feature nameDefinition
Bounding box properties at time t
1Position p t = x t , y t
2Size s i z e t = w t , h t
3Velocity s t = p t t = x t t , y t t
4Size derivative s i z e t t = w t t , h t t
5Speed s t = x t t 2 + y t t 2
6Area s t = w t * h t
Properties averaged over T frames
7Mean speed s ¯ T t = 1 T i = t T + 1 T s i
-Mean velocity norm s ¯ T t 3 different methods:
8Averaging vectors s ¯ T t 1 = 1 T i = t T t 1 p t p i t i
9Mean vectors s ¯ T t 2 = p t p t T + 1 T
10Linear fitting s ¯ T t 3 = Linear Least Squares Fitting
11..13speed/velocity ratio R s T , i t = s T t s T t i i = 1 , 2 , 3
Second order moments
14speed σ v , T 2 t 1 = 1 T 1 i = t T + 1 t s 2 i
15..17speed (centered) σ v , T 2 t 1 + j = 1 T 1 i = t T + 1 t s i s ¯ T t j j=1,2,3
-velocity Σ s , T t 1 = 1 T 1 i = t T + 1 t s i s i
-velocity centered Σ s , T t 1 + j = 1 T 1 i = t T + 1 t s i s ¯ T t j s i s ¯ T t j j=1,2,3
18..21trace t r Σ s , T t i = t r a c e Σ v , T t i=1,2,3,4
22..25eigenvalues ratio R Σ s , T t i = λ m i n λ m a x Σ v , T t i=1,2,3,4
Second feature group consist of the seven Hu invariant moments [12] { h u 1 , h u 2 , h u 3 , h u 4 , h u 5 , h u 6 , h u 7 } of the foreground mask enclosed by the blob bounding box. These seven moments are shape descriptors invariant under translation, rotation and scaling. Each Hu moment is numbered from 26 to 32.
Table 2. Third and fourth group: Optical flow measures
Table 2. Third and fourth group: Optical flow measures
# I t , x , y # F g t , x , y Feature nameDefinition
--target pixels P t = ( x , y ) b o u n d i n g b o x t
Instantaneous
--optical flow F t , x , y = f x t , x , y , f y t , x , y , x , y P t
--mean flow F ¯ t 1 = 1 N x , y P t F t , x , y
--mean flow (centered) F ¯ t 2 = F t 1 v t
--flow norm f t , x , y 1 = F t , x , y
--flow norm (centered) f t , x , y 2 = F t , x , y v t
Spatial energy / 2nd order moments
33, 3437, 38motion energy f t i = 1 N x , y P t f 2 t , x , y i i = 1 , 2
--flow cov. Σ F t 1 = 1 N x , y P t F t , x , y F t , x , y
--flow cov. centered Σ F t 2 = 1 N x , y P t F t , x , y F ¯ t 1 F t , x , y F ¯ t 1
35,3639, 40eigenvalues ratio R Σ F t i = λ m i n λ m a x Σ F t i=1,2
Time averaged over T frames
--mean flow F ¯ t i + 2 = F ¯ t i 1 T j = t T + 1 t F ¯ j i i = 1 , 2
4142motion energy f T ¯ T = 1 T i = t T + 1 t f i 1
Temporal energy / 2nd order moments
--mean flow Σ F ¯ t i = 1 T j = t T + 1 t F ¯ j i F ¯ j i i = 1 , 2 , 3 , 4
43…4651…54trace t r Σ F ¯ t i = t r a c e Σ F ¯ t i i = 1 , 2 , 3 , 4
47 …5055…58eigenvalues ratio R Σ F ¯ t i = λ m i n λ m a x Σ F ¯ t i i = 1 , 2 , 3 , 4
Third and Fourth feature groups have been extracted from optical flow meassures from the enclosed blob bounding box of I t , x , y and F g t , x , y . Optical flow is a meassure of the motion of each one of the pixels of a given frame with respect to a previous one. In our work, it is obtained using the Lucas-Kanade method [13]. Again, some properties are taken from the time averaged meassures of different optical flow properties, using a temporal window of size T.

2.3. Feature selection for activity recognition problems

Feature selection could be described as a single objective optimization problem where given a set of features F, a subset f * has to be determined, such:
J f * , E = max f F J f , E
where J : F × Ψ R is an utility function that meassures how good is a feature subset f F for classiffing a set of points E Ψ , being Ψ the space of training examples. In a supervised context, where the class of the points in E is known, the goodness of the subset f could be interpreted as the accuracy of a classifier built using the attributes in f, while in an unsupervised context the goodness of f could be interpreted as the quality of the generated clusters.
The need for feature selections comes from the fact that available training instances usually contain highly correlated and/or noise-dominated attributes. These attributes have to be identified and removed, because they don’t provide useful information for classification and their use only increase the classification cost, even dropping the classification accuracy.
The feature selection problem has been widely studied in the literature. A taxonomy of the existing techniques could be established according to the evaluation criteria used to score the candidate feature subsets [14]: (1) wrapper methods, where a classifier is trained to evaluate the predictive accuracy of the candidate subset; and (2) filter methods, where a meassure of the predictive power of the feature subset is calculated, without training any classifiers. Filter methods are quicker and more general than wrapper methods, but the optimality of the obtained feature subsets could not be ensured even using exhaustive search methods.

2.4. Classifiers for activity recognition

To solve general classification problems, different algorithms have been proposed, supervised and non supervised [15]. Supervised methods build class models using labelled data, whereas non supervised methods try to discover the structure of the data. Examples of supervised methods are kNN, Neural Networks and Classification Trees. Examples of non supervised methods are clustering techniques. In Human Activity Recognition, the most accepted classification techniques are Hidden Markov Models (HMMs), because activities have temporal structure and HMMs can tackle with this kind of processes. However, others methods have been succesfully used [4]: Neural Networks, Neuro-Fuzzy systems, C4.5. etc.
Be a system that can be described using a set of N different states, where random transitions are produced over time, according to a given probability distribution for each state. The state on the system on each moment depends on the state that it was in the previous moments. This kind of stochastic process is called ”Markov Model”. Additionally, if the present state of the system can not be observed, i.e., it could be only measured by an effect that it produces, the system is called ”Hidden Markov Model”.
A Hidden Markov Model λ [16] is defined by the tuple λ S , M , A , B , Π where:
  • S = { S 0 , , S N } is the set of hidden states of the model. The state at time t is denoted by q t .
  • M is the set of observation symbols of the model.
  • A = { a i j } is the state transition probability distribution:
    a i j = p q t + 1 = S j q t = S i
  • B = { b j } is the observation symbol probability distribution in state j:
    b j = p v q t = j
  • Π = { π j } is the initial state probability distribution:
    p i j = p q 0 = j
One of the most powerful features of HMMs is that it can model both large duration and small duration activities. To model large duration activities, an HMM is defined for each activity. Given a sequence O = o _ 1 o _ k of observations to classify as a large duration activity, and a set of HMM activity models Λ = { λ 1 λ n } , the sequence is associated to the HMM with the highest probability of generating O:
arg max i P m i O
P O m i can be estimated using Forward-Backward algorithm [16].
When the recognition target are small duration activities, only one HMM is used, with one state per activity to classify. Given a sequence O = o _ 1 o k of observations to classify as small duration activities, the problem to solve is the association of each observation with the state most likely of generating it:
q t = arg max i P S i o t , q t 1
The solution to estimate the current state of the HMM can be found using Viterbi algorithm [16]. Viterbi algorithm computes the most likely state sequence that has generated the observed output. The algorithm proceeds as fallows:
  • Initialization:
    δ 1 i = π i b i o 1 1 i N
    ψ 1 i = 0
  • Recursion:
    δ t j = max 1 i N δ t 1 i a i j b j o t 2 t K 1 j N
    ψ t j = arg max 1 i N δ t 1 i a i j 2 t N 1 j N
  • Finalization:
    p * = max 1 i N δ K i
    q K * = arg max 1 i N δ K i
  • Path Backtracking:
    q t * = ψ t + 1 q t + 1 * t = K 1 , K 2 , , 1
The algorithm give us the optimal state sequence stored in vector q * . p * is the probability of q * generating O.

3. Feature selection for HMMs using Heuristic and Genetic Algorithms

Discriminant features have to be used in any classification problem to obtain good results. Also, HMMs have been widely used for activity classification, because activities can be easily represented using this abstraction. It seems to be of interest to study how HMMs could be built using feature selection techniques, because it is not clear that some commonly used features [7] provide the necessary separability of the activities.

3.1. Feature subset evaluation

To evaluate the utility of a given feature subset an HMM wrapper is going to be used. The utility of the feature subset will be the accuracy of the HMM for classifying some small duration activity sequences. Given a set of training sequences O = { O 1 O N } , O i = { o i 1 o i k } containing m different activities, and their corresponding set of hidden states A = { A 1 A N } , A i = { a i 1 a i k } , the estimation of HMM parameters is straightforward. The HMM will have m hidden states, one per activity. To estimate he initial state probability distribution, the frequency of each state to be the first in each training sequence has to be computed. The state transition probability distribution is going to be computed using the frequency of each state transition along the training sequences. To simplify even more the estimation of parameters, the observation symbol robability distribution is supposed to be a single multivariate gaussian distribution. Other type of distributions, like Gaussian Mixtures or KDE [17] can be used, but the construction of the HMM will be harder, and it is not important for the objective of this work. The parameters of each state emission distribution to be estimated are the mean and covariance matrices.
Once that the Hidden Markov Model has been estimated, a test sequence O = { o 1 , , o K } can be classified using Viterbi Algorithm. The obtained state sequence q * is compared to ground truth data, and the accuracy of the model for predicting each hidden state is taken as the subset utility.

3.2. Searching the feature space

Given a set of N features, the number of possible feature subsets to evaluate is 2 N . As N grows, the feature selection problem becomes intractable and exhaustive search methods become unpractical. To avoid this course of dimensionality, suboptimal search methods have to be employed. In this paper two different methods are proposed: Best First Search and Genetic Search.
Best First Search (BFS) [18] is a tree search method that explores the search tree expanding the best node at each level. In this case, the best node is the one who has the best accuracy on its level. Algorithm pseudocode is shown on 1. This search method does not guarantee to find the optimal solution, only a local optima, because when selecting only the best child, the path to the optimal solution could be left unexplored.
Algorithm 1 Best First Search
  • O p e n r o o t
  • while O p e n ϕ do
  •     Pick the best node of O p e n
  •      O p e n ϕ
  •     Generate its children and introduce into O p e n
  • end while
To search the feature subset space using Best First Search, the root node of the tree will be the empty feature set. The successors of a node will be that nodes that contain the same features that its parent plus a new one that its parent does not have. The search will finish were the accuracy of children is equal or smaller than the accuracy of their parent.
Genetic Algorithms (GA) [19] are powerful search techniques inspired in the Darwinian Evolution Theory. Populations of individuals, where each one represents a different solution to the problem, are evolved using crossover and mutation operators. The pseudocode of a simple genetic algorithm can be seen on 2. Again, this search method does not guarantee to find the optimal solution, and even if it is found, there is no way to ensure that is the optimum.
Algorithm 2 Genetic Algorithm
  • P o p u l a t i o n i n i t
  • while stop condition is not satisfied do
  •      E v a l u a t e P o p u l a t i o n
  •      N e x t P o p u l a t i o n s e l e c t i o n P o p u l a t i o n
  •      N e x t P o p u l a t i o n r e p r o d u c t i o n P o p u l a t i o n
  •      P o p u l a t i o n r e p l a c e m e n t P o p u l a t i o n , N e x t P o p u l a t i o n
  • end while
  • S o l u t i o n b e s t P o p u l a t i o n
The individual that is going to be used to explore the feature subset space is a bit string of length N, being N the total number of given features. If the ith bit is set to 1, it means that the ith feature is selected, whereas if it is set to 0, it means that the ith feature is not selected.

4. Experimental validation

4.1. Experimental setup

A video has been recorded at 25 fps on a corridor using a Sony SSC-DC598P video surveillance camera located about 3 meters from the ground. An actor has performed six different small duration activities: walking, stopped, bending, falling, lied down and rising (see Figure 4). A blob tracker has been used to extract the actor bounding box on each frame where the actor appears. Using the bounding box, all the features presented on section 2.2. have been extracted, using temporal windows from 3 to 25 frames, giving 780 different features per frame. The blob tracker has divided the original video on 11 video sequences where the activity being performed on each frame has been manually annotated. The content of each sequence could be seen on Table 3.
The blob tracker has produced some erroneous outputs, where the bounding box does not contain the actor for some frames (see Figure 5. These errors may be produced by scene illumination changes, or by the initialization of a false track that is associated to the actor later. In this cases, the frame has been annotated with the especial activity error. It has been considered that is important to have that label to be able to detect these type of false alarms when the system is working.
An HMM is going to be build for each set of small duration activities to be recognized using both BFS and GA. To evaluate the utility of a candidate feature subset, leave-one-out cross-validation (LOOCV) of the generated HMM is going to be performed with the sequences to be employed on each experiment. LOOCV consist on generate the model with all the sequences unless one and test the model with that sequence. This process is repeated for every sequence to be employed. Using this training strategy, possible overfittings are minimized, because the model has not been built with the sequence used to meassure its quality.
Figure 4. Examples of the activities to recognize. First row, walking. Second row, stopped. Third row, lowing. Fourth row, falling. Fifth row, lied down. Last row, rising
Figure 4. Examples of the activities to recognize. First row, walking. Second row, stopped. Third row, lowing. Fourth row, falling. Fifth row, lied down. Last row, rising
Algorithms 02 00282 g004
Table 3. Content of the sequences used. W stands for walking, S for stopped, F for falling, D for lyed down, R for rising, L for bending and E for error. The number of frames that the activity last is shown between parenthesis
Table 3. Content of the sequences used. W stands for walking, S for stopped, F for falling, D for lyed down, R for rising, L for bending and E for error. The number of frames that the activity last is shown between parenthesis
SequenceContent
1W(220), S(214), W(113), E(197), S(161), W(221)
2W(107), S(172), W(116), S(268), W(12)
3W(50), S(134), W(112), S(202), W(224), S(181), W(145), S(239), E(5), S(30), W(269), S(190), W(250), E(55)
4W(56), S(131), W(215), F(27), D(211), R(64), W(174), F(57), D(171), R(97), W(115), S(69), W(156), F(68), D(58), R(61)
5W(139), F(57), D(79), R(88), W(130), F(54), D(172), R(80), W(176), F(52), D(52), R(82), W(236), E(22)
6L(281), W(283), L(208), W(213), L(173), W(156), L(168), W(167), E(23)
7W(62), L(134), W(146), L(43), W(222), L(157), W(158), L(110), W(80), L(180), W(133), L(168), W(28), E(18)
8W(733), E(10)
9W(760), E(48)
10E(155), W(271), E(307)
11W(65), E(1)
Figure 5. Example of the error activity. Bounding box (in green) does not contain the actor.
Figure 5. Example of the error activity. Bounding box (in green) does not contain the actor.
Algorithms 02 00282 g005
GA has been employed using a configuration of 780 individuals and a mutation rate of 1 780 . The selection technique employed has been tournament selection of 3 individuals. The population of the genetic algorithm is initialized with one random gen set to true. As the length of the individual is 780 and the population 780, it is expected to have each bit set to true in the population at least one time. Population evolution has stopped when the fitness value of the best individual has been without changing for 10 generations. The algorithm has been run 20 times for each experiment, because GAs are random search techniques and different results may be achieved in different trials. The result shown is the best one found.

4.2. Experiment I: three activities

A first attempt to solve the problem will be done using a set of sequences containing sequences 1, 2 and 3. It contains three different activities: stopped, walking and noise. Results obtained with GA and BFS can be seen on Table 4. Relative confusion matrix for the solution obtained by BFS is shown on Table 5 and relative confusion matrix of the solution obtained by GA is shown on Table 6. Relative confusion matrices show the frequency of predicting errors per class. The HMM induced can be seen on Figure 6
Table 4. Features found by by Best First Search (BFS) and Genetic Algorithm(GA) for classificating three activities
Table 4. Features found by by Best First Search (BFS) and Genetic Algorithm(GA) for classificating three activities
AlgorithmAccuracySelection
BFS0.8699 f t , F , R Σ F 4 ¯ t , F 1 , R Σ F 3 ¯ t , F 1 , R Σ F 4 ¯ t , I 2 , R Σ F 5 ¯ t , I 1
GA0.9046 y t t , R Σ F 3 ¯ t , F 1 , f 8 t , F 1 ,
The solution found by GA is better than the solution found by BFS according to the global accuracy. Also, taking a look in the confusion matrices of both classifiers (Table 5 and Table 6) reveals that the solution found by GA is better than the solution found by BFS for the accuracy of all the classes to predict. In both cases, the hardest class to predict is the error class, maybe because it is the class with less examples.
Solution obtained by BFS algorithm includes different meassures of the optical flow. Solution obtained by GA includes velocity along y axes and different optical flow meassures.
Table 5. Relative Confusion matrix for the HMM generated using BFS to classify the activities walking, stopped and error
Table 5. Relative Confusion matrix for the HMM generated using BFS to classify the activities walking, stopped and error
predictionerrorwalkingstopped
class
error0.58690.41610
walking0.05050.89840.0511
stopped0.05310.07960.8673
Figure 6. Structure of the HMM used in experiment 1
Figure 6. Structure of the HMM used in experiment 1
Algorithms 02 00282 g006
Table 6. Relative Confusion matrix for the HMM generated using GA to classify the activities walking, stopped and error
Table 6. Relative Confusion matrix for the HMM generated using GA to classify the activities walking, stopped and error
predictionerrorwalkingstopped
class
error0.75210.04270.2051
walking0.02380.92600.0502
stopped0.08280.02510.8921

4.3. Experiment II: six activities

Two more sequences have been added to the set previously used: X and Y. Three new activities are going to be classified: falling,rising, and lyed down. Results obtained with GA and BFS can be seen on Table 7.
Table 7. Features found by by Best First Search (BFS) and Genetic Algorithm(GA) for classificating six activities
Table 7. Features found by by Best First Search (BFS) and Genetic Algorithm(GA) for classificating six activities
AlgorithmAccuracySelection
BFS0.7051 f 8 ¯ t , F ; R s 6 , 1 t , R s 6 , 2 t , t r Σ F 2 5 ¯ t , F 3 , R Σ s , 6 t 2 , R Σ F 3 ¯ t 2
GA0.7779 h u 1 , R Σ s , 4 t 2 , R s 9 , 1 t , R s 1 1 , 1 t , R s 1 4 , 1 t , f 1 7 ¯ t , F , t r Σ F 2 2 ¯ t , F 3
Table 8. Relative Confusion matrix for the HMM generated using BFS to classify the activities walking, stopped, falling, lyed down, rising and error
Table 8. Relative Confusion matrix for the HMM generated using BFS to classify the activities walking, stopped, falling, lyed down, rising and error
predictionerrorwalkingstoppedfallinglyed downrising
class
error0.27600.50820.08200.09840.00270.0328
walking0.00980.81340.02250.06400.00600.0843
stopped0.00500.12670.80920.00560.01950.0340
falling0.09570.03190.01060.65960.20210
lyed down0.06870.01020.31630.00370.57930.0219
rising0.15880.58290.011800.00240.2441
Figure 7. Structure of the HMM used in experiment 2
Figure 7. Structure of the HMM used in experiment 2
Algorithms 02 00282 g007
Again, the solution found by GA is better than the solution found by BFS if the global accuracy of the HMM induced (see Figure 7 ) by the feature subset is considered. But now, taking a look into the
Table 9. Relative Confusion matrix for the HMM generated using GA to classify the activities walking, stopped, falling, lyed down, rising and error
Table 9. Relative Confusion matrix for the HMM generated using GA to classify the activities walking, stopped, falling, lyed down, rising and error
predictionerrorwalkingstoppedfallinglyed downrising
class
error0.25090.58180.09820.040.00360.0255
walking0.01630.86560.02910.04040.00410.0445
stopped0.06710.04390.78030.00440.06230.0421
falling0.09390.07120.04210.52430.11650.1521
lyed down0.00120.07310.08990.00240.80460.0288
rising0.02270.44340.00970.00320.01620.5049
confusion matrices of both classifiers (Table 8 and Table 9) reveals that the solution found by BFS is better for predicting classes error, stopped and falling, while the solution found by GA is better for walking, lyed down and rising.
Solution found by BFS includes optical flow and trajectory eccentricity meassures. GA includes also a shape descriptor.
Table 10. Results obtained by Best First Search (BFS) and Genetic Algorithm(GA) using all the training sequences
Table 10. Results obtained by Best First Search (BFS) and Genetic Algorithm(GA) using all the training sequences
AlgorithmAccuracySelection
BFS0.7442 h u 1 , f t , F , R s 4 , 1 t , R Σ s , 3 t 2 , R s 6 , 1 t , R s 2 2 , 2 t , R s 3 , 1 t , R s 1 8 , 1 t
GA0.7501 h u 1 , R s 3 , 1 t , R s 8 , 2 t , R Σ s , 11 t 1 , f 8 ¯ t , F , R s 1 2 , 2 t , t r Σ F 1 3 ¯ t , F 1
Table 11. Relative Confusion matrix for de HMM generated using Best First Search
Table 11. Relative Confusion matrix for de HMM generated using Best First Search
predictionerrorwalkingbendingstoppedfallinglied downrising
class
error0.55140.14790.02100.11890.00400.09990.0569
walking0.01110.86760.06030.01360.01180.00350.0320
bending0.03180.30430.63040.00930.00390.02020
stopped0.06060.08180.05960.75130.01390.01760.0153
falling0.12930.22390.11390.06560.33010.06760.0695
lied down0.00580.03610.23000.09260.00390.62180.0097
rising0.01080.34590.18110.04050.06220.02160.3378

4.4. Experiment III: seven activities

Finally, the eleven available sequences have been used to train the HMM models, adding a new activity to the previous experiment configurations: bending (see HMM on Figure 8). Results obtained with GA and BFS can be seen on Table 10.
Table 12. Relative Confusion matrix for de HMM generated using a Genetic Algorithm
Table 12. Relative Confusion matrix for de HMM generated using a Genetic Algorithm
predictionerrorwalkingbendingstoppedfallinglied downrising
class
error0.48810.15120.01480.14130.02670.11760.0603
walking0.01010.85620.07010.01740.01660.00320.0264
bending0.01960.28440.68150.01160.002900
stopped0.05710.04210.08700.77580.00860.01220.0172
falling0.08740.60530.03900.00310.20590.02650.0328
lied down0.06920.01250.09990.00110.01700.77530.0250
rising0.14860.26350.02030.00340.03040.00340.5304
Figure 8. Structure of the HMM used in experiment 3
Figure 8. Structure of the HMM used in experiment 3
Algorithms 02 00282 g008
Taking as the quality meassure the global accuracy of the HMM induced by the result subsets, the solution found by GA is a bit better than the solution found by BFS. According to cobfusion matrices of both classifiers (Table 11 and Table 12), solution found by BFS is better for predicting falling and error while the easiest are walking and stopped, whereas the solution found by GA is better for bending, stopped, lied down and rising.
Solutions found by both BFS and GA include shape descriptors and optical flow and trajectory eccentricity meassures.

4.5. Overall Discussion

Observing confusion matrices, it can be said that the activities performed on more frames are the best classified. That is due to the feature subset evaluation function used, that tries to maximize the global accuracy of the classifier. This effect can be avoided modifying the feature subset evaluation function to reward the correct classification of activities with fewer frames.
Solutions found by GA are smaller than the solutions found by BFS, so classification using the HMM generated by GA is easier than with the obtained by BFS. Besides classification cost, considering that the time needed to extract all the features is the same, extracting five features need less time than extracting nine, so the overall system using the solution by GA will be quicker. Also, the temporal window needed to extract the features found by GA tend to be smaller than for the features found by BFS (22 vs 11 frames), so the system using that feature set will have to store less information in its memory.

5. Conclusion

In this paper a set of candidate features that may represent human activities has been presented. Searching over the possible feature subsets, Hidden Markov Models for classifying activities have been built using only features that are relevant for target classes, over the traditional application where the features employed have to be defined a prioir. The Genetic Algorithm used to explore the feature subsets space has shown to be better than the Best First Search for almost every criteria. It has been observed that the accuracy predicting activities has decreased when the number of activities to predict has grown, because the probability of assigning an observation to the wrong class increases.
When different sensors are used, the selected feature subset used for activity recognition may come from different ones, discarding that sensors that don’t provide any information for the activities to recognized.

Acknowledgments

This work was supported in part by Projects CICYT TIN2008-06742-C02-02/TSI, CICYT TEC2008-06732-C02-02/TEC, SINPROB and CAM MADRINET S-0505/TIC/0255

References and Notes

  1. Corchado, J.; Bajo, J.; de Paz, Y.; Tapia, D. Intelligent environment for monitoring Alzheimer patients, agent technology for health care. Decision Support Sys. 2008, 44, 382–396. [Google Scholar] [CrossRef]
  2. Pavón, J.; Gómez-Sanz, J.; Fernández-Caballero, A.; Valencia-Jiménez, J. Development of intelligent multisensor surveillance systems with agents. Robot. Autonom. Sys. 2007, 55, 892–903. [Google Scholar] [CrossRef]
  3. Robertson, N.; Reid, I. A general method for human activity recognition in video. Comput. Vis. Image Understand. 2006, 104, 232–248. [Google Scholar] [CrossRef]
  4. Perez, O.; Piccardi, M.; Garcia, J.; Molina, J. Comparison of Classifiers for Human Activity Recognition. Lecture Notes Comput. Sci. 2007, 4528, 192. [Google Scholar]
  5. Yang, M.; Ahuja, N. Recognizing hand gesture using motion trajectories. Proc. IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1999; Vol. 1. [Google Scholar]
  6. Ribeiro, P.; Santos-Victor, J. Human activity recognition from Video: modeling, feature selection and classification architecture. Proc. International Workshop on Human Activity Recognition and Modelling, 2005; Vol. 1, 61–78. [Google Scholar]
  7. Perez, O.; Piccardi, M.; García, J.; Patricio, M.; Molina, J. Comparison Between Genetic Algorithms and the Baum-Welch Algorithm in Learning HMMs for Human Activity Classification. Lecture Notes Comput. Sci. 2007, 4448, 399. [Google Scholar]
  8. Brand, M.; Kettnaker, V. Discovery and segmentation of activities in video. IEEE Trans. Patt. Anal. Machine Intell. 2000, 22, 844–851. [Google Scholar] [CrossRef]
  9. Oliver, N.; Rosario, B.; Pentland, A. A Bayesian computer vision system for modeling human interactions. IEEE Trans. Patt. Anal. Machine Intell. 2000, 22, 831–843. [Google Scholar] [CrossRef]
  10. Patricio, M.; Castanedo, F.; Berlanga, A.; Perez, O.; García, J.; Molina, J. Computational Intelligence in Visual Sensor Networks: Improving Video Processing Systems. Comput. Intell. Multimedia Process.: Recent Adv. 2008. [Google Scholar]
  11. Yilmaz, A.; Javed, O.; Shah, M. Object tracking: A survey. ACM Comput. Surveys 2006, 38, 13. [Google Scholar] [CrossRef]
  12. Hu, M. Visual pattern recognition by moment invariants. IEEE Trans. Inform. Theory 1962, 8, 179–187. [Google Scholar]
  13. Lucas, B.; Kanade, T. An iterative image registration technique with an application to stereo vision. Proc. International Joint Conference on Artificial Intelligence, 1981; Vol. 81, 674–679. [Google Scholar]
  14. Liu, H.; Yu, L. Toward integrating feature selection algorithms for classification and clustering. Trans. Knowl Data Engineer. 2005, 17, 491–502. [Google Scholar]
  15. Witten, I.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann Publishers: San Francisco, CA, USA, 2005. [Google Scholar]
  16. Rabiner, L. A tutorial on hidden Markov models and selected applications inspeech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef]
  17. Piccardi, M.; Perez, O. Hidden Markov Models with Kernel Density Estimation of Emission Probabilities and their Use in Activity Recognition. Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2007; 1–8. [Google Scholar]
  18. Russell, S.; Norvig, P.; Canny, J.; Malik, J.; Edwards, D. Artificial intelligence: a modern approach; Prentice Hall: Englewood Cliffs, NJ, USA, 1995. [Google Scholar]
  19. Holland, J. Adaptation in natural and artificial systems; MIT Press: Cambridge, MA, USA, 1992. [Google Scholar]

Share and Cite

MDPI and ACS Style

Cilla, R.; Patricio, M.A.; García, J.; Berlanga, A.; Molina, J.M. Recognizing Human Activities from Sensors Using Hidden Markov Models Constructed by Feature Selection Techniques. Algorithms 2009, 2, 282-300. https://doi.org/10.3390/a2010282

AMA Style

Cilla R, Patricio MA, García J, Berlanga A, Molina JM. Recognizing Human Activities from Sensors Using Hidden Markov Models Constructed by Feature Selection Techniques. Algorithms. 2009; 2(1):282-300. https://doi.org/10.3390/a2010282

Chicago/Turabian Style

Cilla, Rodrigo, Miguel A. Patricio, Jesús García, Antonio Berlanga, and Jose M. Molina. 2009. "Recognizing Human Activities from Sensors Using Hidden Markov Models Constructed by Feature Selection Techniques" Algorithms 2, no. 1: 282-300. https://doi.org/10.3390/a2010282

APA Style

Cilla, R., Patricio, M. A., García, J., Berlanga, A., & Molina, J. M. (2009). Recognizing Human Activities from Sensors Using Hidden Markov Models Constructed by Feature Selection Techniques. Algorithms, 2(1), 282-300. https://doi.org/10.3390/a2010282

Article Metrics

Back to TopTop