Download - An efficient algorithm for recognition of human actions

Transcript

Research ArticleAn Efficient Algorithm for Recognition of Human Actions

Yaser Daanial Khan,1 Nabeel Sabir Khan,1 Shoaib Farooq,1 Adnan Abid,1 Sher Afzal Khan,2

Farooq Ahmad,3 and M. Khalid Mahmood4

1 School of Science and Technology, University of Management and Technology, Lahore 54000, Pakistan2Department of Computer Science, Abdul Wali Khan University, Mardan 23200, Pakistan3 Faculty of Information Technology, University of Central Punjab, 1-Khayaban-e-Jinnah Road, Johar Town, Lahore 54000, Pakistan4Department of Mathematics, University of the Punjab, Lahore 54000, Pakistan

Correspondence should be addressed to Yaser Daanial Khan; [email protected]

Received 4 April 2014; Revised 26 June 2014; Accepted 27 June 2014; Published 27 August 2014

Academic Editor: Yu-Bo Yuan

Copyright ยฉ 2014 Yaser Daanial Khan et al. This is an open access article distributed under the Creative Commons AttributionLicense, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properlycited.

Recognition of human actions is an emerging need.Various researchers have endeavored to provide a solution to this problem. Someof the current state-of-the-art solutions are either inaccurate or computationally intensive while others require human intervention.In this paper a sufficiently accurate while computationally inexpensive solution is provided for the same problem. Image momentswhich are translation, rotation, and scale invariant are computed for a frame. A dynamic neural network is used to identify thepatterns within the stream of image moments and hence recognize actions. Experiments show that the proposed model performsbetter than other competitive models.

1. Introduction

Human action recognition is an important field in computervision. The implications of robust human action recognitionsystem, requiring minimal computations, include a widearray of potential applications such as sign language recog-nition, keyboard or a remote control emulation, humancomputer interaction, surveillance, and video analysis. Suchsystems are developed to enable a computer to intelligentlyrecognize a stream of complex human actions being inputvia a digital camera. It thrives for the need of a multitude ofefficiently designed algorithms pertaining to pattern recogni-tion and computer vision. Background noise, cameramotion,and position and shape of the object are major impairmentfactors against the resolution to this problem. This paperpresents an efficient and sufficiently accurate algorithm forhuman action recognition making use of image moments. Acomprehensive understanding of image moments describescharacteristics information of an image.The proposed systemaims to recognize human actions regardless of its position,scale, colors, size, and phase of the human. The paperdescribes a robust feature extraction and comprehensive

classification and training processes. The primary focus isto facilitate video retrieval classified on the basis of featuredhuman action. Inherently it requires methods to identifyand discover objects of interest by providing comprehensivefeatures after video segmentation, feature extraction, andfeature vector organization. These features are designedsuch that they are immune to encumbrances such as noiseand background view. This calls for methods incessantlycapable of tackling video descriptors which are repeatableand most relevant. An efficient computational paradigm forextraction of such descriptors needs to be devised becauseonly those areas of an image are matters of concern, whichcontain deciphering features. A real-time implementation isrealized for detection of nominated human actions. Variousresearchers have addressed the proposed problem usingdifferent methodologies. Tran et al. represent human actionas a combination of the movements of the body part [1].They provide a representation described by a combinationof movements of the body part to which a certain actioncorrelate. Their proposed method makes use of polar patternof the space for representing the movement of the individualparts of the body. In another article Ali and Shah [2] represent

Hindawi Publishing Corporatione Scientific World JournalVolume 2014, Article ID 875879, 11 pageshttp://dx.doi.org/10.1155/2014/875879

2 The Scientific World Journal

kinematic functions computed from optical flow for therecognition of human action in video tribes.These kinematicfeatures represent the spatiotemporal properties of the video.It further performs principal component analysis (PCA) onthe kinematic feature volume. Multiple instance learning(MIL) is used for the purpose of classification of humanaction using succinct data after PCA. Busaryev and Doolittlerecognize hand gestures captured from a webcam in realtime. Such classification of gestures is applied to controlreal-world applications. Background subtraction and HSV-based extraction are compared as methods for getting aclean hand image for further analysis. The gesture in eachhand image is then determined with Hu moments or a localfeature classifier, and each gesture is mapped to a certainkeystroke or mouse function [3]. Cao et al. combine multiplefeatures for action detection. They build a novel frameworkwhich combines GMM-based representation of STIPs baseddetection [4]. In order to detect moving objects from compli-cated backgrounds, Zhang et al. improved Gaussian mixturemodel, which uses K-means clustering to initialize the modeland gets better motion detection results for surveillancevideos [5]. They demonstrate that the proposed silhouetterepresentation, namely, โ€œenvelope shape,โ€ solves the view-point problem in surveillance videos. Shao et al. present amethod that extracts histogram of oriented gradients (HOG)descriptors corresponding to primitive actions prototype [6].The output contains only the region of interest (ROI). Usingthis information the gradient of motion is computed formotion estimation. The gradient vectors are obtained for thepartitioning of periodic effect. Once it detects a completecycle of movement, two key frames are selected for encodingthemotion. Finally, the current class action descriptors for theclassification of features are extracted while the correspond-ing classifier is trained offline. Ullah et al. implemented thebag of features (BoF) approach for classification of humanactions in realistic videos.Themain idea is to segment videosinto semantically meaningful regions (both spatially andtemporally) and then to compute histogram of local featuresfor each region separately [7, 8].

Certain weaknesses of the recognition algorithm forhuman actions in video with the kinematic features [8]and multiple instance learning are quite evident. Firstly thekinematic properties selected are not scale, translation, androtation invariant, as the same action from different anglesinduces different optical flow. Secondly, occlusion presentsserious consequences for the performance of the algorithm,especially in cases where a significant part of the body isclosed. Moreover the training step is the slowest part ofthe algorithm which makes excessive use of memory due toits iterative nature. The method using the HSV model forsegmentation of hands will have problems if another objectof the same hue is present in the frame. Other methods usingsparse representations of human action recognition cannothandle several actions in a video clip. This is because they donot take into account the spatial and temporal orientation ofthe extracted features. The method discussed in [9, 10] usescolor intensities to segment the action by manually selectinga region. Using this approach a region must be selectedevery time when the scene changes; this undesirably requires

human intervention. Furthermore, most of the algorithmswork only for a specific illumination; it will fail to giveresults on high or low illumination. The approach used in[11] is based upon the assumption that each independentobservation follows the same distribution. Certainly thisapproach is bound to fail in case the distribution of theobservations is quite the reverse. Although the approachseems to be scale invariant still it is not rotation invariant.

The paper is organized into several sections. Section 1gives a brief introduction of the problem and the currentstate of the art. Section 2 gives an overview of the proposedsystem. Section 3 describes the feature extraction process.Section 4 gives a comprehensive description of the trainingprocess. Section 5 provides some detailed results from themodel while Section 6 adds some conclusive remarks.

2. An Overview of the Proposed System

The system is designed to retrieve semantically similar videoclips for a given criterion video clip, from a large set ofsemantically different videos. The video dataset containsfeatures of every proposed action and on query, video featureswill be extracted and matched with the stored features inthe feature library. Since gestures are sequence of postures(static frames), therefore the system is expected to recognizegestures by identifying constituent postures one by one.Ultimately a temporal classifier is used to classify the inputstream of spatial postures into an action. Figure 1 showsthe flow of the initial training process. Firstly, individualframes are extracted from the video input. Secondly eachextracted frame is preprocessed to make it suitable formoment extraction. These moments form a feature vectorwhich is initially used for training of the system.

The system is exhaustively trained using the training pro-cess described later. A sufficiently trained system is deemedappropriate for classification of the proposed actions. Figure 2shows the process used for classification of human actions.

Extracted features from a live video feed are fed intoa trained dynamic neural network (DNN) for the purposeof classification. The neural network classifies the actionperformed in a few successive frames. The dynamic neuralnetwork is designed such that its behavior varies temporallybased on the video input.

3. Preprocessing and Feature Extraction

Initially a number of preprocessing steps must be performedon video frames beforemoments based features are extracted.Computations of moments require that the image is ofmonochrome nature. The chromatic frame extracted fromthe video is firstly binarized using a threshold. The thresholdis carefully chosen based on themean illumination level of theframe. Mean illumination is computed by taking the mean ofluminosity value of each pixel in the frame. Once binarized,the image will hold either black or white pixels. Further toremove noise and other impairments dilation and erosion isperformed [12]. Figures 3 and 4 show the result of this processon a sample frame.

The Scientific World Journal 3

Training video Frame extraction Frame Digital image Preprocessing

Training Label the frame Moments extraction

Resu

ltant

fram

e

Figure 1: The steps of the training process.

Testing video Frame extraction Frame Digital image Preprocessing

Neural networkClassifed action Feature vectorClassifcation Moments extractionRe

sulta

nt fr

ame

Figure 2: The classification process.

Before any intricate processing is performed on the dataset, the background is removed from each frame. Heretwo alternate approaches are adopted. In the first approachinitial few frames are captured without any foreground actioncontaining only the background. Any frame from this initialfootage is used as a representative. This frame is subtractedfrom each frame containing foreground to obtain the filteredforeground. In the other approach each successive frame isXORed. The resultant frame represents the change in actionduring the period of the latter frame. The difference frame inthis case also excludes the background.

3.1. Moments Extraction. Moments are scalar quantitieswhich are used to categorize the shape and its features.They are computed from the shape boundary and its entireregion. The concept of moments in images is quite similarto the concept of moments in physics. One major differencebetween the two is that image moments are inherently two-dimensional in nature. The resolution to the proposed prob-lem is sought with the help of various moments such as raw,central, and scale invariant and rotation invariant moments

along with certain corporeal properties of the image likethe centroid and eccentricity. Invariant moments are thosemoments which are impervious to certain deformations inthe shape and are most suited for comparison between twoimages. The scale, location, and rotation invariant momentsare used to extract features regardless of size, position, androtation, respectively.

3.2. Raw Moments. Raw moments are calculated along theorigin of the image. Let ๐‘“(๐‘ฅ, ๐‘ฆ) be a function that definesan image where (๐‘ฅ, ๐‘ฆ) are any arbitrary coordinates of theimage. In case of two-dimensional continuous signal the rawmoment function๐‘€๐‘๐‘ž for themoment of order (๐‘+๐‘ž) is givenas

๐‘€๐‘๐‘ž = โˆ‘๐‘ฅ

โˆ‘๐‘ฆ

๐‘ฅ๐‘๐‘ฆ๐‘ž๐‘“ (๐‘ฅ, ๐‘ฆ) , (1)

where ๐‘“(๐‘ฅ, ๐‘ฆ) is ๐‘ฅth pixel along ๐‘ฅ-axis and ๐‘ฆth pixel along๐‘ฆ-axis and ๐‘, ๐‘ž are the ๐‘th and ๐‘žth indices of the moments.These moments are computed throughout the span of the

4 The Scientific World Journal

(a) (b)

Figure 3: (a) The original frame. (b) The frame after.

(a) (b)

Figure 4: (a) The binarized image. (b) The same image after erosion dilation operations.

image. The raw moments provide information about proper-ties like area and size of the image; for example, the moment๐‘€00 will give the area of object.

3.3. Central Moments. The moments which are invariant totranslation of objects in an image are called central momentsas they are computed along the centroid rather than theorigin. From the equation of raw moments central momentsare calculated such that the first two order moments from(18), that is, ๐‘€10 and ๐‘€01, are used to locate the centroid ofthe image.

Let ๐‘“(๐‘ฅ, ๐‘ฆ) be a digital image; then reducing the coordi-nates in previous equation by center of gravity (๐‘ฅ and ๐‘ฆ) ofthe object we get

๐œ‡๐‘๐‘ž = โˆ‘๐‘ฅ

โˆ‘๐‘ฆ

(๐‘ฅ โˆ’ ๐‘ฅ)๐‘(๐‘ฆ โˆ’ ๐‘ฆ)

๐‘ž๐‘“ (๐‘ฅ, ๐‘ฆ) . (2)

The coordinates of the center of mass (๐‘ฅ, ๐‘ฆ) are the point ofintersection of the lines ๐‘ฅ = ๐‘ฅ and ๐‘ฆ = ๐‘ฆ, parallel to the๐‘ฅ and ๐‘ฆ-axis, where the first order moments are zero. Thecoordinates of the center of gravity are the components of thecentroid given as follows:

๐‘ฅ =๐‘€10๐‘€00

, ๐‘ฆ =๐‘€01๐‘€00

, (3)

while๐œ‡00 = ๐‘€

00,

๐œ‡01 = ๐œ‡10 = 0.(4)

Moments of order up to three are simplified in [13] and aregiven as follows:

๐œ‡11 = ๐‘€11 โˆ’ ๐‘ฅ๐‘€01 = ๐‘€11 โˆ’ ๐‘ฆ๐‘€10,

๐œ‡20 = ๐‘€20 โˆ’ ๐‘ฅ๐‘€10,

๐œ‡20 = ๐‘€20 โˆ’ ๐‘ฅ๐‘€10,

๐œ‡21 = ๐‘€21 โˆ’ 2๐‘ฅ๐‘€11 โˆ’ ๐‘ฅ๐‘€20 + 2๐‘ฅ2๐‘€01,

๐œ‡12 = ๐‘€12 โˆ’ 2๐‘ฆ๐‘€11 โˆ’ ๐‘ฅ๐‘€02 + 2๐‘ฆ2๐‘€10,

๐œ‡30= ๐‘€30 โˆ’ 3๐‘ฅ๐‘€20 โˆ’ 2๐‘ฅ2๐‘€10,

๐œ‡03= ๐‘€30 โˆ’ 3๐‘ฆ๐‘€02 โˆ’ 2๐‘ฆ2๐‘€10.

(5)

It is shown in [14] that the generalized form of centralmoments is

๐œ‡๐‘๐‘ž =

๐‘

โˆ‘๐‘š

๐‘ž

โˆ‘๐‘›

(๐‘๐‘š)(

๐‘ž๐‘›) (โˆ’๐‘ฅ)

(๐‘โˆ’๐‘š)(โˆ’๐‘ฆ)(๐‘žโˆ’๐‘›)

๐‘€๐‘š๐‘›. (6)

The main advantage of central moments is their invariancesto translations of the object. Therefore they are suited well to

The Scientific World Journal 5

describe the form or shape of the object while the centroidpertains to information about the location of the object.

3.4. Scale Invariant Moments. The raw moments and thecentral moments depend on the size of object. This createsa problem when the same object is compared but both theimages are captured from different distances. To deal withthis encumbrance scale invariant moments are calculated.Moments ๐œ‡๐‘–๐‘— are invariant to changes in scale and areobtained by dividing the central moment by scaled (00)thmoment as given in the following:

๐œ‡๐‘–๐‘— =๐œ‡๐‘–๐‘—

๐œ‡(1+(๐‘–+๐‘—)/2)

00

. (7)

3.5. Rotational Invariant Moments. Rotational moments arethose moments which are invariant to changes in scale andalso in rotation. Most frequently used are the Hu set ofinvariant moments:

๐ผ1 = ๐œ‚20 + ๐œ‚02, ๐ผ2 = (๐œ‚20 โˆ’ ๐œ‚02)2+ 4๐œ‚211,

๐ผ3 = (๐œ‚30 โˆ’ 3๐œ‚12)2+ (3๐œ‚21 โˆ’ ๐œ‚03)

2,

๐ผ4 = (๐œ‚30 + ๐œ‚12)2+ (๐œ‚21 โˆ’ ๐œ‚03)

2,

๐ผ5 = (๐œ‚30 + 3๐œ‚12) (๐œ‚30 โˆ’ ๐œ‚12)

ร— [(๐œ‚30 + ๐œ‚12)2โˆ’ 3(๐œ‚21 + ๐œ‚03)

2]

+ (3๐œ‚21โˆ’ ๐œ‚03) (๐œ‚21 + ๐œ‚03)

ร— [3(๐œ‚30 + ๐œ‚12)2โˆ’ (๐œ‚21 + ๐œ‚03)

2] ,

๐ผ6 = (๐œ‚20 โˆ’ ๐œ‚02) [(๐œ‚30 + ๐œ‚12)2โˆ’ (๐œ‚21 + ๐œ‚03)

2]

+ 4๐œ‚11 (๐œ‚30 + ๐œ‚12) (๐œ‚21 + ๐œ‚03) ,

๐ผ7 = (๐œ‚30 + 3๐œ‚12) (๐œ‚30 โˆ’ ๐œ‚12)

ร— [(๐œ‚30 + ๐œ‚12)2โˆ’ 3(๐œ‚21 + ๐œ‚03)

2]

โˆ’ (๐œ‚30 โˆ’ 3๐œ‚12) (๐œ‚21 + ๐œ‚03)

ร— [3(๐œ‚30 + ๐œ‚12)2โˆ’ (๐œ‚21 + ๐œ‚03)

2] .

(8)

All the moments discussed in this section are computed foreach frame.The collection of themoments is used as a featurevector.This feature vector provides characteristic informationabout the contents of the frame in numerical form. Thevariation of patterns formed by periodic frames in a videodefines the action being performed. Further a framework ispresented capable of recognizing the hidden patterns withinthe stream of feature vectors for each defined human action[14โ€“16].

4. Training the Network

A drawback of supervised training is that training data needsto be labeled. Initially each frame in the training video is

Inpu

ts

Input layer Hidden layer

...

...

...

...

Output layer

Out

puts

Figure 5: A recurrent neural network, notice the output beingrecurrently fed into the input layer.

assigned a class number. A specific number is assigned to eachclass, inherently; the frame related to any class will be givena class number. A target matrix is organized such that eachcolumn represents a label of a frame within the training data.Another input matrix is correspondingly organized in whicheach column contains the extracted moments of the frame.Further a neural network is designed such that neurons in theinput layer could be clamped to each element in the obtainedfeature vector. The neurons in hidden layer are variable andwill be changed to fine-tune the results, while the output layerhas neurons equivalent to the number of designated classes.Moreover the network is of recurrent nature; that is, theoutput at output layer is recurrently clamped with the inputas shown in Figure 5. Initially all the inputs and outputs ofhidden layer are assigned random weights. Back propagationalgorithm is used to adjust these weights and converge theoutput.This algorithmmakes use of the sigmoid function (๐œŽ)for the training purpose given as

๐œŽ (๐‘ฅ) =1

1 + ๐‘’โˆ’๐‘ฅ. (9)

The derivative of this function is given as

๐‘‘๐œŽ (๐‘ฅ)

๐‘‘๐‘ฅ= ๐œŽ (๐‘ฅ) โ‹… (1 โˆ’ ๐œŽ (๐‘ฅ)) . (10)

The feature vector for each frame is fed into the input layerand the output is computed. Initially randomly assignedweights are used for each edge. The difference betweenthe actual and labeled output determines the error. Backpropagation technique back-tracks this error and readjuststhe weights so that the error is reduced. The weights areadjusted in a backward direction. In case of proposednetworkweights are adjusted in the hidden layer and then the sameis done for input layer. Several iterations are performed foreach input until convergence is achieved and no appreciablechange in weights is perceived.

Let the weight of an edge between an arbitrary neuron ๐‘–in input layer and an arbitrary neuron in hidden layer ๐‘— begiven as ๐‘ค๐‘–๐‘— while the weight of an edge between an arbitraryneuron ๐‘— in hidden and arbitrary neuron ๐‘˜ in output layer

6 The Scientific World Journal

is given as ๐‘ค๐‘—๐‘˜. For each neuron in input layer the followingoperations are performed:

๐œ“๐‘— =๐‘

โˆ‘๐‘–=1

๐‘ฅ๐‘–๐‘ค๐‘–๐‘— โˆ’ ๐œ๐‘—, ๐œ’๐‘— =1

1 + ๐‘’โˆ’๐œ“๐‘—, (11)

where ๐‘ represents the number of input layer neurons and๐œ๐‘— the threshold used by the ๐‘—th neuron in the hidden layer.Outputs at the hidden layer are given as follows:

๐œ“๐‘˜ =๐‘€

โˆ‘๐‘—=1

๐œ’๐‘—๐‘ค๐‘—๐‘˜ โˆ’ ๐œ๐‘˜, ๐œ’๐‘˜ =1

1 + ๐‘’โˆ’๐œ“๐‘˜, (12)

while ๐œ๐‘˜ is the threshold of the ๐‘˜th neuron at the output layer,๐œ’๐‘˜ is the neuron output, and ๐‘€ is the number of neurons inhidden layer.

The obtained feature vector for a single video frame isclamped to the neural network in order to produce an output๐‘ง๐‘˜. Here the difference between the desired and actual outputis computed as the error ๐œ–๐‘˜ given as

๐œ–๐‘˜ = ๐œ†๐‘˜ โˆ’ ๐‘ง๐‘˜, (13)

while ๐œ†๐‘˜ is the desired output.Further error gradient is used to determine the incremen-

tal iterative change in the weight so that the actual outputapproaches the expected output.The error gradient is definedas the product of error and the rate of change in the actualoutput. Analytically it is given as

๐›ฟ๐‘˜ = ๐œ–๐‘˜๐œ•๐œ’๐‘˜๐œ•๐œ“๐‘˜

. (14)

Using the partial derivative of ๐œ’๐‘˜ and putting it in aboveequation the following equation is formed:

๐›ฟ๐‘— = ๐œ–๐‘˜๐œ’๐‘— โ‹… (1 โˆ’ ๐œ’๐‘—) . (15)

The weight of edges between input and hidden layer alsoneeds to be adjusted. For this purpose the error gradientfor hidden layer should also be calculated. In the backpropagation techniques the errors are back-tracked.The errorgradient at output layer is primarily used to calculate errorgradient at hidden layer. Here, the following equation is usedto calculate it:

๐›ฟ๐‘— = ๐œ’๐‘— โ‹… (1 โˆ’ ๐œ’๐‘—)๐‘€

โˆ‘๐‘˜=1

๐‘ค๐‘—๐‘˜๐›ฟ๐‘˜. (16)

Using these error gradients the renewed weights for neuronat each layer are computed.The following equations are used:

๐‘ค๐‘–๐‘— = ๐‘ค๐‘–๐‘— + ๐›พ โ‹… ๐‘ฅ๐‘– โ‹… ๐›ฟ๐‘—,

๐‘ค๐‘—๐‘˜ = ๐‘ค๐‘—๐‘˜ + ๐›พ โ‹… ๐œ’๐‘— โ‹… ๐›ฟ๐‘˜,(17)

where ๐›พ is the learning rate. Generally it is a tiny positivevalue lesser than 1 and is adjustable according to the learningbehavior of the network. Similarly the threshold used for

computing the renewed weights should also be recalculatedfor the next iteration. The following equations are used torecalculate the weights:

๐œƒ๐‘˜ = ๐œƒ๐‘˜ + ๐›พ โ‹… (โˆ’1) โ‹… ๐›ฟ๐‘˜, (18)

๐œƒ๐‘— = ๐œƒ๐‘— + ๐›พ โ‹… (โˆ’1) โ‹… ๐›ฟ๐‘—. (19)

Equations (18) and (19) represent the threshold for arbitraryneuron in output and hidden layer, respectively.This methodof gradient descent is quite effective and works splendidlyfor almost all sorts of problems. It has the capability tominimize the error and optimize the network to provideaccurate results. Although the training process is iterative,it ultimately needs to terminate. This termination conditionis indicated by the convergence of results. The result is saidto have converged when no appreciable change in weights ispossible. This termination condition is determined using themean square error given as

๐ธ =1

๐‘€

๐‘€

โˆ‘๐‘˜=1

(๐œ†๐‘˜ โˆ’ ๐‘ง๐‘˜)2. (20)

In the current problem a learning rate of ๐›ผ = 0.0001 wasused. The output of a recurrent neural network is not onlydependent on the current input but also dependent on theprevious output. This recurrent nature of these networksmakes them useful for problems which require a continuousinput of dynamic data changing temporally as well. Identifi-cation of an action is not necessarily dependent on a singleframe; rather previous and subsequent frames may also tell astory. Hence the use of recurrent network caters for the needfor previous temporal information [17].

5. Results and Discussion

A large database of videos was collected containing hundredsof videos of varied length. Each video contained actions like

(i) walking,(ii) clapping,(iii) hugging,(iv) single hand waving,(v) double hand waving,(vi) hand shaking.

Figure 6 shows some of the sample actions. Several videoscontaining different actions were taken under varied con-ditions in terms of illumination, location, and background.Frame by frame extraction from these videos is performed.Each frame is firstly labeled in accordance with its semanticcontent manually. Each stream of frames belonging to aspecified class is bifurcated and kept in a folder maintainingits sequence. Hence several samples of each action aresegmented from the videosmanually. Each sample is a streamof videos belonging to specific action. In the next step thebackground or the effect of background is removed from theframe. Two different strategies are followed for this purpose.With the first method, background is removed by firstlytaking a blank frame which only contains the background

The Scientific World Journal 7

Wav

e2

Wav

eW

alki

ngH

uggi

ngH

ands

hake

Cl

appi

ng

Figure 6: Action database: examples of sequences corresponding to different types of actions and scenarios.

Table 1: The numerical comparison of raw moments for each of the actions.

Clapping Handshake Hugging Walking Wave 1 Wave 2

Spatial/raw moments

1.75๐ธ + 03 4.90๐ธ + 03 7.31๐ธ + 03 6.70๐ธ + 01 1.30๐ธ + 03 9.47๐ธ + 02

4.46๐ธ + 05 1.07๐ธ + 06 1.32๐ธ + 06 1.59๐ธ + 04 2.98๐ธ + 05 1.15๐ธ + 05

3.91๐ธ + 05 8.59๐ธ + 05 1.55๐ธ + 06 6.94๐ธ + 03 4.10๐ธ + 05 3.23๐ธ + 05

9.85๐ธ + 07 2.13๐ธ + 08 3.10๐ธ + 08 1.33๐ธ + 06 9.39๐ธ + 07 3.94๐ธ + 07

1.16๐ธ + 08 2.63๐ธ + 08 2.82๐ธ + 08 4.16๐ธ + 06 7.34๐ธ + 07 1.63๐ธ + 07

9.87๐ธ + 07 2.51๐ธ + 08 5.95๐ธ + 08 2.83๐ธ + 06 1.34๐ธ + 08 1.13๐ธ + 08

2.53๐ธ + 10 5.62๐ธ + 10 6.85๐ธ + 10 3.18๐ธ + 08 2.32๐ธ + 10 5.53๐ธ + 09

2.48๐ธ + 10 6.61๐ธ + 10 1.20๐ธ + 11 4.80๐ธ + 08 3.08๐ธ + 10 1.37๐ธ + 10

3.05๐ธ + 10 6.86๐ธ + 10 6.67๐ธ + 10 1.14๐ธ + 09 1.91๐ธ + 10 2.70๐ธ + 09

2.66๐ธ + 10 9.24๐ธ + 10 2.53๐ธ + 11 1.37๐ธ + 09 4.49๐ธ + 10 3.99๐ธ + 10

and then subtracting this frame from the one containing aforeground. As a result background will be eliminated. Theother method used for this purpose takes the difference oftwo successive frames. The resultant frame will contain justthe change that occurred due to motion dynamics. Oncethe effect of background has been countered then for allresultant frames a corresponding feature vector is formed.The feature vector of a frame contains the raw, central, scale,and rotation invariant moments of the image besides itscentroid and eccentricity. Tables 1, 2, 3, 4, and 5 show thequantified values of these features. The computed vector foreach frame is fed into the recurrent neural network, iterativelytraining the network as described previously. The trainingstops when the mean square error is minimized. Not allthe database is used for the training purpose. One-fourthof database samples are not used for training; rather theyare reserved for testing. At the point when the model has

been sufficiently trained it is time to test it. The remainingsamples are similarly transformed into feature vectors and fedinto the trained model. The accuracy of the model is basedon its ability to correctly identify these untrained samples.Figure 7 represents the confusion matrix which shows thatthe overall accuracy of the system is 80.8%. Also it is noticedthat the system is better able to recognize medial frames inan action rather than initial or terminal ones. The accuracyof the system is further increased to 95% if only the accuracyof recognition for the medial frames is considered.

Various experiments were conducted to verify the accu-racy, efficiency, and effectiveness of the system in comparisonwith other competitive models. A technique described in [18]extracts the features in terms of spatial as well as temporalterms. These features are used to train SVM and henceclassify the video. The authors in [19] use a technique whichsignificantly reduces the training overhead. A patch based

8 The Scientific World Journal

Table 2: The numerical comparison of central moments for each of the actions.

Clapping Handshake Hugging Walking Wave 1 Wave 2

Central moments

1.75๐ธ + 03 4.90๐ธ + 03 7.31๐ธ + 03 6.70๐ธ + 01 1.30๐ธ + 03 9.47๐ธ + 02

0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00

0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00 0.00๐ธ + 00

โˆ’8.96๐ธ + 05 2.57๐ธ + 07 3.14๐ธ + 07 โˆ’3.23๐ธ + 05 7.40๐ธ + 04 โˆ’1.81๐ธ + 04

2.25๐ธ + 06 2.94๐ธ + 07 4.52๐ธ + 07 3.77๐ธ + 05 5.14๐ธ + 06 2.21๐ธ + 06

1.15๐ธ + 07 1.01๐ธ + 08 2.68๐ธ + 08 2.11๐ธ + 06 4.60๐ธ + 06 2.28๐ธ + 06

9.36๐ธ + 06 โˆ’1.26๐ธ + 09 โˆ’2.57๐ธ + 09 4.05๐ธ + 07 8.61๐ธ + 07 โˆ’1.40๐ธ + 07

9.23๐ธ + 07 2.15๐ธ + 09 โˆ’9.86๐ธ + 08 โˆ’1.26๐ธ + 08 1.37๐ธ + 08 9.95๐ธ + 06

โˆ’3.76๐ธ + 07 โˆ’1.80๐ธ + 09 โˆ’4.80๐ธ + 08 โˆ’2.51๐ธ + 07 โˆ’7.32๐ธ + 07 1.84๐ธ + 08

โˆ’5.31๐ธ + 08 1.30๐ธ + 10 1.40๐ธ + 10 6.42๐ธ + 08 โˆ’1.91๐ธ + 08 โˆ’1.50๐ธ + 08

Table 3: The numerical comparison of image orientation for each of the actions.

Clapping Handshake Hugging Walking Wave 1 Wave 2

Image orientation

โˆ’5.11๐ธ + 02 5.24๐ธ + 03 4.29๐ธ + 03 โˆ’4.82๐ธ + 03 5.69๐ธ + 01 โˆ’1.91๐ธ + 01

1.29๐ธ + 03 6.00๐ธ + 03 6.18๐ธ + 03 5.63๐ธ + 03 3.95๐ธ + 03 2.33๐ธ + 03

6.56๐ธ + 03 2.06๐ธ + 04 3.66๐ธ + 04 3.15๐ธ + 04 3.53๐ธ + 03 2.40๐ธ + 03

โˆ’9.58๐ธ โˆ’ 02 3.12๐ธ โˆ’ 01 1.37๐ธ โˆ’ 01 โˆ’1.78๐ธ โˆ’ 01 โˆ’1.32๐ธ โˆ’ 01 โˆ’2.49๐ธ โˆ’ 01

Table 4: The numerical comparison of scale invariant moments for each of the actions.

Clapping Handshake Hugging Walking Wave 1 Wave 2

Scale invariant

โˆ’2.91๐ธ โˆ’ 01 1.07๐ธ + 00 5.87๐ธ โˆ’ 01 โˆ’7.20๐ธ + 01 4.37๐ธ โˆ’ 02 โˆ’2.02๐ธ โˆ’ 02

7.33๐ธ โˆ’ 01 1.22๐ธ + 00 8.46๐ธ โˆ’ 01 8.40๐ธ + 01 3.04๐ธ + 00 2.46๐ธ + 00

3.74๐ธ + 00 4.20๐ธ + 00 5.01๐ธ + 00 4.70๐ธ + 02 2.72๐ธ + 00 2.54๐ธ + 00

3.04๐ธ + 00 โˆ’5.27๐ธ + 01 โˆ’4.81๐ธ + 01 9.01๐ธ + 03 5.09๐ธ + 01 โˆ’1.56๐ธ + 01

3.00๐ธ + 01 8.99๐ธ + 01 โˆ’1.84๐ธ + 01 โˆ’2.81๐ธ + 04 8.07๐ธ + 01 1.11๐ธ + 01

โˆ’1.22๐ธ + 01 โˆ’7.50๐ธ + 01 โˆ’8.99๐ธ + 00 โˆ’5.59๐ธ + 03 โˆ’4.33๐ธ + 01 2.05๐ธ + 02

โˆ’1.73๐ธ + 02 5.41๐ธ + 02 2.61๐ธ + 02 1.43๐ธ + 05 โˆ’1.13๐ธ + 02 โˆ’1.67๐ธ + 02

Table 5: The numerical comparison of rotation invariant moments for each of the actions.

Clapping Handshake Hugging Walking Wave 1 Wave 2

Rotation invariants

4.47๐ธ + 00 5.43๐ธ + 00 5.86๐ธ + 00 5.54๐ธ + 02 5.75๐ธ + 00 5.00๐ธ + 00

9.37๐ธ + 00 1.34๐ธ + 01 1.87๐ธ + 01 1.70๐ธ + 05 1.12๐ธ โˆ’ 01 7.13๐ธ โˆ’ 03

4.35๐ธ + 04 6.08๐ธ + 05 1.67๐ธ + 05 1.97๐ธ + 10 1.52๐ธ + 05 4.41๐ธ + 04

2.91๐ธ + 04 2.39๐ธ + 05 4.61๐ธ + 04 2.43๐ธ + 10 5.23๐ธ + 03 8.02๐ธ + 04

1.02๐ธ + 09 8.50๐ธ + 10 3.89๐ธ + 09 5.32๐ธ + 20 1.17๐ธ + 08 โˆ’6.09๐ธ + 09

8.91๐ธ + 04 7.40๐ธ + 05 1.72๐ธ + 05 9.97๐ธ + 12 โˆ’1.19๐ธ + 03 2.21๐ธ + 03

โˆ’2.05๐ธ + 08 3.26๐ธ + 10 1.08๐ธ + 09 3.07๐ธ + 19 1.07๐ธ + 08 โˆ’1.96๐ธ + 09

7.61๐ธ + 02 2.33๐ธ + 05 5.05๐ธ + 04 3.95๐ธ + 11 โˆ’6.43๐ธ + 02 3.20๐ธ + 03

motion descriptor and matching technique is developed bythe author. A concept of transferrable learning distance isintroduced which extracts the generic obscure knowledgewithin patches and is used to identify actions in newer videos.Both of these techniques were implemented. The accuracy ofthe proposed technique was evaluated in comparison withboth these techniques. Figure 8 shows the obtained resultswhile using the assembled action database. It can be seen that

the proposed technique performs reasonably well and ismoreconsistent as compared to other competitive techniques.

Figure 9 somehow depicts the efficiency of the system,illustrating the number of frames against time required toclassify an action. The graph shows that with the increasingnumber of frames the computed time for each frame lengthremains constant. Time required for recognition does notseem to rapidly increase if the number of frames is rapidly

The Scientific World Journal 9

Hugging

00.0%

00.0%

80.4%

10.1%

00.0%

00.0%

60.3%

452.4%

522.8%

211.1%

00.0%

40.2%

40.2%

40.2%

40.2%

30.2%

30.2%

30.2%

30.2%

70.4%

100.5%

110.6%

100.5%

120.6%

150.8%

140.7%

673.5%

27614.6%

1236.5%

22011.7%

39721.0%

30015.9%

140.7%

191.0%

20911.1%

231.2%

90.8%9.2%

87.6%12.4%

77.4%22.6%

70.8%29.2%

80.8%19.2%

81.9%18.1%

74.6%25.4%

72.3%27.7%

57.7%42.3%

94.4%5.6%

95.9%4.1%

73.6%26.4%

86.4%13.6%

Handshake

Walking

Wave 1

Wave 2

Clapping

(a)

00.0%

00.0%

00.0%

00.0%

00.0%

00.0%

00.0%

00.0%

00.0%

40.5%

00.0%

00.0%

00.0%

00.0%

00.0%

00.0%

00.0%

00.0%

20.3%

20.3%

20.3%

20.3%

20.3%

10.1%

10.1%

10.1%

10.1%

10.1%

10.1%

30.4%

13217.2%

11214.6%

648.3%

10013.0%

15720.4%

18123.5%

97.8%2.2%

97.4%2.6%

97.0%3.0%

97.3%2.7%

97.0%3.0%

96.9%3.1%

97.3%2.7%

96.2%3.8%

94.1%5.9%

98.2%1.8%

97.8%2.2%

95.2%4.8%

98.0%2.0%

Hugging

Handshake

Walking

Wave 1

Wave 2

Clapping

(b)

Figure 7: (a) The confusion matrix formed for all the frames. (b) The confusion matrix formed for medial frames.

0

10

20

30

40

50

60

70

80

90

100

Clapping Hand shaking Hugging Walking

MBFE (moment based feature extraction)Human action recognition from single clip per actionRecognition of human action by local SVM

(a)

Clapping Hand shaking Hugging Walking0

10

20

30

40

50

60

70

80

90

100

MBFE (moment based feature extraction)Human action recognition from single clip per actionRecognition of human action by local SVM

(b)

Figure 8: (a) The comparison of results using only the medial frames. (b) The comparison using all of the frames.

increased. This shows that the rapidly increasing numberof frames does not have much effect on the efficiency ofproposed algorithm.

A receiver operating characteristics (ROC) analysis is alsoperformed for videos within the database. Figure 10 givesROC graph for all the frames in the database while Figure 11gives the ROC distribution for only the medial frames inthe video database. Both graphs suggest that the accuracy ofthe proposed system is better than the current state of theart. Also the accuracy is greatly increased if only the medialframes in an action video are considered.

6. Conclusions

The paper presents a robust technique for recognition ofhuman actions. It provides a robust framework for featureextraction of frame and training of classifiers. The momentsbased feature extraction technique proves to be computa-tionally inexpensive while providing higher accuracy thancurrent state-of-the-art approaches. Hence it is more ben-eficial than other competitive techniques discussed in thepaper. Experimental data exhibits that the system has anaccuracy of 97% if used for medial frames. Furthermore

10 The Scientific World Journal

180

160

140

120

100

80

60

40

0.465 0.47 0.475 0.48 0.485 0.49

Time (s)

Num

ber o

f fra

mes

Figure 9: An analysis of number of frames required to identify anaction.

ProposedHuman action recognition using local SVMHuman action recognition using single clip per action

1

0.95

0.9

0.85

0.8

0.75

0.7

0.65

0.6

0.55

0.5

0.001 0.01 0.1 0

False acceptance rate

True

acce

ptan

ce ra

te

Figure 10: An ROC analysis for the proposed and other competitivetechniques using all the frames of video.

1

0.95

0.9

0.85

0.8

0.75

0.7

0.65

0.6

0.55

0.5

0.001 0.01 0.1 0

False acceptance rate

True

acce

ptan

ce ra

te

ProposedHuman action recognition using local SVMHuman action recognition using single clipper action

Figure 11: An ROC analysis for the proposed and other competitivetechniques using only the medial frames of video.

the experimental results show that the described system isimmune to acceptable illumination changes while dealingwith indoor and outdoor actions.

Conflict of Interests

The authors declare that there is no conflict of interestsregarding the publication of this paper.

References

[1] K. N. Tran, I. A. Kakadiaris, and S. K. Shah, โ€œPart-basedmotion descriptor image for human action recognition,โ€PatternRecognition, vol. 45, no. 7, pp. 2562โ€“2572, 2012.

[2] S. Ali and M. Shah, โ€œHuman action recognition in videosusing kinematic features and multiple instance learning,โ€ IEEETransactions on Pattern Analysis and Machine Intelligence, vol.32, no. 2, pp. 288โ€“303, 2010.

[3] O. Busaryev and J. Doolittle, Gesture Recognition with Applica-tions, Department of Computer Science and Engineering, TheOhio State University, Columbus, Ohio, USA.

[4] L. Cao, Y. L. Tian, Z. Liu, B. Yao, Z. Zhang, and T. S. Huang,โ€œAction detection usingmultiple spatial-temporal interest Pointfeatures,โ€ in Proceedings of the IEEE International Conference onMultimedia and Expo (ICME โ€™10), pp. 340โ€“345, July 2010.

[5] F. Zhang, Y. Wang, and Z. Zhang, โ€œView-invariant actionrecognition in surveillance videos,โ€ in Proceedings of the 1stAsian Conference on Pattern Recognition (ACPR โ€™11), pp. 580โ€“583, Beijing, China, November 2011.

[6] L. Shao, L. Ji, Y. Liu, and J. Zhang, โ€œHuman action segmentationand recognition via motion and shape analysis,โ€ Pattern Recog-nition Letters, vol. 33, no. 4, pp. 438โ€“445, 2012.

[7] M. M. Ullah, S. N. Parizi, and I. Laptev, โ€œImproving bag-of-features action recognition with non-local cues,โ€ in Proceed-ings of the British Machine Vision Conference (BMVC โ€™10), F.Labrosse, R. Zwiggelaar, Y. H. Liu, and B. Tiddeman, Eds., vol.10, pp. 95.1โ€“95.11, BMVA Press, September 2010.

[8] D. deMenthon andD.Doermann, โ€œVideo retrieval using spatio-temporal descriptors,โ€ in Proceedings of the 11th ACM Interna-tional Conference on Multimedia (MM '03), pp. 508โ€“517, NewYork, NY, USA, November 2003.

[9] T. Volkmer, Semantics of Video Shots for Content-BasedRetrieval, 2007.

[10] T. Guha and R. K. Ward, โ€œLearning sparse representationsfor human action recognition,โ€ IEEE Transactions on PatternAnalysis and Machine Intelligence, vol. 34, no. 8, pp. 1576โ€“1588,2012.

[11] Z. Jiang, Z. Lin, and L. Davis, โ€œRecognizing human actionsby learning and matching shape-motion prototype trees,โ€ IEEETransactions on Pattern Analysis and Machine Intelligence, vol.34, no. 3, pp. 533โ€“547, 2012.

[12] R. C. Gonzalez and R. E.Woods,Digital Image Processing, 2002.[13] J. Flusser, Z. Barbara, and T. Suk, Moments and Moment

Invariants in Pattern Recognition, JohnWiley& Sons, NewYork,NY, USA, 2009.

[14] J. Flusser, B. Zitova, and T. Suk, Moments and Moment Invari-ants in Pattern Recognition, John Wiley & Sons, New York, NY,USA, 2009.

[15] H. Jose Antonio Martฤฑn, M. Santos, and J. de Lope, โ€œOrthog-onal variant moments features in image analysis,โ€ InformationSciences, vol. 180, no. 6, pp. 846โ€“860, 2010.

[16] H. Ming-Kuei, โ€œVisual pattern recognition by moment invari-ants,โ€ IRE Transactions on Information Theory, vol. 8, no. 2, pp.179โ€“187, 1962.

[17] B. Coppin, Artificial Intelligence Illuminated, Jones & BartlettLearning, Sudbury, Mass, USA, 2004.

[18] C. Schuldt, I. Laptev, and B. Caputo, โ€œRecognizing humanactions: a local SVM approach,โ€ in Proceedings of the 17thInternational Conference on Pattern Recognition (ICPR โ€™04), vol.3, pp. 32โ€“36, IEEE, August 2004.