SlideShare a Scribd company logo
1 of 9
Download to read offline
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
ACTIVITY RECOGNITION USING HISTOGRAM OF 
ORIENTED GRADIENT PATTERN HISTORY 
Dipankar Das 
Department of Information and Communication Engineering, University of Rajshahi, 
Rajshahi-6205, Bangladesh 
ABSTRACT 
Human activity recognition is an important task in computer vision because it has many application areas 
such as, healthcare, security, entertainment, and tactical scenarios. This paper presents a methodology to 
automatically recognize human activity from input video stream using Histogram of Oriented Gradient 
Pattern History (HOGPH) features and SVM classifier. For this purpose, the proposed system extracts 
HOG features from a sequence of consecutive video frames and analyzes them to construct HOGPH feature 
vector. The HOGPH feature vectors are used to train a multi-class SVM classifier for different human 
activities. In test mode, we use the classifier with HOGPH feature vector to recognize human activity. We 
have experimented with video data of human activity in real environments for three different tasks 
(browsing, reading, and writing). The experimental result and its accuracy reveal that the proposed system 
is applicable to recognize human activity in real-life. 
KEYWORDS 
Human Activity, Activity Recognition, Histogram of Oriented Gradient, SVM. 
1. INTRODUCTION 
Recognizing human activity or task from real-time video data is one of the promising and 
challenging applications of computer vision. Recently, this research has attracted the attention of 
many researchers from different disciplines including Human-Computer-Interaction (HCI), and 
Human-Robot-Interaction (HRI). The main objective of human task or activity recognition is to 
provide useful information on a user’s behaviour that allows computing system or robot to 
proactively assist users or to make a comfortable interaction with him/her [1]. Researchers in 
computer vision and machine learning have investigated gestures and activity recognition from 
static images and video in constrained environment or stationary settings[2][3][4]. However, there 
are very limited number of works to recognize human task or activity in unconstrained daily life 
settings. In this research, we recognize human task or acitivity in unconstrained real-world 
environments, such as Internet browsing in an office environment, reading and writing in an 
library environment etc., using Histogram of Oriented Gradient Pattern History (HOGPH) and 
Support Vector Machine (SVM) classifier. 
Human activity is very complex and diverse characteristics because an activity can be performed 
in many different ways, depending on the different context and for a multitude of reasons. 
Although some state-of-the-art system obtained good performance on many activity recognition 
tasks, however, most of the researchers so far mainly focus on recognizing “which” activity is 
DOI : 10.5121/ijcseit.2014.4403 23
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
being performed at a specific point in time. In contrast, only small number of researches 
investigated means to extract qualitative information from the sensor data that allows us to infer 
additional characteristics. It can be shown that such qualitative assessment are very difficult to 
perform automatically, and has so far only been demonstrated for constrained settings, such as 
sports[5][6]. For general activities or tasks, activity recognition research work is still far from 
reaching a similar understanding. However, in this research, we proposed an activity recognition 
system that tries to assess the qualitative characteristics of the human activity. For this purpose, 
the proposed activity recognition system first extracts HOG features of activity during a certain 
period of time to build up an activity pattern known as the histogram of oriented gradient pattern 
history (HOGPH). Then the HOGPH feature vector are used to classify the human activity using 
support vector machine (SVM) classifier. 
To develop a good human activity recognition system, we need to develop a pattern recognition 
system that is robust to intra-class variability. In activity, such type of intra-class variability exists 
because same type of activity may be executed different ways by different person. Intra-class 
variability can also happen if an activity is performed by the same person at different time. Thus, 
for activity recognition, we need a classifier that adapts this type of intra-class variability. To 
address this issue, the proposed system uses the multi-class SVM classifier with an increase 
amount of training data for different activity classes using liblinear kernel. The proposed HOGPH 
features are highly discriminative and person independent that also make help the system to adapt 
intra-class variability. 
24 
2. RELATED WORKS 
Since there are some successful researches in activity recognition, many researchers are 
motivated toward more challenging and application-specific scenarios. Some real-world 
application are benefited from task or activity recognition such as the office scenarios, the sports 
and entertainment sector[7][8], the industrial sector[9][10], and healthcare. The activity of daily 
living [11] attracted a significant amount of attention of the vision researchers [12] [13] [14]. The 
traditional medical methods can be enhanced by analyzing the daily human activities [6]. Another 
important aim of human activity analysis is to provide a healthy lifestyle. Thus many researchers 
are encouraged to research in the related human activities, e.g. food intake [15] and 
medication[16], brushing teeth or hand washing[17], or transportation routines[18]. 
Currently different companies or agencies use activity recognition as a key component in their 
consumer products. For instant, to fundamentally change the game experience in game consoles 
the NintendoWii or the Microsoft Kinect rely on the recognition of gestures or even full body 
movements. Although they are mainly developed for the entertainment purpose, however, these 
systems have used in other application areas, such as for personal fitness training and 
rehabilitation, and also stimulated new activity recognition research [19]. These examples 
highlight the importance of human activity or task recognition in both academic area and 
industrial section. In spite of considerable improvement in inferring activities from on-body 
inertial sensors and in prototyping and deploying activity recognition systems [20], developing 
human activity recognition systems that meet application and user requirements remains a 
challenging task. In this research, we recognize human task or acitivity in unconstrained real-world 
environments, such as Internet browsing in an office environment, reading and writing in 
an library environment etc., using Histogram of Oriented Gradient Pattern History (HOGPH) and 
SVM classifer.
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
In human activity recognition system, Aggarwal et al.[21] propose state-space and template 
matching based techniques. Some researchers use context-based decisions making techniques to 
recognize human activity and action [22][23]. In [22], the authors propose an action recognition 
technique that includes entering a room, using a computer, opening a file cabinet, picking up the 
telephone, etc. In their system, they able to recognize actions based on prior knowledge about the 
layout of the room. However, the system is limited to actions like sitting and standing. Also, it is 
only able to recognize a picking action by knowledge of where the object is and tracking it after 
the person has come within a certain distance of it. Moreover, both of the above contextual based 
approaches require prior knowledge of the precise location of certain objects in the environment. 
Davis et al. use temporal plates for matching and recognition of human activity [24]. The system 
computes motion history images (MHI's) of the persons in the scene. Although template matching 
procedures have a lower computational cost, they are usually more sensitive to the variance in the 
duration of the movement. 
25 
3. PROPOSED APPROACH 
The proposed human activity recognition system is illustrated in Figure 1 . In our approach, we 
recoded the video data of human activity, such as browsing, reading, and writing, in a real-world 
environment. The recorded video data is divided into two sections: training activity data, and 
testing activity data. The proposed system is also divided into two phases: training and testing. In 
training stage, we extract the histogram of oriented gradient features (HOG) [25] from each video 
frame. 
FIGURE 1: THE PROPOSED HUMAN ACTIVITY RECOGNITION APPROACH.
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
The HOG feature vectors from n consecutive video frames are analyzed to generate the histogram 
of oriented gradient pattern history (HOGPH). The generated HOGPH feature vectors are used to 
train the multi-class SVM classifier model for all activities. In testing stage, for each human 
activity we generate the HOGPH feature vector and feed into the SVM model in recognition 
mode. The model classifies the human activity into an appropriate class. 
26 
4. EXTRACTION OF HOG FEATURE 
In this research, we use the histogram of oriented gradient (HOG) feature from each video frame 
to represent human activity. The method is originally developed by N. Dalal and B. Triggs in [25] 
for human detection purposes. Their method is based on evaluating well-normalized local 
histograms of image gradient orientations in a dense grid. The fundamental principle is that local 
object shape and appearance can often be represented rather well by the distribution of local 
intensity gradients, even without precise knowledge of the corresponding gradient or edge 
positions. In practical implementation, we divide the image window into small spatial regions 
known as cell. Then for each cell, we build up a local 1-D histogram of gradient directions over 
the pixels of the cell. The combined histogram entries form the gradient orientation 
representation. The histogram is also useful to contrast-normalize the local responses to make it 
invariance to illumination, and shadowing. This can be done by building up a measure of local 
histogram energy over somewhat larger spatial regions (block) and using the results to normalize 
all of the cells in the block. Here the normalized descriptor blocks are known as Histogram of 
Oriented Gradient (HOG) descriptors. 
5. CONSTRUCTION OF HOGPH FEATURE VECTOR 
The human activity is determined by recognizing the pattern of task history in which the target 
person is involved. For instance, if the target person is involved in a “reading” task, then the 
pattern history such as “downward the head” indicates that his/her attention is toward the book. 
Such type of activity related pattern is determined by analysing HOG feature vector as follows. 
Given a video sequence, we extract the histogram of orientation gradient (HOG) feature for each 
frame. The HOG features are combined for n consecutive frames to build a HOG feature pattern 
history, HOGPH according to the following equation,
where HF0 and HFi are the HOG features of the first and i-th frames, respectively. The first frame 
captures the human appearance features involved in a task, and the rest of the HOG feature 
frames indicate the change of behaviour pattern in the activity during the observation history. 
Thus, the HOGPH feature captures the appearance of the activity and the activity related 
behaviour pattern history. Here, each bin in the histogram represents the number of edges that 
have the orientation within a given angular range. The angular range is set to 20 degrees and we 
use unsigned gradient. Thus, the bin size is equal to180/20 = 9. With this bin size, if we create the 
HOGPH feature vector for 10 consecutive video frames (i = 1,.....,9) then the HOGPH feature 
vector is of size 90. A multi-class support vector machine (SVM) is learned using HOGPH 
feature.
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
6. ACTIVITY CLASSIFICATION 
LASSIFICATION 
In our activity recognition approach 
learned with HOGPH features 
descriptors are extracted from the 
pattern history we analyze the subsequent HOG feature and calculate the difference of two 
consecutive frames. Here the act 
represented by orientations of an edge histogram within an object's subregion quantized into 
bin and each edge's contribution is weighted by its magnitude. Therefore, each bin in the 
histogram represents the number of edges that have orientations within a given angular range. 
final pattern history of human activity is represented by the HOGPH feature vector. 
class SVM classifier [26] is learned using 
for our experiments in a multi-class mode with the 
approach, a multi-class support vector machine (SVM) classifier is 
vector. To represent the appearance of an activity 
first frame of an activity image. In order to describe the 
activity, HOG 
. activity 
activity pattern is represented by its local shape. The local shape is 
In the verification step, the HOGPH feature vector is extracted from activity video frames and 
into the multi-class SVM classifier in recognition mode. Only the hypotheses for which a positive 
confidence measurement is returned are kept for each 
confidence level is recognized as the correct 
probabilistic output of the SVM classifier. 
7. EXPERIMENTAL RESULTS 
We conducted experiments to investigate the performance of the vision 
activity when s/he is involved in a task 
7.1. Experimental Data 
We implemented the proposed system and conducted experiments to verify 
performance. To collect experimental data, we asked 14 
females) for three different tasks: browsing, readin 
Saitama University, Japan, with an average age of 27 years old. They did not receive any 
information for their activity or task 
average recorded lengths for each person for the 
8, 9, and 9 minutes, respectively. 
were involved in browsing, reading and writing tasks, respectively. 
Figure2: Human involved in activities 
togram HOGPH feature vector. We use the LIBSVM 
liblinear and rbf exponential kernel 
activity. Activity with 
activity. The confidence level is measured using the 
ESULTS 
vision-based system to 
task. 
activity recognition 
. 14-non-paid participants (11 males and 3 
reading, and writing. They were graduate students at 
task. All the tasks were recorded using a video camera. The 
or activities of browsing, reading, and writing 
Figure 2 show some snapshot of activities where the humans 
browsing, reading, and writing, respectively 
27 
. K-bin 
The 
The multi-use 
package 
kernels. 
fed 
the highest 
idence human 
g, . were 
, respectively.
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
28 
7.2. Training and Testing the SVM Classifier 
Each of the recorded tasks was divided into test and train sample videos. As described in Section 
5, one sample features a vector consisting of ten frames of the video stream. Table 1 shows the 
number of training and testing sample used in the experiments and their total length for three 
different activities. The multi-class SVM classifier was learned using the training samples for 
three activities. In the recognition mode, each test sample value was tested against three tasks and 
the task with the highest probability was selected as the detected task. Total 8740 test samples for 
three activities were used to test the system. The SVM classifier was used with two different 
kernels: liblinear and rbf. The performance of the classifier was compared for different kernels on 
the proposed activity test sample dataset. 
Table1: Sample Training and Test Data for Experiments 
Activity Training Data Test Data 
Length 
(in Min) 
No. of 
Samples 
Length 
(in Min) 
No. of Samples 
Browsing 19.30 2482 22.93 3104 
Reading 16.00 2058 19.10 2904 
Writing 15.00 1929 19.91 2732 
Total 50.30 6469 61.94 8740 
7.3. Activity Recognition Performance 
We measured the activity recognition performance on the test video data. From the test video 
data, we used 8740 test activity samples. First, we measured the performance of the SVM 
classifier with rbf exponential kernel. In this case, among 8740 test samples 6064 activity samples 
were correctly recognized with 69.38% accuracy. Table 2 shows the activity specific recognition 
performance and cross category confusion of activity recognition system using the rbf 
exponential kernel. 
Table 2: Activity Recognition Performance Using SVM Classifier and rbf Exponential Kernel 
Activity Browsing Reading Writing 
Browsing 2439 302 363 
Reading 318 1935 651 
Writing 278 764 1690 
Second, the performance of the proposed system was also measured using SVM classifier and 
liblinear kernel. Here, among 8740 test samples 8277 activity samples were correctly recognized 
with 94.70% recognition accuracy. Table 3 shows the activity specific recognition performance 
and cross category confusion of activity recognition system using the liblinear kernel. From the 
experimental result, it was shown that the proposed system performed significant improvement 
for human activity recognition task with liblinear kernel than rbf kernel. The liblinear kernel was 
also decreased the cross-category confusion rate. Figure 3 shows some snapshots of the 
recognized human activities for “browsing”, “reading”, and “writing”, respectively.
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
29 
Table 3: Activity Recognition Performance Using SVM Classifier and liblinear Kernel 
Activity Browsing Reading Writing 
Browsing 3101 3 0 
Reading 28 2448 428 
Writing 0 4 2728 
FIGURE 3: SNAPSHOT OF RECOGNIZED ACTIVITIES 
8. CONCLUSION 
In this paper, we propose an automatic human activity recognition system that can accurately 
recognize three different activities (browsing, reading, and writing) from real-life video data. The 
proposed system uses the HOG pattern history (HOGPH) feature vectors that represent human 
activity and its behavior accurately and precisely. We analyze and combine the HOG features of 
human activity for few consecutive video frames. The first frame represents appearance of the 
activity and the remaining frames indicate the change of behavior pattern during activity. The 
HOGPH feature vectors are used to learn and classify human activity using SVM classifier. In 
this research, the performance of the classifier is compared using two different kernels: rbf and 
liblinear. The experimental result shows that the proposed system produces higher accuracy with
International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 
liblinear kernel than rbf kernel. In future, we want to extend and compare the proposed system 
for human activity recognition with more versatile databases. 
30 
REFERENCES 
[1] G. D. Abowd, A. K. Dey, R. Orr and J. Brotherton (1998), “Context-awareness in wearable and 
ubiquitous computing,” Virtual Reality, vol. 3, no. 3, pp. 200-211. 
[2] S. Mitra and T. Acharya (2007), “Gesture Recognition: A Survey,” IEEE Trans. on Systems, Man, 
and Cybernetics, Part C: Applications and Reviews, vol. 37, no. 3, pp. 311-324. 
[3] P. Turaga, R. Chellappa, V. S. Subrahmanian and O. Udrea (2008), “ Machine Recognition of Human 
Activities: A Survey,” IEEE Trans. on Circuits and Systems for Video Technology, vol. 18, no. 11, 
pp. 1473-1488. 
[4] J. K. Aggarwal and M. S. Ryoo (2011), “ Human activity analysis: A review,” Comput. Surveys , vol. 
43, no. 3, pp. 16:1-16:43. 
[5] E. Velloso, A. Bulling, H. Gellersen, W. Ugulino and H. Fuks (2013), “Qualitative Activity 
Recognition of Weight Lifting Exercises,” in 4th Augmented Human International Conference 
(AugmentedHuman 2013). 
[6] B. Tessendorf, F. Gravenhorst, B. Arnrich and G. Troster (2011), “An IMU-based Sensor Network to 
Continuously Monitor Rowing Technique on the Water,” in ISSNIP. 
[7] K. Kunze, M. Barry, E. Heinz, P. Lukowicz, D. Majoe and J. Gutknecht (2006), “Towards 
Recognizing Tai Chi–An Initial Experiment Using Wearable Sensors,” in FAWC. 
[8] D. Minnen, T. Starner, I. Essa and C. Isbell (2006), “Discovering Characteristic Actions from On- 
Body Sensor Data,” in 10th IEEE International Symposium on Wearable Computers (ISWC). 
[9] I. Maurtua, P. T. Kirisci, T. Stiefmeier, M. L. Sbodio and H.Witt (2007), “A Wearable Computing 
Prototype for Supporting Training Activities in Automative Production,” in 4th International Forum 
on Applied Wearable Computing. 
[10] T. Stiefmeier, D. Roggen, G. Ogris, P. Lukowicz and G. Troster (2008), “Wearable Activity Tracking 
in Car Manufacturing,” IEEE Pervasive Computing, vol. 7, no. 2, pp. 42-50. 
[11] S. Katz, T. Downs, H. Cash and R. Grotz (1970), “ Progress in development of the index of ADL,” 
The Gerontologist, vol. 10, no. 1, p. 20. 
[12] L. Bao and S. S. Intille (2004), “Activity Recognition from User-Annotated Acceleration Data,” in 
Pervasive. 
[13] N. Ravi, N. Dandekar, P. Mysore and M. L. Littman (2005), “Activity recognition from accelerometer 
data,” in 17th International Conference on Innovative Applications of Artificial Intelligence. 
[14] B. Logan, J. Healey, M. Philipose, E. Tapia and S. Intille (2007), “A long-term evaluation of sensing 
modalities for activity recognition,” in UbiComp. 
[15] G. Pirkl, K. Stockinger, K. Kunze and P. Lukowicz (2008), “Adapting magnetic resonant coupling 
based relative positioning technology for wearable activitiy recogniton,” in ISWC. 
[16] R. d. Oliveira, M. Cherubini and N. Oliver (2010), “MoviPill: Improving Medication Compliance for 
Elders Using a Mobile Persuasive Social Game,” in UbiComp. 
[17] J. Lester, T. Choudhury and G. Borriello (2006), “A Practical Approach to Recognizing Physical 
Activities,” in International Conference on Pervasive Computing. 
[18] J. Krumm and E. Horvitz (2006), “Predestination: Inferring Destinations from Partial Trajectories,” in 
UbiComp. 
[19] J. Sung, C. Ponce, B. Selman and A. Saxena (2011), “ Human Activity Detection from RGBD 
Images,” in AAAI Workshop on Plan, Activity, and Intent Recognition. 
[20] D. Ashbrook and T. Starner (2010), “MAGIC: a motion gesture design tool,” in CHI. 
[21] J. K. Aggarwal and Q. Cai (1999), “ Human motion analysis:A review,” Computer Vision and Image 
Understanding, pp. 428-440. 
[22] D. Ayers and M. Shah (1998), “Recognizing human action in a static room,” in Computer Vision and 
Pattern Recognition. 
[23] S. S. Intille, J. Davis and A. Bobick (1997), “Real time closed world tracking,” in IEEE International 
Conference on Computer Vision and Pattern Recognition. 
[24] J. Davis and A. Bobick (1997), “The representation and recognition of action using temporal plates,” 
in Computer Vision and Pattern Recognition.

More Related Content

What's hot

An Efficient Activity Detection System based on Skeleton Joints Identification
An Efficient Activity Detection System based on Skeleton Joints Identification An Efficient Activity Detection System based on Skeleton Joints Identification
An Efficient Activity Detection System based on Skeleton Joints Identification IJECEIAES
 
A virtual analysis on various techniques using ann with data mining
A virtual analysis on various techniques using ann with data miningA virtual analysis on various techniques using ann with data mining
A virtual analysis on various techniques using ann with data miningeSAT Journals
 
Automatic Emotion Recognition Using Facial Expression: A Review
Automatic Emotion Recognition Using Facial Expression: A ReviewAutomatic Emotion Recognition Using Facial Expression: A Review
Automatic Emotion Recognition Using Facial Expression: A ReviewIRJET Journal
 
human activity recognization using machine learning with data analysis
human activity recognization using machine learning with data analysishuman activity recognization using machine learning with data analysis
human activity recognization using machine learning with data analysisVenkat Projects
 
MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG
MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG
MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG sipij
 
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...IRJET Journal
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity RecognitionAshwinGill1
 
TOURIST PLACE RECOMMENDATION SYSTEM
TOURIST PLACE RECOMMENDATION SYSTEMTOURIST PLACE RECOMMENDATION SYSTEM
TOURIST PLACE RECOMMENDATION SYSTEMIJARIIT
 
A Review on Feature Extraction Techniques and General Approach for Face Recog...
A Review on Feature Extraction Techniques and General Approach for Face Recog...A Review on Feature Extraction Techniques and General Approach for Face Recog...
A Review on Feature Extraction Techniques and General Approach for Face Recog...Editor IJCATR
 
Optimized Biometric System Based on Combination of Face Images and Log Transf...
Optimized Biometric System Based on Combination of Face Images and Log Transf...Optimized Biometric System Based on Combination of Face Images and Log Transf...
Optimized Biometric System Based on Combination of Face Images and Log Transf...sipij
 
Feature based head pose estimation for controlling movement of
Feature based head pose estimation for controlling movement ofFeature based head pose estimation for controlling movement of
Feature based head pose estimation for controlling movement ofIAEME Publication
 
Novel Hybrid Approach to Visual Concept Detection Using Image Annotation
Novel Hybrid Approach to Visual Concept Detection Using Image AnnotationNovel Hybrid Approach to Visual Concept Detection Using Image Annotation
Novel Hybrid Approach to Visual Concept Detection Using Image AnnotationCSCJournals
 
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...Scale Invariant Feature Transform Based Face Recognition from a Single Sample...
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...ijceronline
 
An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...
An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...
An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...ijtsrd
 
Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01Ai Zhang
 

What's hot (18)

An Efficient Activity Detection System based on Skeleton Joints Identification
An Efficient Activity Detection System based on Skeleton Joints Identification An Efficient Activity Detection System based on Skeleton Joints Identification
An Efficient Activity Detection System based on Skeleton Joints Identification
 
Dq4301702706
Dq4301702706Dq4301702706
Dq4301702706
 
A virtual analysis on various techniques using ann with data mining
A virtual analysis on various techniques using ann with data miningA virtual analysis on various techniques using ann with data mining
A virtual analysis on various techniques using ann with data mining
 
Automatic Emotion Recognition Using Facial Expression: A Review
Automatic Emotion Recognition Using Facial Expression: A ReviewAutomatic Emotion Recognition Using Facial Expression: A Review
Automatic Emotion Recognition Using Facial Expression: A Review
 
human activity recognization using machine learning with data analysis
human activity recognization using machine learning with data analysishuman activity recognization using machine learning with data analysis
human activity recognization using machine learning with data analysis
 
Facial Expression Recognitino
Facial Expression RecognitinoFacial Expression Recognitino
Facial Expression Recognitino
 
MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG
MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG
MULTI SCALE ICA BASED IRIS RECOGNITION USING BSIF AND HOG
 
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...
Recognition of Silverleaf Whitefly and Western Flower Thrips Via Image Proces...
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity Recognition
 
TOURIST PLACE RECOMMENDATION SYSTEM
TOURIST PLACE RECOMMENDATION SYSTEMTOURIST PLACE RECOMMENDATION SYSTEM
TOURIST PLACE RECOMMENDATION SYSTEM
 
G0333946
G0333946G0333946
G0333946
 
A Review on Feature Extraction Techniques and General Approach for Face Recog...
A Review on Feature Extraction Techniques and General Approach for Face Recog...A Review on Feature Extraction Techniques and General Approach for Face Recog...
A Review on Feature Extraction Techniques and General Approach for Face Recog...
 
Optimized Biometric System Based on Combination of Face Images and Log Transf...
Optimized Biometric System Based on Combination of Face Images and Log Transf...Optimized Biometric System Based on Combination of Face Images and Log Transf...
Optimized Biometric System Based on Combination of Face Images and Log Transf...
 
Feature based head pose estimation for controlling movement of
Feature based head pose estimation for controlling movement ofFeature based head pose estimation for controlling movement of
Feature based head pose estimation for controlling movement of
 
Novel Hybrid Approach to Visual Concept Detection Using Image Annotation
Novel Hybrid Approach to Visual Concept Detection Using Image AnnotationNovel Hybrid Approach to Visual Concept Detection Using Image Annotation
Novel Hybrid Approach to Visual Concept Detection Using Image Annotation
 
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...Scale Invariant Feature Transform Based Face Recognition from a Single Sample...
Scale Invariant Feature Transform Based Face Recognition from a Single Sample...
 
An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...
An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...
An Improved Self Organizing Feature Map Classifier for Multimodal Biometric R...
 
Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01Thermal Imaging Emotion Recognition final report 01
Thermal Imaging Emotion Recognition final report 01
 

Similar to Activity recognition using histogram of

ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORYACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORYijcseit
 
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORYACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORYijcseit
 
IRJET- Human Activity Recognition using Flex Sensors
IRJET- Human Activity Recognition using Flex SensorsIRJET- Human Activity Recognition using Flex Sensors
IRJET- Human Activity Recognition using Flex SensorsIRJET Journal
 
A Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition ProceduresA Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition Proceduresijtsrd
 
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...IRJET Journal
 
Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964Editor IJARCET
 
Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964Editor IJARCET
 
Yoga Posture Classification using Computer Vision
Yoga Posture Classification using Computer VisionYoga Posture Classification using Computer Vision
Yoga Posture Classification using Computer VisionDr. Amarjeet Singh
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity RecognitionIRJET Journal
 
IRJET- Survey on Detection of Crime
IRJET-  	  Survey on Detection of CrimeIRJET-  	  Survey on Detection of Crime
IRJET- Survey on Detection of CrimeIRJET Journal
 
Real-time Activity Recognition using Smartphone Accelerometer
Real-time Activity Recognition using Smartphone AccelerometerReal-time Activity Recognition using Smartphone Accelerometer
Real-time Activity Recognition using Smartphone Accelerometerijtsrd
 
IoT Based Human Activity Recognition and Classification Using Machine Learning
IoT Based Human Activity Recognition and Classification Using Machine LearningIoT Based Human Activity Recognition and Classification Using Machine Learning
IoT Based Human Activity Recognition and Classification Using Machine LearningIRJET Journal
 
A Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal FeaturesA Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal FeaturesCSCJournals
 
Chapter 1_Introduction.docx
Chapter 1_Introduction.docxChapter 1_Introduction.docx
Chapter 1_Introduction.docxKISHWARYA2
 
Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...
Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...
Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...Nelson J. S. Silva
 
Human Activity Recognition (HAR) Using Opencv
Human Activity Recognition (HAR) Using OpencvHuman Activity Recognition (HAR) Using Opencv
Human Activity Recognition (HAR) Using OpencvIRJET Journal
 
Study of AI Fitness Model Using Deep Learning
Study of AI Fitness Model Using Deep LearningStudy of AI Fitness Model Using Deep Learning
Study of AI Fitness Model Using Deep LearningIRJET Journal
 
Human activity recognition updated 1 - Copy.pptx
Human activity recognition updated 1 - Copy.pptxHuman activity recognition updated 1 - Copy.pptx
Human activity recognition updated 1 - Copy.pptxBhaveshKhuje
 
Health monitoring catalogue based on human activity classification using mac...
Health monitoring catalogue based on human activity  classification using mac...Health monitoring catalogue based on human activity  classification using mac...
Health monitoring catalogue based on human activity classification using mac...IJECEIAES
 
Survey on Human Behavior Recognition using CNN
Survey on Human Behavior Recognition using CNNSurvey on Human Behavior Recognition using CNN
Survey on Human Behavior Recognition using CNNIRJET Journal
 

Similar to Activity recognition using histogram of (20)

ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORYACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
 
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORYACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY
 
IRJET- Human Activity Recognition using Flex Sensors
IRJET- Human Activity Recognition using Flex SensorsIRJET- Human Activity Recognition using Flex Sensors
IRJET- Human Activity Recognition using Flex Sensors
 
A Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition ProceduresA Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition Procedures
 
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
 
Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964
 
Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964Volume 2-issue-6-1960-1964
Volume 2-issue-6-1960-1964
 
Yoga Posture Classification using Computer Vision
Yoga Posture Classification using Computer VisionYoga Posture Classification using Computer Vision
Yoga Posture Classification using Computer Vision
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity Recognition
 
IRJET- Survey on Detection of Crime
IRJET-  	  Survey on Detection of CrimeIRJET-  	  Survey on Detection of Crime
IRJET- Survey on Detection of Crime
 
Real-time Activity Recognition using Smartphone Accelerometer
Real-time Activity Recognition using Smartphone AccelerometerReal-time Activity Recognition using Smartphone Accelerometer
Real-time Activity Recognition using Smartphone Accelerometer
 
IoT Based Human Activity Recognition and Classification Using Machine Learning
IoT Based Human Activity Recognition and Classification Using Machine LearningIoT Based Human Activity Recognition and Classification Using Machine Learning
IoT Based Human Activity Recognition and Classification Using Machine Learning
 
A Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal FeaturesA Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal Features
 
Chapter 1_Introduction.docx
Chapter 1_Introduction.docxChapter 1_Introduction.docx
Chapter 1_Introduction.docx
 
Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...
Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...
Leveraging Eye-gaze and Time-series Features to Predict User Interests and Bu...
 
Human Activity Recognition (HAR) Using Opencv
Human Activity Recognition (HAR) Using OpencvHuman Activity Recognition (HAR) Using Opencv
Human Activity Recognition (HAR) Using Opencv
 
Study of AI Fitness Model Using Deep Learning
Study of AI Fitness Model Using Deep LearningStudy of AI Fitness Model Using Deep Learning
Study of AI Fitness Model Using Deep Learning
 
Human activity recognition updated 1 - Copy.pptx
Human activity recognition updated 1 - Copy.pptxHuman activity recognition updated 1 - Copy.pptx
Human activity recognition updated 1 - Copy.pptx
 
Health monitoring catalogue based on human activity classification using mac...
Health monitoring catalogue based on human activity  classification using mac...Health monitoring catalogue based on human activity  classification using mac...
Health monitoring catalogue based on human activity classification using mac...
 
Survey on Human Behavior Recognition using CNN
Survey on Human Behavior Recognition using CNNSurvey on Human Behavior Recognition using CNN
Survey on Human Behavior Recognition using CNN
 

Recently uploaded

Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilV3cube
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Servicegiselly40
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure servicePooja Nehwal
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc
 

Recently uploaded (20)

Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of Brazil
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 

Activity recognition using histogram of

  • 1. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 ACTIVITY RECOGNITION USING HISTOGRAM OF ORIENTED GRADIENT PATTERN HISTORY Dipankar Das Department of Information and Communication Engineering, University of Rajshahi, Rajshahi-6205, Bangladesh ABSTRACT Human activity recognition is an important task in computer vision because it has many application areas such as, healthcare, security, entertainment, and tactical scenarios. This paper presents a methodology to automatically recognize human activity from input video stream using Histogram of Oriented Gradient Pattern History (HOGPH) features and SVM classifier. For this purpose, the proposed system extracts HOG features from a sequence of consecutive video frames and analyzes them to construct HOGPH feature vector. The HOGPH feature vectors are used to train a multi-class SVM classifier for different human activities. In test mode, we use the classifier with HOGPH feature vector to recognize human activity. We have experimented with video data of human activity in real environments for three different tasks (browsing, reading, and writing). The experimental result and its accuracy reveal that the proposed system is applicable to recognize human activity in real-life. KEYWORDS Human Activity, Activity Recognition, Histogram of Oriented Gradient, SVM. 1. INTRODUCTION Recognizing human activity or task from real-time video data is one of the promising and challenging applications of computer vision. Recently, this research has attracted the attention of many researchers from different disciplines including Human-Computer-Interaction (HCI), and Human-Robot-Interaction (HRI). The main objective of human task or activity recognition is to provide useful information on a user’s behaviour that allows computing system or robot to proactively assist users or to make a comfortable interaction with him/her [1]. Researchers in computer vision and machine learning have investigated gestures and activity recognition from static images and video in constrained environment or stationary settings[2][3][4]. However, there are very limited number of works to recognize human task or activity in unconstrained daily life settings. In this research, we recognize human task or acitivity in unconstrained real-world environments, such as Internet browsing in an office environment, reading and writing in an library environment etc., using Histogram of Oriented Gradient Pattern History (HOGPH) and Support Vector Machine (SVM) classifier. Human activity is very complex and diverse characteristics because an activity can be performed in many different ways, depending on the different context and for a multitude of reasons. Although some state-of-the-art system obtained good performance on many activity recognition tasks, however, most of the researchers so far mainly focus on recognizing “which” activity is DOI : 10.5121/ijcseit.2014.4403 23
  • 2. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 being performed at a specific point in time. In contrast, only small number of researches investigated means to extract qualitative information from the sensor data that allows us to infer additional characteristics. It can be shown that such qualitative assessment are very difficult to perform automatically, and has so far only been demonstrated for constrained settings, such as sports[5][6]. For general activities or tasks, activity recognition research work is still far from reaching a similar understanding. However, in this research, we proposed an activity recognition system that tries to assess the qualitative characteristics of the human activity. For this purpose, the proposed activity recognition system first extracts HOG features of activity during a certain period of time to build up an activity pattern known as the histogram of oriented gradient pattern history (HOGPH). Then the HOGPH feature vector are used to classify the human activity using support vector machine (SVM) classifier. To develop a good human activity recognition system, we need to develop a pattern recognition system that is robust to intra-class variability. In activity, such type of intra-class variability exists because same type of activity may be executed different ways by different person. Intra-class variability can also happen if an activity is performed by the same person at different time. Thus, for activity recognition, we need a classifier that adapts this type of intra-class variability. To address this issue, the proposed system uses the multi-class SVM classifier with an increase amount of training data for different activity classes using liblinear kernel. The proposed HOGPH features are highly discriminative and person independent that also make help the system to adapt intra-class variability. 24 2. RELATED WORKS Since there are some successful researches in activity recognition, many researchers are motivated toward more challenging and application-specific scenarios. Some real-world application are benefited from task or activity recognition such as the office scenarios, the sports and entertainment sector[7][8], the industrial sector[9][10], and healthcare. The activity of daily living [11] attracted a significant amount of attention of the vision researchers [12] [13] [14]. The traditional medical methods can be enhanced by analyzing the daily human activities [6]. Another important aim of human activity analysis is to provide a healthy lifestyle. Thus many researchers are encouraged to research in the related human activities, e.g. food intake [15] and medication[16], brushing teeth or hand washing[17], or transportation routines[18]. Currently different companies or agencies use activity recognition as a key component in their consumer products. For instant, to fundamentally change the game experience in game consoles the NintendoWii or the Microsoft Kinect rely on the recognition of gestures or even full body movements. Although they are mainly developed for the entertainment purpose, however, these systems have used in other application areas, such as for personal fitness training and rehabilitation, and also stimulated new activity recognition research [19]. These examples highlight the importance of human activity or task recognition in both academic area and industrial section. In spite of considerable improvement in inferring activities from on-body inertial sensors and in prototyping and deploying activity recognition systems [20], developing human activity recognition systems that meet application and user requirements remains a challenging task. In this research, we recognize human task or acitivity in unconstrained real-world environments, such as Internet browsing in an office environment, reading and writing in an library environment etc., using Histogram of Oriented Gradient Pattern History (HOGPH) and SVM classifer.
  • 3. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 In human activity recognition system, Aggarwal et al.[21] propose state-space and template matching based techniques. Some researchers use context-based decisions making techniques to recognize human activity and action [22][23]. In [22], the authors propose an action recognition technique that includes entering a room, using a computer, opening a file cabinet, picking up the telephone, etc. In their system, they able to recognize actions based on prior knowledge about the layout of the room. However, the system is limited to actions like sitting and standing. Also, it is only able to recognize a picking action by knowledge of where the object is and tracking it after the person has come within a certain distance of it. Moreover, both of the above contextual based approaches require prior knowledge of the precise location of certain objects in the environment. Davis et al. use temporal plates for matching and recognition of human activity [24]. The system computes motion history images (MHI's) of the persons in the scene. Although template matching procedures have a lower computational cost, they are usually more sensitive to the variance in the duration of the movement. 25 3. PROPOSED APPROACH The proposed human activity recognition system is illustrated in Figure 1 . In our approach, we recoded the video data of human activity, such as browsing, reading, and writing, in a real-world environment. The recorded video data is divided into two sections: training activity data, and testing activity data. The proposed system is also divided into two phases: training and testing. In training stage, we extract the histogram of oriented gradient features (HOG) [25] from each video frame. FIGURE 1: THE PROPOSED HUMAN ACTIVITY RECOGNITION APPROACH.
  • 4. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 The HOG feature vectors from n consecutive video frames are analyzed to generate the histogram of oriented gradient pattern history (HOGPH). The generated HOGPH feature vectors are used to train the multi-class SVM classifier model for all activities. In testing stage, for each human activity we generate the HOGPH feature vector and feed into the SVM model in recognition mode. The model classifies the human activity into an appropriate class. 26 4. EXTRACTION OF HOG FEATURE In this research, we use the histogram of oriented gradient (HOG) feature from each video frame to represent human activity. The method is originally developed by N. Dalal and B. Triggs in [25] for human detection purposes. Their method is based on evaluating well-normalized local histograms of image gradient orientations in a dense grid. The fundamental principle is that local object shape and appearance can often be represented rather well by the distribution of local intensity gradients, even without precise knowledge of the corresponding gradient or edge positions. In practical implementation, we divide the image window into small spatial regions known as cell. Then for each cell, we build up a local 1-D histogram of gradient directions over the pixels of the cell. The combined histogram entries form the gradient orientation representation. The histogram is also useful to contrast-normalize the local responses to make it invariance to illumination, and shadowing. This can be done by building up a measure of local histogram energy over somewhat larger spatial regions (block) and using the results to normalize all of the cells in the block. Here the normalized descriptor blocks are known as Histogram of Oriented Gradient (HOG) descriptors. 5. CONSTRUCTION OF HOGPH FEATURE VECTOR The human activity is determined by recognizing the pattern of task history in which the target person is involved. For instance, if the target person is involved in a “reading” task, then the pattern history such as “downward the head” indicates that his/her attention is toward the book. Such type of activity related pattern is determined by analysing HOG feature vector as follows. Given a video sequence, we extract the histogram of orientation gradient (HOG) feature for each frame. The HOG features are combined for n consecutive frames to build a HOG feature pattern history, HOGPH according to the following equation,
  • 5. where HF0 and HFi are the HOG features of the first and i-th frames, respectively. The first frame captures the human appearance features involved in a task, and the rest of the HOG feature frames indicate the change of behaviour pattern in the activity during the observation history. Thus, the HOGPH feature captures the appearance of the activity and the activity related behaviour pattern history. Here, each bin in the histogram represents the number of edges that have the orientation within a given angular range. The angular range is set to 20 degrees and we use unsigned gradient. Thus, the bin size is equal to180/20 = 9. With this bin size, if we create the HOGPH feature vector for 10 consecutive video frames (i = 1,.....,9) then the HOGPH feature vector is of size 90. A multi-class support vector machine (SVM) is learned using HOGPH feature.
  • 6. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 6. ACTIVITY CLASSIFICATION LASSIFICATION In our activity recognition approach learned with HOGPH features descriptors are extracted from the pattern history we analyze the subsequent HOG feature and calculate the difference of two consecutive frames. Here the act represented by orientations of an edge histogram within an object's subregion quantized into bin and each edge's contribution is weighted by its magnitude. Therefore, each bin in the histogram represents the number of edges that have orientations within a given angular range. final pattern history of human activity is represented by the HOGPH feature vector. class SVM classifier [26] is learned using for our experiments in a multi-class mode with the approach, a multi-class support vector machine (SVM) classifier is vector. To represent the appearance of an activity first frame of an activity image. In order to describe the activity, HOG . activity activity pattern is represented by its local shape. The local shape is In the verification step, the HOGPH feature vector is extracted from activity video frames and into the multi-class SVM classifier in recognition mode. Only the hypotheses for which a positive confidence measurement is returned are kept for each confidence level is recognized as the correct probabilistic output of the SVM classifier. 7. EXPERIMENTAL RESULTS We conducted experiments to investigate the performance of the vision activity when s/he is involved in a task 7.1. Experimental Data We implemented the proposed system and conducted experiments to verify performance. To collect experimental data, we asked 14 females) for three different tasks: browsing, readin Saitama University, Japan, with an average age of 27 years old. They did not receive any information for their activity or task average recorded lengths for each person for the 8, 9, and 9 minutes, respectively. were involved in browsing, reading and writing tasks, respectively. Figure2: Human involved in activities togram HOGPH feature vector. We use the LIBSVM liblinear and rbf exponential kernel activity. Activity with activity. The confidence level is measured using the ESULTS vision-based system to task. activity recognition . 14-non-paid participants (11 males and 3 reading, and writing. They were graduate students at task. All the tasks were recorded using a video camera. The or activities of browsing, reading, and writing Figure 2 show some snapshot of activities where the humans browsing, reading, and writing, respectively 27 . K-bin The The multi-use package kernels. fed the highest idence human g, . were , respectively.
  • 7. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 28 7.2. Training and Testing the SVM Classifier Each of the recorded tasks was divided into test and train sample videos. As described in Section 5, one sample features a vector consisting of ten frames of the video stream. Table 1 shows the number of training and testing sample used in the experiments and their total length for three different activities. The multi-class SVM classifier was learned using the training samples for three activities. In the recognition mode, each test sample value was tested against three tasks and the task with the highest probability was selected as the detected task. Total 8740 test samples for three activities were used to test the system. The SVM classifier was used with two different kernels: liblinear and rbf. The performance of the classifier was compared for different kernels on the proposed activity test sample dataset. Table1: Sample Training and Test Data for Experiments Activity Training Data Test Data Length (in Min) No. of Samples Length (in Min) No. of Samples Browsing 19.30 2482 22.93 3104 Reading 16.00 2058 19.10 2904 Writing 15.00 1929 19.91 2732 Total 50.30 6469 61.94 8740 7.3. Activity Recognition Performance We measured the activity recognition performance on the test video data. From the test video data, we used 8740 test activity samples. First, we measured the performance of the SVM classifier with rbf exponential kernel. In this case, among 8740 test samples 6064 activity samples were correctly recognized with 69.38% accuracy. Table 2 shows the activity specific recognition performance and cross category confusion of activity recognition system using the rbf exponential kernel. Table 2: Activity Recognition Performance Using SVM Classifier and rbf Exponential Kernel Activity Browsing Reading Writing Browsing 2439 302 363 Reading 318 1935 651 Writing 278 764 1690 Second, the performance of the proposed system was also measured using SVM classifier and liblinear kernel. Here, among 8740 test samples 8277 activity samples were correctly recognized with 94.70% recognition accuracy. Table 3 shows the activity specific recognition performance and cross category confusion of activity recognition system using the liblinear kernel. From the experimental result, it was shown that the proposed system performed significant improvement for human activity recognition task with liblinear kernel than rbf kernel. The liblinear kernel was also decreased the cross-category confusion rate. Figure 3 shows some snapshots of the recognized human activities for “browsing”, “reading”, and “writing”, respectively.
  • 8. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 29 Table 3: Activity Recognition Performance Using SVM Classifier and liblinear Kernel Activity Browsing Reading Writing Browsing 3101 3 0 Reading 28 2448 428 Writing 0 4 2728 FIGURE 3: SNAPSHOT OF RECOGNIZED ACTIVITIES 8. CONCLUSION In this paper, we propose an automatic human activity recognition system that can accurately recognize three different activities (browsing, reading, and writing) from real-life video data. The proposed system uses the HOG pattern history (HOGPH) feature vectors that represent human activity and its behavior accurately and precisely. We analyze and combine the HOG features of human activity for few consecutive video frames. The first frame represents appearance of the activity and the remaining frames indicate the change of behavior pattern during activity. The HOGPH feature vectors are used to learn and classify human activity using SVM classifier. In this research, the performance of the classifier is compared using two different kernels: rbf and liblinear. The experimental result shows that the proposed system produces higher accuracy with
  • 9. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 liblinear kernel than rbf kernel. In future, we want to extend and compare the proposed system for human activity recognition with more versatile databases. 30 REFERENCES [1] G. D. Abowd, A. K. Dey, R. Orr and J. Brotherton (1998), “Context-awareness in wearable and ubiquitous computing,” Virtual Reality, vol. 3, no. 3, pp. 200-211. [2] S. Mitra and T. Acharya (2007), “Gesture Recognition: A Survey,” IEEE Trans. on Systems, Man, and Cybernetics, Part C: Applications and Reviews, vol. 37, no. 3, pp. 311-324. [3] P. Turaga, R. Chellappa, V. S. Subrahmanian and O. Udrea (2008), “ Machine Recognition of Human Activities: A Survey,” IEEE Trans. on Circuits and Systems for Video Technology, vol. 18, no. 11, pp. 1473-1488. [4] J. K. Aggarwal and M. S. Ryoo (2011), “ Human activity analysis: A review,” Comput. Surveys , vol. 43, no. 3, pp. 16:1-16:43. [5] E. Velloso, A. Bulling, H. Gellersen, W. Ugulino and H. Fuks (2013), “Qualitative Activity Recognition of Weight Lifting Exercises,” in 4th Augmented Human International Conference (AugmentedHuman 2013). [6] B. Tessendorf, F. Gravenhorst, B. Arnrich and G. Troster (2011), “An IMU-based Sensor Network to Continuously Monitor Rowing Technique on the Water,” in ISSNIP. [7] K. Kunze, M. Barry, E. Heinz, P. Lukowicz, D. Majoe and J. Gutknecht (2006), “Towards Recognizing Tai Chi–An Initial Experiment Using Wearable Sensors,” in FAWC. [8] D. Minnen, T. Starner, I. Essa and C. Isbell (2006), “Discovering Characteristic Actions from On- Body Sensor Data,” in 10th IEEE International Symposium on Wearable Computers (ISWC). [9] I. Maurtua, P. T. Kirisci, T. Stiefmeier, M. L. Sbodio and H.Witt (2007), “A Wearable Computing Prototype for Supporting Training Activities in Automative Production,” in 4th International Forum on Applied Wearable Computing. [10] T. Stiefmeier, D. Roggen, G. Ogris, P. Lukowicz and G. Troster (2008), “Wearable Activity Tracking in Car Manufacturing,” IEEE Pervasive Computing, vol. 7, no. 2, pp. 42-50. [11] S. Katz, T. Downs, H. Cash and R. Grotz (1970), “ Progress in development of the index of ADL,” The Gerontologist, vol. 10, no. 1, p. 20. [12] L. Bao and S. S. Intille (2004), “Activity Recognition from User-Annotated Acceleration Data,” in Pervasive. [13] N. Ravi, N. Dandekar, P. Mysore and M. L. Littman (2005), “Activity recognition from accelerometer data,” in 17th International Conference on Innovative Applications of Artificial Intelligence. [14] B. Logan, J. Healey, M. Philipose, E. Tapia and S. Intille (2007), “A long-term evaluation of sensing modalities for activity recognition,” in UbiComp. [15] G. Pirkl, K. Stockinger, K. Kunze and P. Lukowicz (2008), “Adapting magnetic resonant coupling based relative positioning technology for wearable activitiy recogniton,” in ISWC. [16] R. d. Oliveira, M. Cherubini and N. Oliver (2010), “MoviPill: Improving Medication Compliance for Elders Using a Mobile Persuasive Social Game,” in UbiComp. [17] J. Lester, T. Choudhury and G. Borriello (2006), “A Practical Approach to Recognizing Physical Activities,” in International Conference on Pervasive Computing. [18] J. Krumm and E. Horvitz (2006), “Predestination: Inferring Destinations from Partial Trajectories,” in UbiComp. [19] J. Sung, C. Ponce, B. Selman and A. Saxena (2011), “ Human Activity Detection from RGBD Images,” in AAAI Workshop on Plan, Activity, and Intent Recognition. [20] D. Ashbrook and T. Starner (2010), “MAGIC: a motion gesture design tool,” in CHI. [21] J. K. Aggarwal and Q. Cai (1999), “ Human motion analysis:A review,” Computer Vision and Image Understanding, pp. 428-440. [22] D. Ayers and M. Shah (1998), “Recognizing human action in a static room,” in Computer Vision and Pattern Recognition. [23] S. S. Intille, J. Davis and A. Bobick (1997), “Real time closed world tracking,” in IEEE International Conference on Computer Vision and Pattern Recognition. [24] J. Davis and A. Bobick (1997), “The representation and recognition of action using temporal plates,” in Computer Vision and Pattern Recognition.
  • 10. International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol. 4, No.4, August 2014 [25] N. Dalal and B. Triggs (2005), “Histograms of oriented gradients for human detection,” in CVPR (1). [26] I. Tsochantaridis, T. Joachims, T. Hofmann and Y. Altun (2005), “Large margin methods for structured and interdependent output variables,” Journal of Machine Learning Research, vol. 6, pp. 1453-1484. 31 Authors Dipankar Das received his B.Sc. and M.Sc. degree in Computer Science and Technology from the University of Rajshahi, Rajshahi, Bangladesh in 1996 and 1997, respectively. He also received his PhD degree in Computer Vision from Saitama University Japan in 2010. He was a Postdoctoral fellow in Robot Vision from October 2011 to March 2014 at the same university. He is currently working as an associate professor of the Department of Information and Communication Engineering, University of Rajshahi. His research interests include Object Recognition and Human Computer Interaction.