Due to concerns like backdrop cluttering, incomplete obstruction, scale disparities, viewpoint, illumination, and appearance, identifying activities of humans from a sequence of video or still photos are a complex issue. Multiple movement recognition structures is necessary for numerous applications, such as a video investigation mechanism, human computer interface HCI , and robotics for characterising human behaviour. In this work, we bestow a comprehensive assessment of recent and advanced designs involved in the classification of human activity. We outline a classification of human activity approaches and go through their benefits and drawbacks. Specifically, we classify human activities categorization approaches into two broad classes based on if or not they make use of information from several modes. Next, each of these classes is broken down into its subclasses, which illustrate how each category models human activity. Monisa Nazir | Shalini Bhadola | Kirti Bhaia | Rohini Sharma "A Complete Analysis of Human Action Recognition Procedures" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-5 , August 2022, URL: https://www.ijtsrd.com/papers/ijtsrd50522.pdf Paper URL: https://www.ijtsrd.com/computer-science/cognitive-science/50522/a-complete-analysis-of-human-action-recognition-procedures/monisa-nazir
2. International Journal of Trend in Scientific Research and Development @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD50522 | Volume – 6 | Issue – 5 | July-August 2022 Page 594
has been shown to be essential in games console design, individual fitness tracing, sporting study, and
operational and behavioral well-being monitoring to name a few. Human activity recognition is a branch of
Image processing [2-6] and improvement [7-8].
A prototype in the source can use the data it has retrieved from that area to skill a model in the final domain with
less labelled data. A smaller number of data may be needed to skill the prototype in the target domain as a result
of applying knowledge from the prior skilled prototype. The authors [9] have examined transferring information
among the models with various probability distributions and have enlarged on the various moods, data labeling
method, and nomenclature of kind of information conveyed in transfer learning. The accuracy of the model was
utilised as an assessment parameter by the writers to assess the concert of their AR models, which were
evaluated using the RF and DT(Decision Tree) algorithms. The authors demonstrated that a recognition
prototype can be skilled through a smaller number of examples by testing this methodology on the HAR [10],
Daily and Sports [11], and MHealth [12] datasets. The Likelihood distribution of the device data differs
significantly between peoples, and if a prototype is trained specifically for a person, its performance will suffer.
HUMAN ACTIVITY RECOGNITION (HAR) METHODS
Sensor-based HAR
To mimic a diversity of human actions, it combines the newly developed field of sensor networks with cutting-
edge DM (data mining) and ML methods [13]. Mobile gadgets, like smartphones, have enough sensor data and
processing capability to recognise physical activity and estimate how much energy is used in daily life.
Researchers in sensor-centered movement identification think that by giving omnipresent computers and sensors
the ability to watch agents' behaviour (with their permission), these computers will be better able to represent us.
Sensor-based HAR is depicted in Fig 2.
Sensor level, many people action identification
Recognising movements for many peoples employing body sensors originally present in [14]. In order to detect
patterns of group activity in office environments, acceleration sensors and other sensor technology were used.
Gu et al. investigate the activities of several users in intelligent environments [15]. In this study, they look at the
basic issue of recognizing multi-user activities from sensor readings in a home setting, and they present a fresh
pattern burrowing method to identify both single-user and multi-user activities in a unified way.
Sensor-level assembly action identification
The purpose of group activity recognition is to identify the behaviour of the assembly as an entity rather than the
actions of the group's separate members, which makes it fundamentally distinct from one or many user activity
identification. Group behaviour is emergent, which means that its characteristics are essentially distinct from
those of the individuals who make up the group or any combination of those characteristics [16]. The key
difficulties are in describing the behaviour of the individual group members as well as their roles within the
dynamics of the group and their connections to parallel emergent group behaviour. The quantization of the
3. International Journal of Trend in Scientific Research and Development @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD50522 | Volume – 6 | Issue – 5 | July-August 2022 Page 595
behaviour and roles of people that join the group, the incorporation of explicit models for role description into
inference methods, and extensibility assessments for very large groups and crowds are all issues that still need to
be resolved. Crowd control, emergency response, social networking.
Data Mining Based HAR
A data mining (DM) strategy has recently been offered as an alternative to conventional ML procedures. The
issue of HAR is presented as a pattern-based cataloging issue. To identify sequential, overlapping, and
continuous processes in a unified solution, they suggested a DM method based on discriminatory patterns that
characterize substantial variations between any two activity groups of data. Authors make use of 2D corners in
time and space. Through a hierarchical method and an expanding search field, these are assembled both spatial-
temporal [17]. Data mining is effectively used to learn the most distinguishing and descriptive features at each
stage of the hierarchy. Fig 2 provides a Data Mining procedure for HAR Recognition.
Fig 2: Data mining Based HAR
Machine Learning (ML) Centered HAR
The study of ML methods for HAR issues has increased recently since they are particularly successful at
learning from and extracting insight from the activity records. The methods include hand-made, trait-centered
conventional ML procedures obtained by heuristics as well as more recently created hierarchically self-evolving
functionality based deep learning methods. Considering all of the research that has been done in this area, AR
still presents a difficult difficulty in unmanaged smart environments. Numerous difficulties are presented by the
activity data's complicated, unstable, and chaotic character, which have an impact on how well HAR systems
work in the forest. The success of the ML techniques is heavily reliant on the data representation in HAR
systems created via shallow learning and often employed attribute heuristics are based on the researcher's
domain expertise [17]. The Deep learning (DL) uses some nonlinear transformation to learn the attributes from
the raw data gradationally. The kind of deep learning network is determined by the nonlinear transformation.
Recently, DL has received approval, and several works have used deep learning in augmented reality. The DL
networks are some of the most used deep learning approaches in augmented reality. The impact of using deep
learning to analyse time series activity data has been thoroughly discussed in [44]. Authors in [18] investigated
DL approaches for the activity dataset came to the conclusion that RNN outperformed the most recent results for
OPPORTUNITY [19].
Authors in [20] have suggested a unique deep network composed of convolutional layers. The writers tuned the
hyper-parameters of their network and fused the different sensors paradigms like acceleration, tachometer,
compass in possible permutations and comparing the outcomes by assessing them using OPPORTUNITY
dataset. The authors documented 30 different users' ADLs, then A multilayer CNN framework with alternate
convolutional and pooling layers was introduced by authors [21], who demonstrated that their system
outperformed the existing ADL accuracy benchmark. When the short-term fourier transform of the
accelerometer data was used as an input to the suggested CNN network, Ravi et al. produced accuracy that was
virtually on par with cutting-edge outcomes.
A (RBM) based activity identification model for smart watches was created and developed by [22], who also
demonstrated that the model is hardware independent. OPPORTUNITY datasets, which use various cutting-edge
classifiers for each of the activities, are real-world data sets that the authors used to validate their correctness.
The most recent advancements in deep learning [23] proposed a novel method of using ensembles of deep
LSTM networks with input from wearable sensors. Since this is the only study to use a combination of deep
learning approaches, more research in this area is possible. Employing activities of daily living sensor data
gathered from the wrist and waist, authors in [24] analysed the hand-crafted features and the CNN derived
4. International Journal of Trend in Scientific Research and Development @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD50522 | Volume – 6 | Issue – 5 | July-August 2022 Page 596
features using kNN. The researcher will have a better understanding of deep learning networks by studying the
features obtained from a deep learning network in more detail and contrasting them with heuristic features. A
multichannel Framework for numerous sensor data has been proposed by San et al. [25]. The accuracy of
activity recognition has significantly increased according to the authors' evaluations using Decision Tree, kNN,
and Naive Bayes classifiers. Fig 3 illustrates the machine learning based HAR.
Fig 3: Machine Learning Based Basic Human Activity Recognition
CONCLUSION
A difficult time series classification task is human
activity recognition, or HAR. In order to accurately
engineer features from the raw data in order to build a
machine learning model, it generally requires
extensive domain understanding and methodologies
from signal processing. It entails predicting a person's
movement based on sensor data. By automatically
extracting features from the unprocessed sensor data,
deep learning techniques like convolutional neural
networks and recurrent neural networks have recently
proven capable and even attain cutting-edge
outcomes.
REFERENCES
[1] Aggarwal, J.K. and Ryoo, M.S., Human
activity analysis: A review. ACM Computing
Surveys (CSUR), , 2011, 43 (3), p. 16.
[2] Jasdeep Singh, Kirti Bhatia, Shalini Bhadola,
Rohini Sharma, An Analytical Survey on
Dimension Reduction Based Face Recognition
Systems, International Journal of
Multidisciplinary Research in Science,
Engineering, Technology & Management
(IJMRSETM), Volume 9, Issue 7, July 2022,
pp.1493-1498.
[3] Akash Dagar, Princy, Kirti Bhatia, Rohini
Sharma, Demosaicing Techniques Thorough
Analysis, International Journal of Innovative
Research in Science, Engineering, and
Technology (IJIRSET), Volume 11, Issue 7,
July 2022, pp.9902-9907.
[4] Jyoti, Kirti Bhatia, Shalini Bhadola, Rohini
Sharma, An Analysis of Facsimile
Demosaicing Procedures, International journal
of Innovative Research in computer and
communication engineering, Vol-08, Issue-07,
July 2020.
[5] Deepak Dahiya, Kirti Bhatia, Rohini Sharma,
Shalini Bhadola, A Deep Overview on Image
Denoising Approaches, International Journal of
Innovative Research in Computer and
Communication Engineering, Volume 9, Issue
7, July 2021.
[6] Nakul Nalwa, Shalini Bhadola, Kirti Bhatia,
Rohini Sharma, A Detailed Study on People
Tracking Methodologies in Different Scenarios,
International Journal of Innovative Research in
Computer and Communication Engineering,
Volume 10, Issue 6, June 2022.
[7] Mahesh Kumar Attri, Kirti Bhatia, Shalini
Bhadola, Rohini Sharma, An Image Sharpening
and Smoothing Approaches Analysis,
International Journal of Innovative Research in
Computer and Communication Engineering,
Volume 10, Issue 6, June 2022, pp 5573-5580.
[8] Yogita, Shalini Bhadola, Kirti Bhatia, Rohini
Sharma, A Deep Overview of Image Inpainting
5. International Journal of Trend in Scientific Research and Development @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD50522 | Volume – 6 | Issue – 5 | July-August 2022 Page 597
Approaches, International Journal of Innovative
Research in Science, Engineering and
Technology, Volume 11, Issue 6, June 2022,
pp. 8744-8749.
[9] Khan, M. A. A. H. and Roy, N., Transact:
Transfer learning enabled activity recognition.
In Pervasive Computing and Communications
Workshops (PerComWorkshops), 2017 IEEE
International Conference on, 2011, pp. 545–
550.
[10] Anguita, D., Ghio, A., Oneto, L., Parra, X. and
Reyes-Ortiz, J. L., A public domain dataset for
human activity recognition using smartphones.
In ESANN, (2013).
[11] Barshan, B. and Yüksek, M. C., Recognizing
daily and sports activities in two open source
machine learning environments using body-
worn sensor units. The Computer Journal,
(2013), 57, pp. 1649–1667.
[12] Banos,O., Garcia, R., Holgado-Terriza, J. A.,
Damas, M., Pomares, H., Rojas, I., Saez, A. and
Villalonga, C. mhealthdroid: a novel
framework for agile development of mobile
health applications. In International Workshop
on Ambient Assisted Living, 2014, pp. 91–98.
[13] Choudhury, T., and Borriello, G., The Mobile
Sensing Platform: An Embedded System for
Activity Recognition. IEEE Pervasive
Magazine – Special Issue on Activity-Based
Computing, 2008.
[14] Want R., Hopper A., Falcao V., Gibbons J.
(1992). The Active Badge Location System,
ACM Transactions on Information, Systems, 40
(1), pp. 91–102.
[15] Gu, T., Wu, Z., Wang, L., Tao, X. and Lu., J.
(2009). Mining Emerging Patterns for
Recognizing Activities of Multiple Users in
Pervasive Computing, In Proc. of the 6th
International Conference on Mobile and
Ubiquitous Systems: Computing, Networking
and Services (MobiQuitous '09), Toronto,
Canada, July 13–16, 2009.
[16] Gu, T., Wu, Z., Tao, X., Pung, H. and Lu.,
J.(20090. epSICAR: An Emerging Patterns
based Approach to Sequential, Interleaved and
Concurrent Activity Recognition. In Proc. of
the 7th Annual IEEE International Conference
on Pervasive Computing and Communications
(Percom '09), Galveston, Texas, March 9–13,
2009.
[17] Bengio, Y. (2013) Deep learning of
representations: Looking forward. In
International Conference on Statistical
Language and Speech Processing, 2013, pp.1–
37.
[18] Hammerla, N. Y., Halloran, S. and Ploetz, T.
(2016) Deep, convolutional, and recurrent
models for human activity recognition using
wearables. arXiv preprint arXiv:1604.08880.
[19] Roggen, D., Forster, K., Calatroni, A.,
Holleczek, T., Fang, Y., Troster, G., Ferscha,
A., Holzmann, C., Riener, A., Lukowicz, P. et
al. (2009) Opportunity: Towards opportunistic
activity and context recognition systems. In
World of Wireless, Mobile and Multimedia
Networks &Workshops, 2009.WoWMoM
2009. IEEE International Symposium on a, 1–6.
IEEE.
[20] Ordóñez, F. J. and Roggen, D. (2016) Deep
convolutional and lstm recurrent neural
networks for multimodal wearable activity
recognition. Sensors, 16, 115.
[21] Ronao, C. A. and Cho, S.-B. (2016) Human
activity recognition with smartphone sensors
using deep learning neural networks. Expert
Systems with Applications, 59, pp. 235–244.
[22] Bhattacharya, S. and Lane, N.D. (2016) from
smart to deep: Robust activity recognition on
smartwatches using deep learning. In Pervasive
Computing and Communication Workshops
(PerComWorkshops), 2016 IEEE International
Conference on, 1–6. IEEE.
[23] Panwar, M., Dyuthi, S. R., Prakash, K. C.,
Biswas, D., Acharyya, A., Maharatna, K.,
Gautam, A. and Naik, G. R. (2017) Cnn based
approach for activity recognition using a wrist-
worn accelerometer. In Engineering in
Medicine and Biology Society (EMBC), 2017
39th Annual International Conference of the
IEEE, pp. 2438–2441.
[24] Sani, S., Wiratunga, N. and Massie, S. (2017)
Learning deep features for knn-based human
activity recognition.
[25] San, P. P., Kakar, P., Li, X.-L., Krishnaswamy,
S., Yang, J.-B. and Nguyen, M. N. (2017) Deep
learning for human activity recognition.