This paper presents and discusses the creation of a sound event classification model using deep learning. In the design of service robots, it is necessary to include routines that improve the response of both the robot and the human being throughout the interaction. These types of tasks are critical when the robot is taking care of children, the elderly, or people in vulnerable situations. Certain dangerous situations are difficult to identify and assess by an autonomous system, and yet, the life of the users may depend on these robots. Acoustic signals correspond to events that can be detected at a great distance, are usually present in risky situations, and can be continuously sensed without incurring privacy risks. For the creation of the model, a customized database is structured with seven categories that allow to categorize a problem, and eventually allow the robot to provide the necessary help. These audio signals are processed to produce graphical representations consistent with human acoustic identification. These images are then used to train three convolutional models identified as high-performing in this type of problem. The three models are evaluated with specific metrics to identify the best-performing model. Finally, the results of this evaluation are discussed and analyzed.
Scheme for motion estimation based on adaptive fuzzy neural networkTELKOMNIKA JOURNAL
Many applications of robots in collaboration with humans require the robot to follow the person autonomously. Depending on the tasks and their context, this type of tracking can be a complex problem. The paper proposes and evaluates a principle of control of autonomous robots for applications of services to people, with the capacity of prediction and adaptation for the problem of following people without the use of cameras (high level of privacy) and with a low computational cost. A robot can easily have a wide set of sensors for different variables, one of the classic sensors in a mobile robot is the distance sensor. Some of these sensors are capable of collecting a large amount of information sufficient to precisely define the positions of objects (and therefore people) around the robot, providing objective and quantitative data that can be very useful for a wide range of tasks, in particular, to perform autonomous tasks of following people. This paper uses the estimated distance from a person to a service robot to predict the behavior of a person, and thus improve performance in autonomous person following tasks. For this, we use an adaptive fuzzy neural network (AFNN) which includes a fuzzy neural network based on Takagi-Sugeno fuzzy inference, and an adaptive learning algorithm to update the membership functions and the rule base. The validity of the proposal is verified both by simulation and on a real prototype. The average RMSE of prediction over the 50 laboratory tests with different people acting as target object was 7.33.
LEARNING OF ROBOT NAVIGATION TASKS BY PROBABILISTIC NEURAL NETWORKcscpconf
This paper reports results of artificial neural network for robot navigation tasks. Machine learning methods have proven usability in many complex problems concerning mobile robots
control. In particular we deal with the well-known strategy of navigating by “wall-following”. In this study, probabilistic neural network (PNN) structure was used for robot navigation tasks.
The PNN result was compared with the results of the Logistic Perceptron, Multilayer Perceptron, Mixture of Experts and Elman neural networks and the results of the previous
studies reported focusing on robot navigation tasks and using same dataset. It was observed the PNN is the best classification accuracy with 99,635% accuracy using same dataset.
This paper reports results of artificial neural network for robot navigation tasks. Machine
learning methods have proven usability in many complex problems concerning mobile robots
control. In particular we deal with the well-known strategy of navigating by “wall-following”.
In this study, probabilistic neural network (PNN) structure was used for robot navigation tasks.
The PNN result was compared with the results of the Logistic Perceptron, Multilayer
Perceptron, Mixture of Experts and Elman neural networks and the results of the previous
studies reported focusing on robot navigation tasks and using same dataset. It was observed the
PNN is the best classification accuracy with 99,635% accuracy using same dataset.
LEARNING OF ROBOT NAVIGATION TASKS BY PROBABILISTIC NEURAL NETWORKcsandit
This paper reports results of artificial neural network for robot navigation tasks. Machine
learning methods have proven usability in many complex problems concerning mobile robots
control. In particular we deal with the well-known strategy of navigating by “wall-following”.
In this study, probabilistic neural network (PNN) structure was used for robot navigation tasks.
The PNN result was compared with the results of the Logistic Perceptron, Multilayer
Perceptron, Mixture of Experts and Elman neural networks and the results of the previous
studies reported focusing on robot navigation tasks and using same dataset. It was observed the
PNN is the best classification accuracy with 99,635% accuracy using same dataset.
A novel visual tracking scheme for unstructured indoor environmentsIJECEIAES
In the ever-expanding sphere of assistive robotics, the pressing need for advanced methods capable of accurately tracking individuals within unstructured indoor settings has been magnified. This research endeavours to devise a realtime visual tracking mechanism that encapsulates high performance attributes while maintaining minimal computational requirements. Inspired by the neural processes of the human brain’s visual information handling, our innovative algorithm employs a pattern image, serving as an ephemeral memory, which facilitates the identification of motion within images. This tracking paradigm was subjected to rigorous testing on a Nao humanoid robot, demonstrating noteworthy outcomes in controlled laboratory conditions. The algorithm exhibited a remarkably low false detection rate, less than 4%, and target losses were recorded in merely 12% of instances, thus attesting to its successful operation. Moreover, the algorithm’s capacity to accurately estimate the direct distance to the target further substantiated its high efficacy. These compelling findings serve as a substantial contribution to assistive robotics. The proficient visual tracking methodology proposed herein holds the potential to markedly amplify the competencies of robots operating in dynamic, unstructured indoor settings, and set the foundation for a higher degree of complex interactive tasks.
Non-contact, camera-based physiological measurements -
such as blood volume pulse and respiration rate - can now
be inferred by neural networks based on facial videos. This
technology has the potential to enable medical professionals
to make more informed telehealth decisions. Currently, this
software only runs on PCs, without a user interface. The neu-
ral network has a significant computational cost, making it
difficult to deploy on low-cost mobile devices. It also performs
poorly in varied environmental, sensor, personal, and contex-
tual conditions - such as darker skin tones. In this project,
we implement this neural network as an Android app that
runs in real-time; develop a more efficient architecture; evalu-
ate these architectures on older smartphones; and provide an
open-source, simple personalization pipeline to enable users to
calibrate the app. This all serves to make the technology more
democratic: making it available to as many users as possible,
while giving them the means to train and develop it further.
DYNAMIC AND REALTIME MODELLING OF UBIQUITOUS INTERACTIONcscpconf
Ubiquitous systems require user to be dynamically and realtime informed in order to make his current activity increasingly easy. First, this paper presents and discusses a method to model the realtime interaction of the user with a ubiquitous system based on Petri-nets modelling technology. The goal deals with investigating dynamically the appropriate form of interaction depending on the context of the user. Thus, the interaction model structure should be dynamically improved with respect to the current and particular activity or goal of the user to better cope with his runtime requirements. This mechanism has been characterized as “models mutation”. Secondly, this paper proves the dynamic construction of models while basing on the dynamic composition of services. The ultimate purpose is to take advantage of the ontology of service written in OWL-S in order to describe the dynamic aspect of Petri-nets based models, especially, the realtime and automatic composition of such models. Simulation work has been conducted to validate the proposed approach.
Scheme for motion estimation based on adaptive fuzzy neural networkTELKOMNIKA JOURNAL
Many applications of robots in collaboration with humans require the robot to follow the person autonomously. Depending on the tasks and their context, this type of tracking can be a complex problem. The paper proposes and evaluates a principle of control of autonomous robots for applications of services to people, with the capacity of prediction and adaptation for the problem of following people without the use of cameras (high level of privacy) and with a low computational cost. A robot can easily have a wide set of sensors for different variables, one of the classic sensors in a mobile robot is the distance sensor. Some of these sensors are capable of collecting a large amount of information sufficient to precisely define the positions of objects (and therefore people) around the robot, providing objective and quantitative data that can be very useful for a wide range of tasks, in particular, to perform autonomous tasks of following people. This paper uses the estimated distance from a person to a service robot to predict the behavior of a person, and thus improve performance in autonomous person following tasks. For this, we use an adaptive fuzzy neural network (AFNN) which includes a fuzzy neural network based on Takagi-Sugeno fuzzy inference, and an adaptive learning algorithm to update the membership functions and the rule base. The validity of the proposal is verified both by simulation and on a real prototype. The average RMSE of prediction over the 50 laboratory tests with different people acting as target object was 7.33.
LEARNING OF ROBOT NAVIGATION TASKS BY PROBABILISTIC NEURAL NETWORKcscpconf
This paper reports results of artificial neural network for robot navigation tasks. Machine learning methods have proven usability in many complex problems concerning mobile robots
control. In particular we deal with the well-known strategy of navigating by “wall-following”. In this study, probabilistic neural network (PNN) structure was used for robot navigation tasks.
The PNN result was compared with the results of the Logistic Perceptron, Multilayer Perceptron, Mixture of Experts and Elman neural networks and the results of the previous
studies reported focusing on robot navigation tasks and using same dataset. It was observed the PNN is the best classification accuracy with 99,635% accuracy using same dataset.
This paper reports results of artificial neural network for robot navigation tasks. Machine
learning methods have proven usability in many complex problems concerning mobile robots
control. In particular we deal with the well-known strategy of navigating by “wall-following”.
In this study, probabilistic neural network (PNN) structure was used for robot navigation tasks.
The PNN result was compared with the results of the Logistic Perceptron, Multilayer
Perceptron, Mixture of Experts and Elman neural networks and the results of the previous
studies reported focusing on robot navigation tasks and using same dataset. It was observed the
PNN is the best classification accuracy with 99,635% accuracy using same dataset.
LEARNING OF ROBOT NAVIGATION TASKS BY PROBABILISTIC NEURAL NETWORKcsandit
This paper reports results of artificial neural network for robot navigation tasks. Machine
learning methods have proven usability in many complex problems concerning mobile robots
control. In particular we deal with the well-known strategy of navigating by “wall-following”.
In this study, probabilistic neural network (PNN) structure was used for robot navigation tasks.
The PNN result was compared with the results of the Logistic Perceptron, Multilayer
Perceptron, Mixture of Experts and Elman neural networks and the results of the previous
studies reported focusing on robot navigation tasks and using same dataset. It was observed the
PNN is the best classification accuracy with 99,635% accuracy using same dataset.
A novel visual tracking scheme for unstructured indoor environmentsIJECEIAES
In the ever-expanding sphere of assistive robotics, the pressing need for advanced methods capable of accurately tracking individuals within unstructured indoor settings has been magnified. This research endeavours to devise a realtime visual tracking mechanism that encapsulates high performance attributes while maintaining minimal computational requirements. Inspired by the neural processes of the human brain’s visual information handling, our innovative algorithm employs a pattern image, serving as an ephemeral memory, which facilitates the identification of motion within images. This tracking paradigm was subjected to rigorous testing on a Nao humanoid robot, demonstrating noteworthy outcomes in controlled laboratory conditions. The algorithm exhibited a remarkably low false detection rate, less than 4%, and target losses were recorded in merely 12% of instances, thus attesting to its successful operation. Moreover, the algorithm’s capacity to accurately estimate the direct distance to the target further substantiated its high efficacy. These compelling findings serve as a substantial contribution to assistive robotics. The proficient visual tracking methodology proposed herein holds the potential to markedly amplify the competencies of robots operating in dynamic, unstructured indoor settings, and set the foundation for a higher degree of complex interactive tasks.
Non-contact, camera-based physiological measurements -
such as blood volume pulse and respiration rate - can now
be inferred by neural networks based on facial videos. This
technology has the potential to enable medical professionals
to make more informed telehealth decisions. Currently, this
software only runs on PCs, without a user interface. The neu-
ral network has a significant computational cost, making it
difficult to deploy on low-cost mobile devices. It also performs
poorly in varied environmental, sensor, personal, and contex-
tual conditions - such as darker skin tones. In this project,
we implement this neural network as an Android app that
runs in real-time; develop a more efficient architecture; evalu-
ate these architectures on older smartphones; and provide an
open-source, simple personalization pipeline to enable users to
calibrate the app. This all serves to make the technology more
democratic: making it available to as many users as possible,
while giving them the means to train and develop it further.
DYNAMIC AND REALTIME MODELLING OF UBIQUITOUS INTERACTIONcscpconf
Ubiquitous systems require user to be dynamically and realtime informed in order to make his current activity increasingly easy. First, this paper presents and discusses a method to model the realtime interaction of the user with a ubiquitous system based on Petri-nets modelling technology. The goal deals with investigating dynamically the appropriate form of interaction depending on the context of the user. Thus, the interaction model structure should be dynamically improved with respect to the current and particular activity or goal of the user to better cope with his runtime requirements. This mechanism has been characterized as “models mutation”. Secondly, this paper proves the dynamic construction of models while basing on the dynamic composition of services. The ultimate purpose is to take advantage of the ontology of service written in OWL-S in order to describe the dynamic aspect of Petri-nets based models, especially, the realtime and automatic composition of such models. Simulation work has been conducted to validate the proposed approach.
Visual victim detection and quadrotor-swarm coordination control in search an...IJECEIAES
We propose a distributed victim-detection algorithm through visual information on quadrotors using convolutional neuronal networks (CNN) in a search and rescue environment. Describing the navigation algorithm, which allows quadrotors to avoid collisions. Secondly, when one quadrotor detects a possible victim, it causes its closest neighbors to disconnect from the main swarm and form a new sub-swarm around the victim, which validates the victim’s status. Thus, a formation control that permits to acquire information is performed based on the well-known rendezvous consensus algorithm. Finally, images are processed using CNN identifying potential victims in the area. Given the uncertainty of the victim detection measurement among quadrotors’ cameras in the image processing, estimation consensus (EC) and max-estimation consensus (M-EC) algorithms are proposed focusing on agreeing over the victim detection estimation. We illustrate that M-EC delivers better results than EC in scenarios with poor visibility and uncertainty produced by fire and smoke. The algorithm proves that distributed fashion can obtain a more accurate result in decision-making on whether or not there is a victim, showing robustness under uncertainties and wrong measurements in comparison when a single quadrotor performs the mission. The well-functioning of the algorithm is evaluated by carrying out a simulation using V-Rep.
Identifier of human emotions based on convolutional neural network for assist...TELKOMNIKA JOURNAL
This paper proposes a solution for the problem of continuous prediction in real-time of the emotional state of a human user from the identification of characteristics in facial expressions. In robots whose main task is the care of people (children, sick or elderly people) is important to maintain a close relationship man-machine, anld a rapid response of the robot to the actions of the person under care. We propose to increase the level of intimacy of the robot, and its response to specific situations of the user, identifying in real time the emotion reflected by the person's face. This solution is integrated with algorithms of the research group related to the tracking of people for use on an assistant robot. The strategy used involves two stages of processing, the first involves the detection of faces using HOG and linear SVM, while the second identifies the emotion in the face using a CNN. The strategy was completely tested in the laboratory on our robotic platform, demonstrating high performance with low resource consumption. Through various controlled laboratory tests with different people, which forced a certain emotion on their faces, the scheme was able to identify the emotions with a success rate of 92%.
Analysis and assessment software for multi-user collaborative cognitive radi...IJECEIAES
Computer simulations are without a doubt a useful methodology that allows to explore research queries and develop prototypes at lower costs and timeframes than those required in hardware processes. The simulation tools used in cognitive radio networks (CRN) are undergoing an active process. Currently, there is no stable simulator that enables to characterize every element of the cognitive cycle and the available tools are a framework for discrete-event software. This work presents the spectral mobility simulator in CRN called “App MultiColl-DCRN”, developed with MATLAB’s app designer. In contrast with other frameworks, the simulator uses real spectral occupancy data and simultaneously analyzes features regarding spectral mobility, decision-making, multi-user access, collaborative scenarios and decentralized architectures. Performance metrics include bandwidth, throughput level, number of failed handoffs, number of total handoffs, number of handoffs with interference, number of anticipated handoffs and number of perfect handoffs. The assessment of the simulator involves three scenarios: the first and second scenarios present a collaborative structure using the multi-criteria optimization and compromise solution (VIKOR) decision-making model and the naïve Bayes prediction technique respectively. The third scenario presents a multi-user structure and uses simple additive weighting (SAW) as a decision-making technique. The present development represents a contribution in the cognitive radio network field since there is currently no software with the same features.
Robot operating system based autonomous navigation platform with human robot ...TELKOMNIKA JOURNAL
In emerging technologies, indoor service robots are playing a vital role for people who are physically challenged and visually impaired. The service robots are efficient and beneficial for people to overcome the challenges faced during their regular chores. This paper proposes the implementation of autonomous navigation platforms with human-robot interaction which can be used in service robots to avoid the difficulties faced in daily activities. We used the robot operating system (ROS) framework for the implementation of algorithms used in auto navigation, speech processing and recognition, and object detection and recognition. A suitable robot model was designed and tested in the Gazebo environment to evaluate the algorithms. The confusion matrix that was created from 125 different cases points to the decent correctness of the model.
Human-machine interactions based on hand gesture recognition using deep learn...IJECEIAES
Human interaction with computers and other machines is becoming an increasingly important and relevant topic in the modern world. Hand gesture recognition technology is an innovative approach to managing computers and electronic devices that allows users to interact with technology through gestures and hand movements. This article presents deep learning methods that allow you to efficiently process and classify hand gestures and hand gesture recognition technologies for interacting with computers. This paper discusses modern deep learning methods such as convolutional neural networks (CNN) and recurrent neural networks (RNN), which show excellent results in gesture recognition tasks. Next, the development and implementation of a human-machine interaction system based on hand gesture recognition is discussed. System architectures are described, as well as technical and practical aspects of their application. In conclusion, the article summarizes the research results and outlines the prospects for the development of hand gesture recognition technology to improve human- machine interaction. The advantages and limitations of the technology are analyzed, as well as possible areas of its application in the future.
Swarm robotics : Design and implementationIJECEIAES
This project presents a swarming and herding behaviour using simple robots. The main goal is to demonstrate the applicability of artificial intelligence (AI) in simple robotics that can then be scaled to industrial and consumer markets to further the ability of automation. AI can be achieved in many different ways; this paper explores the possible platforms on which to build a simple AI robots from consumer grade microcontrollers. Emphasis on simplicity is the main focus of this paper. Cheap and 8 bit microcontrollers were used as the brain of each robot in a decentralized swarm environment were each robot is autonomous but still a part of the whole. These simple robots don’t communicate directly with each other. They will utilize simple IR sensors to sense each other and simple limit switches to sense other obstacles in their environment. Their main objective is to assemble at certain location after initial start from random locations, and after converging they would move as a single unit without collisions. Using readily available microcontrollers and simple circuit design, semi-consistent swarming behaviour was achieved. These robots don’t follow a set path but will react dynamically to different scenarios, guided by their simple AI algorithm.
Taking advantage of state of the art underwater vehicles and current networking capabilities, the visionary
double objective of this work is to “open to people connected to the Internet, an access to ocean depths
anytime, anywhere.” Today, these people can just perceive the changing surface of the sea from the shores,
but ignore almost everything on what is hidden. If they could explore seabed and become knowledgeable,
they would get involved in finding alternative solutions for our vital terrestrial problems – pollution,
climate changes, destruction of biodiversity and exhaustion of Earth resources. The second objective is to
assist professionals of underwater world in performing their tasks by augmenting the perception of the
scene and offering automated actions such as wildlife monitoring and counting. The introduction of Mixed
Reality and Internet in aquatic activities constitutes a technological breakthrough when compared with the
status of existing related technologies. Through Internet, anyone, anywhere, at any moment will be
naturally able to dive in real-time using a Remote Operated Vehicle (ROV) in the most remarkable sites
around the world. The heart of this work is focused on Mixed Reality. The main challenge is to reach real
time display of digital video stream to web users, by mixing 3D entities (objects or pre-processed
underwater terrain surfaces), with 2D videos of live images collected in real time by a teleoperated ROV.
A review paper: optimal test cases for regression testing using artificial in...IJECEIAES
The goal of the testing process is to find errors and defects in the software being developed so that they can be fixed and corrected before they are delivered to the customer. Regression testing is an essential quality testing technique during the maintenance phase of the program as it is performed to ensure the integrity of the program after modifications have been made. With the development of the software, the test suite becomes too large to be fully implemented within the given test cost in terms of budget and time. Therefore, the cost of regression testing using different techniques should be reduced, here we dealt many methods such as retest all technique, regression test selection technique (RTS) and test case prioritization technique (TCP). The efficiency of these techniques is evaluated through the use of many metrics such as average percentage of fault detected (APFD), average percentage block coverage (APBC) and average percentage decision coverage (APDC). In this paper we dealt with these different techniques used in test case selection and test case prioritization and the metrics used to evaluate their efficiency by using different techniques of artificial intelligent and describe the best of all.
Proactive Intelligent Home System Using Contextual Information and Neural Net...IJERA Editor
Nowadays, cities around the world intend to use information technology to improve the lives of their citizens.
Future smart cities will incorporate digital data and technology to interact differently with their human
inhabitants.
Among the key component of a smart city, we find the smart home component. It is an autonomic environment
that can provide various smart services by considering the user’s context information. Several methods are used
in context-aware system to provide such services. In this paper, we propose an approach to offer the most
relevant services to the user according to any significant change of his context environment. The proposed
approach is based on the use of context history information together with user profiling and machine learning
techniques. Experimentations show that the proposed solution can efficiently provide the most useful services to
the user in an intelligent home environment.
Ambiences on the-fly usage of available resources through personal devicesijasuc
In smart spaces such as smart homes, computation is
embedded everywhere: in toys, appliances, or the
home’s infrastructure. Most of these devices provid
e a pool of available resources which the user can
take
advantage, interacting and creating a friendly envi
ronment. The inherent composability of these system
s
and other unique characteristics such as low-cost e
nergy, simplicity in module programming, and even
their small size, make them a suitable candidate fo
r dynamic and adaptive ambient systems. This resear
ch
work focuses on what is defined as an “ambience”, a
space with a user-defined set of computational
devices. A smart-home is modeled as a collection of
ambiences, where every ambience is capable of
providing a pool of available resources to the user
. In turn, the user is supposed to carry one or sev
eral
personal devices able to interact with the ambience
s, taking advantage of his inherent mobility. In th
is way,
the whole system can benefit from resources discove
red in the spatial proximity. A software architectu
re is
designed, which is based on the implementation of l
ow-cost algorithms able to detect and update the sy
stem
when changes in an ambience occur. Ambience middlew
are implementation works in a wide range of
architectures and OSs, while showing a negligible o
verhead in the time to perform the basic output
operations.
In recent times, the definition of internet of things has evolved to great extent and it means different things
to different people. However the basic idea of Internet of Things remains the same which is use of information and
communication Technologies(ICT’s) as well as Internet to provide efficient services to the mankind for their
optimum benefits. While implementing the Internet of Things itself is complex due to large of devices various link
layer technologies and services that are involved in such systems. In this paper, we specifically implement an urban
Internet of Things system which primarily gives an emphasis on smart educational institute which aims at
exploiting advanced communication technologies to support for administration of the institute and for the new
visitor who find it difficult to find their way around the institute. This paper thereby provides a comprehensive
survey of various technologies such as Android , Java, protocols and architecture for using Internet of Things in
Smart Institute. Furthermore ,this system will provide an overlook of the institute via the administration office
,playground and canteens. The user of the system will also get an static image map to roam around the institute.
Lastly the system will provide technical solutions and best practice guidelines for finding the way around the
institute and also help a new visitor in reaching his desired location.
Intelligent Robotics Navigation System: Problems, Methods, and Algorithm IJECEIAES
This paper set out to supplement new studies with a brief and comprehensible review of the advanced development in the area of the navigation system, starting from a single robot, multi-robot, and swarm robots from a particular perspective by taking insights from these biological systems. The inspiration is taken from nature by observing the human and the social animal that is believed to be very beneficial for this purpose. The intelligent navigation system is developed based on an individual characteristic or a social animal biological structure. The discussion of this paper will focus on how simple agent’s structure utilizes flexible and potential outcomes in order to navigate in a productive and unorganized surrounding. The combination of the navigation system and biologically inspired approach has attracted considerable attention, which makes it an important research area in the intelligent robotic system. Overall, this paper explores the implementation, which is resulted from the simulation performed by the embodiment of robots operating in real environments.
Bibliometric analysis highlighting the role of women in addressing climate ch...IJECEIAES
Fossil fuel consumption increased quickly, contributing to climate change
that is evident in unusual flooding and draughts, and global warming. Over
the past ten years, women's involvement in society has grown dramatically,
and they succeeded in playing a noticeable role in reducing climate change.
A bibliometric analysis of data from the last ten years has been carried out to
examine the role of women in addressing the climate change. The analysis's
findings discussed the relevant to the sustainable development goals (SDGs),
particularly SDG 7 and SDG 13. The results considered contributions made
by women in the various sectors while taking geographic dispersion into
account. The bibliometric analysis delves into topics including women's
leadership in environmental groups, their involvement in policymaking, their
contributions to sustainable development projects, and the influence of
gender diversity on attempts to mitigate climate change. This study's results
highlight how women have influenced policies and actions related to climate
change, point out areas of research deficiency and recommendations on how
to increase role of the women in addressing the climate change and
achieving sustainability. To achieve more successful results, this initiative
aims to highlight the significance of gender equality and encourage
inclusivity in climate change decision-making processes.
Voltage and frequency control of microgrid in presence of micro-turbine inter...IJECEIAES
The active and reactive load changes have a significant impact on voltage
and frequency. In this paper, in order to stabilize the microgrid (MG) against
load variations in islanding mode, the active and reactive power of all
distributed generators (DGs), including energy storage (battery), diesel
generator, and micro-turbine, are controlled. The micro-turbine generator is
connected to MG through a three-phase to three-phase matrix converter, and
the droop control method is applied for controlling the voltage and
frequency of MG. In addition, a method is introduced for voltage and
frequency control of micro-turbines in the transition state from gridconnected mode to islanding mode. A novel switching strategy of the matrix
converter is used for converting the high-frequency output voltage of the
micro-turbine to the grid-side frequency of the utility system. Moreover,
using the switching strategy, the low-order harmonics in the output current
and voltage are not produced, and consequently, the size of the output filter
would be reduced. In fact, the suggested control strategy is load-independent
and has no frequency conversion restrictions. The proposed approach for
voltage and frequency regulation demonstrates exceptional performance and
favorable response across various load alteration scenarios. The suggested
strategy is examined in several scenarios in the MG test systems, and the
simulation results are addressed.
More Related Content
Similar to Acoustic event characterization for service robot using convolutional networks
Visual victim detection and quadrotor-swarm coordination control in search an...IJECEIAES
We propose a distributed victim-detection algorithm through visual information on quadrotors using convolutional neuronal networks (CNN) in a search and rescue environment. Describing the navigation algorithm, which allows quadrotors to avoid collisions. Secondly, when one quadrotor detects a possible victim, it causes its closest neighbors to disconnect from the main swarm and form a new sub-swarm around the victim, which validates the victim’s status. Thus, a formation control that permits to acquire information is performed based on the well-known rendezvous consensus algorithm. Finally, images are processed using CNN identifying potential victims in the area. Given the uncertainty of the victim detection measurement among quadrotors’ cameras in the image processing, estimation consensus (EC) and max-estimation consensus (M-EC) algorithms are proposed focusing on agreeing over the victim detection estimation. We illustrate that M-EC delivers better results than EC in scenarios with poor visibility and uncertainty produced by fire and smoke. The algorithm proves that distributed fashion can obtain a more accurate result in decision-making on whether or not there is a victim, showing robustness under uncertainties and wrong measurements in comparison when a single quadrotor performs the mission. The well-functioning of the algorithm is evaluated by carrying out a simulation using V-Rep.
Identifier of human emotions based on convolutional neural network for assist...TELKOMNIKA JOURNAL
This paper proposes a solution for the problem of continuous prediction in real-time of the emotional state of a human user from the identification of characteristics in facial expressions. In robots whose main task is the care of people (children, sick or elderly people) is important to maintain a close relationship man-machine, anld a rapid response of the robot to the actions of the person under care. We propose to increase the level of intimacy of the robot, and its response to specific situations of the user, identifying in real time the emotion reflected by the person's face. This solution is integrated with algorithms of the research group related to the tracking of people for use on an assistant robot. The strategy used involves two stages of processing, the first involves the detection of faces using HOG and linear SVM, while the second identifies the emotion in the face using a CNN. The strategy was completely tested in the laboratory on our robotic platform, demonstrating high performance with low resource consumption. Through various controlled laboratory tests with different people, which forced a certain emotion on their faces, the scheme was able to identify the emotions with a success rate of 92%.
Analysis and assessment software for multi-user collaborative cognitive radi...IJECEIAES
Computer simulations are without a doubt a useful methodology that allows to explore research queries and develop prototypes at lower costs and timeframes than those required in hardware processes. The simulation tools used in cognitive radio networks (CRN) are undergoing an active process. Currently, there is no stable simulator that enables to characterize every element of the cognitive cycle and the available tools are a framework for discrete-event software. This work presents the spectral mobility simulator in CRN called “App MultiColl-DCRN”, developed with MATLAB’s app designer. In contrast with other frameworks, the simulator uses real spectral occupancy data and simultaneously analyzes features regarding spectral mobility, decision-making, multi-user access, collaborative scenarios and decentralized architectures. Performance metrics include bandwidth, throughput level, number of failed handoffs, number of total handoffs, number of handoffs with interference, number of anticipated handoffs and number of perfect handoffs. The assessment of the simulator involves three scenarios: the first and second scenarios present a collaborative structure using the multi-criteria optimization and compromise solution (VIKOR) decision-making model and the naïve Bayes prediction technique respectively. The third scenario presents a multi-user structure and uses simple additive weighting (SAW) as a decision-making technique. The present development represents a contribution in the cognitive radio network field since there is currently no software with the same features.
Robot operating system based autonomous navigation platform with human robot ...TELKOMNIKA JOURNAL
In emerging technologies, indoor service robots are playing a vital role for people who are physically challenged and visually impaired. The service robots are efficient and beneficial for people to overcome the challenges faced during their regular chores. This paper proposes the implementation of autonomous navigation platforms with human-robot interaction which can be used in service robots to avoid the difficulties faced in daily activities. We used the robot operating system (ROS) framework for the implementation of algorithms used in auto navigation, speech processing and recognition, and object detection and recognition. A suitable robot model was designed and tested in the Gazebo environment to evaluate the algorithms. The confusion matrix that was created from 125 different cases points to the decent correctness of the model.
Human-machine interactions based on hand gesture recognition using deep learn...IJECEIAES
Human interaction with computers and other machines is becoming an increasingly important and relevant topic in the modern world. Hand gesture recognition technology is an innovative approach to managing computers and electronic devices that allows users to interact with technology through gestures and hand movements. This article presents deep learning methods that allow you to efficiently process and classify hand gestures and hand gesture recognition technologies for interacting with computers. This paper discusses modern deep learning methods such as convolutional neural networks (CNN) and recurrent neural networks (RNN), which show excellent results in gesture recognition tasks. Next, the development and implementation of a human-machine interaction system based on hand gesture recognition is discussed. System architectures are described, as well as technical and practical aspects of their application. In conclusion, the article summarizes the research results and outlines the prospects for the development of hand gesture recognition technology to improve human- machine interaction. The advantages and limitations of the technology are analyzed, as well as possible areas of its application in the future.
Swarm robotics : Design and implementationIJECEIAES
This project presents a swarming and herding behaviour using simple robots. The main goal is to demonstrate the applicability of artificial intelligence (AI) in simple robotics that can then be scaled to industrial and consumer markets to further the ability of automation. AI can be achieved in many different ways; this paper explores the possible platforms on which to build a simple AI robots from consumer grade microcontrollers. Emphasis on simplicity is the main focus of this paper. Cheap and 8 bit microcontrollers were used as the brain of each robot in a decentralized swarm environment were each robot is autonomous but still a part of the whole. These simple robots don’t communicate directly with each other. They will utilize simple IR sensors to sense each other and simple limit switches to sense other obstacles in their environment. Their main objective is to assemble at certain location after initial start from random locations, and after converging they would move as a single unit without collisions. Using readily available microcontrollers and simple circuit design, semi-consistent swarming behaviour was achieved. These robots don’t follow a set path but will react dynamically to different scenarios, guided by their simple AI algorithm.
Taking advantage of state of the art underwater vehicles and current networking capabilities, the visionary
double objective of this work is to “open to people connected to the Internet, an access to ocean depths
anytime, anywhere.” Today, these people can just perceive the changing surface of the sea from the shores,
but ignore almost everything on what is hidden. If they could explore seabed and become knowledgeable,
they would get involved in finding alternative solutions for our vital terrestrial problems – pollution,
climate changes, destruction of biodiversity and exhaustion of Earth resources. The second objective is to
assist professionals of underwater world in performing their tasks by augmenting the perception of the
scene and offering automated actions such as wildlife monitoring and counting. The introduction of Mixed
Reality and Internet in aquatic activities constitutes a technological breakthrough when compared with the
status of existing related technologies. Through Internet, anyone, anywhere, at any moment will be
naturally able to dive in real-time using a Remote Operated Vehicle (ROV) in the most remarkable sites
around the world. The heart of this work is focused on Mixed Reality. The main challenge is to reach real
time display of digital video stream to web users, by mixing 3D entities (objects or pre-processed
underwater terrain surfaces), with 2D videos of live images collected in real time by a teleoperated ROV.
A review paper: optimal test cases for regression testing using artificial in...IJECEIAES
The goal of the testing process is to find errors and defects in the software being developed so that they can be fixed and corrected before they are delivered to the customer. Regression testing is an essential quality testing technique during the maintenance phase of the program as it is performed to ensure the integrity of the program after modifications have been made. With the development of the software, the test suite becomes too large to be fully implemented within the given test cost in terms of budget and time. Therefore, the cost of regression testing using different techniques should be reduced, here we dealt many methods such as retest all technique, regression test selection technique (RTS) and test case prioritization technique (TCP). The efficiency of these techniques is evaluated through the use of many metrics such as average percentage of fault detected (APFD), average percentage block coverage (APBC) and average percentage decision coverage (APDC). In this paper we dealt with these different techniques used in test case selection and test case prioritization and the metrics used to evaluate their efficiency by using different techniques of artificial intelligent and describe the best of all.
Proactive Intelligent Home System Using Contextual Information and Neural Net...IJERA Editor
Nowadays, cities around the world intend to use information technology to improve the lives of their citizens.
Future smart cities will incorporate digital data and technology to interact differently with their human
inhabitants.
Among the key component of a smart city, we find the smart home component. It is an autonomic environment
that can provide various smart services by considering the user’s context information. Several methods are used
in context-aware system to provide such services. In this paper, we propose an approach to offer the most
relevant services to the user according to any significant change of his context environment. The proposed
approach is based on the use of context history information together with user profiling and machine learning
techniques. Experimentations show that the proposed solution can efficiently provide the most useful services to
the user in an intelligent home environment.
Ambiences on the-fly usage of available resources through personal devicesijasuc
In smart spaces such as smart homes, computation is
embedded everywhere: in toys, appliances, or the
home’s infrastructure. Most of these devices provid
e a pool of available resources which the user can
take
advantage, interacting and creating a friendly envi
ronment. The inherent composability of these system
s
and other unique characteristics such as low-cost e
nergy, simplicity in module programming, and even
their small size, make them a suitable candidate fo
r dynamic and adaptive ambient systems. This resear
ch
work focuses on what is defined as an “ambience”, a
space with a user-defined set of computational
devices. A smart-home is modeled as a collection of
ambiences, where every ambience is capable of
providing a pool of available resources to the user
. In turn, the user is supposed to carry one or sev
eral
personal devices able to interact with the ambience
s, taking advantage of his inherent mobility. In th
is way,
the whole system can benefit from resources discove
red in the spatial proximity. A software architectu
re is
designed, which is based on the implementation of l
ow-cost algorithms able to detect and update the sy
stem
when changes in an ambience occur. Ambience middlew
are implementation works in a wide range of
architectures and OSs, while showing a negligible o
verhead in the time to perform the basic output
operations.
In recent times, the definition of internet of things has evolved to great extent and it means different things
to different people. However the basic idea of Internet of Things remains the same which is use of information and
communication Technologies(ICT’s) as well as Internet to provide efficient services to the mankind for their
optimum benefits. While implementing the Internet of Things itself is complex due to large of devices various link
layer technologies and services that are involved in such systems. In this paper, we specifically implement an urban
Internet of Things system which primarily gives an emphasis on smart educational institute which aims at
exploiting advanced communication technologies to support for administration of the institute and for the new
visitor who find it difficult to find their way around the institute. This paper thereby provides a comprehensive
survey of various technologies such as Android , Java, protocols and architecture for using Internet of Things in
Smart Institute. Furthermore ,this system will provide an overlook of the institute via the administration office
,playground and canteens. The user of the system will also get an static image map to roam around the institute.
Lastly the system will provide technical solutions and best practice guidelines for finding the way around the
institute and also help a new visitor in reaching his desired location.
Intelligent Robotics Navigation System: Problems, Methods, and Algorithm IJECEIAES
This paper set out to supplement new studies with a brief and comprehensible review of the advanced development in the area of the navigation system, starting from a single robot, multi-robot, and swarm robots from a particular perspective by taking insights from these biological systems. The inspiration is taken from nature by observing the human and the social animal that is believed to be very beneficial for this purpose. The intelligent navigation system is developed based on an individual characteristic or a social animal biological structure. The discussion of this paper will focus on how simple agent’s structure utilizes flexible and potential outcomes in order to navigate in a productive and unorganized surrounding. The combination of the navigation system and biologically inspired approach has attracted considerable attention, which makes it an important research area in the intelligent robotic system. Overall, this paper explores the implementation, which is resulted from the simulation performed by the embodiment of robots operating in real environments.
Bibliometric analysis highlighting the role of women in addressing climate ch...IJECEIAES
Fossil fuel consumption increased quickly, contributing to climate change
that is evident in unusual flooding and draughts, and global warming. Over
the past ten years, women's involvement in society has grown dramatically,
and they succeeded in playing a noticeable role in reducing climate change.
A bibliometric analysis of data from the last ten years has been carried out to
examine the role of women in addressing the climate change. The analysis's
findings discussed the relevant to the sustainable development goals (SDGs),
particularly SDG 7 and SDG 13. The results considered contributions made
by women in the various sectors while taking geographic dispersion into
account. The bibliometric analysis delves into topics including women's
leadership in environmental groups, their involvement in policymaking, their
contributions to sustainable development projects, and the influence of
gender diversity on attempts to mitigate climate change. This study's results
highlight how women have influenced policies and actions related to climate
change, point out areas of research deficiency and recommendations on how
to increase role of the women in addressing the climate change and
achieving sustainability. To achieve more successful results, this initiative
aims to highlight the significance of gender equality and encourage
inclusivity in climate change decision-making processes.
Voltage and frequency control of microgrid in presence of micro-turbine inter...IJECEIAES
The active and reactive load changes have a significant impact on voltage
and frequency. In this paper, in order to stabilize the microgrid (MG) against
load variations in islanding mode, the active and reactive power of all
distributed generators (DGs), including energy storage (battery), diesel
generator, and micro-turbine, are controlled. The micro-turbine generator is
connected to MG through a three-phase to three-phase matrix converter, and
the droop control method is applied for controlling the voltage and
frequency of MG. In addition, a method is introduced for voltage and
frequency control of micro-turbines in the transition state from gridconnected mode to islanding mode. A novel switching strategy of the matrix
converter is used for converting the high-frequency output voltage of the
micro-turbine to the grid-side frequency of the utility system. Moreover,
using the switching strategy, the low-order harmonics in the output current
and voltage are not produced, and consequently, the size of the output filter
would be reduced. In fact, the suggested control strategy is load-independent
and has no frequency conversion restrictions. The proposed approach for
voltage and frequency regulation demonstrates exceptional performance and
favorable response across various load alteration scenarios. The suggested
strategy is examined in several scenarios in the MG test systems, and the
simulation results are addressed.
Enhancing battery system identification: nonlinear autoregressive modeling fo...IJECEIAES
Precisely characterizing Li-ion batteries is essential for optimizing their
performance, enhancing safety, and prolonging their lifespan across various
applications, such as electric vehicles and renewable energy systems. This
article introduces an innovative nonlinear methodology for system
identification of a Li-ion battery, employing a nonlinear autoregressive with
exogenous inputs (NARX) model. The proposed approach integrates the
benefits of nonlinear modeling with the adaptability of the NARX structure,
facilitating a more comprehensive representation of the intricate
electrochemical processes within the battery. Experimental data collected
from a Li-ion battery operating under diverse scenarios are employed to
validate the effectiveness of the proposed methodology. The identified
NARX model exhibits superior accuracy in predicting the battery's behavior
compared to traditional linear models. This study underscores the
importance of accounting for nonlinearities in battery modeling, providing
insights into the intricate relationships between state-of-charge, voltage, and
current under dynamic conditions.
Smart grid deployment: from a bibliometric analysis to a surveyIJECEIAES
Smart grids are one of the last decades' innovations in electrical energy.
They bring relevant advantages compared to the traditional grid and
significant interest from the research community. Assessing the field's
evolution is essential to propose guidelines for facing new and future smart
grid challenges. In addition, knowing the main technologies involved in the
deployment of smart grids (SGs) is important to highlight possible
shortcomings that can be mitigated by developing new tools. This paper
contributes to the research trends mentioned above by focusing on two
objectives. First, a bibliometric analysis is presented to give an overview of
the current research level about smart grid deployment. Second, a survey of
the main technological approaches used for smart grid implementation and
their contributions are highlighted. To that effect, we searched the Web of
Science (WoS), and the Scopus databases. We obtained 5,663 documents
from WoS and 7,215 from Scopus on smart grid implementation or
deployment. With the extraction limitation in the Scopus database, 5,872 of
the 7,215 documents were extracted using a multi-step process. These two
datasets have been analyzed using a bibliometric tool called bibliometrix.
The main outputs are presented with some recommendations for future
research.
Use of analytical hierarchy process for selecting and prioritizing islanding ...IJECEIAES
One of the problems that are associated to power systems is islanding
condition, which must be rapidly and properly detected to prevent any
negative consequences on the system's protection, stability, and security.
This paper offers a thorough overview of several islanding detection
strategies, which are divided into two categories: classic approaches,
including local and remote approaches, and modern techniques, including
techniques based on signal processing and computational intelligence.
Additionally, each approach is compared and assessed based on several
factors, including implementation costs, non-detected zones, declining
power quality, and response times using the analytical hierarchy process
(AHP). The multi-criteria decision-making analysis shows that the overall
weight of passive methods (24.7%), active methods (7.8%), hybrid methods
(5.6%), remote methods (14.5%), signal processing-based methods (26.6%),
and computational intelligent-based methods (20.8%) based on the
comparison of all criteria together. Thus, it can be seen from the total weight
that hybrid approaches are the least suitable to be chosen, while signal
processing-based methods are the most appropriate islanding detection
method to be selected and implemented in power system with respect to the
aforementioned factors. Using Expert Choice software, the proposed
hierarchy model is studied and examined.
Enhancing of single-stage grid-connected photovoltaic system using fuzzy logi...IJECEIAES
The power generated by photovoltaic (PV) systems is influenced by
environmental factors. This variability hampers the control and utilization of
solar cells' peak output. In this study, a single-stage grid-connected PV
system is designed to enhance power quality. Our approach employs fuzzy
logic in the direct power control (DPC) of a three-phase voltage source
inverter (VSI), enabling seamless integration of the PV connected to the
grid. Additionally, a fuzzy logic-based maximum power point tracking
(MPPT) controller is adopted, which outperforms traditional methods like
incremental conductance (INC) in enhancing solar cell efficiency and
minimizing the response time. Moreover, the inverter's real-time active and
reactive power is directly managed to achieve a unity power factor (UPF).
The system's performance is assessed through MATLAB/Simulink
implementation, showing marked improvement over conventional methods,
particularly in steady-state and varying weather conditions. For solar
irradiances of 500 and 1,000 W/m2
, the results show that the proposed
method reduces the total harmonic distortion (THD) of the injected current
to the grid by approximately 46% and 38% compared to conventional
methods, respectively. Furthermore, we compare the simulation results with
IEEE standards to evaluate the system's grid compatibility.
Enhancing photovoltaic system maximum power point tracking with fuzzy logic-b...IJECEIAES
Photovoltaic systems have emerged as a promising energy resource that
caters to the future needs of society, owing to their renewable, inexhaustible,
and cost-free nature. The power output of these systems relies on solar cell
radiation and temperature. In order to mitigate the dependence on
atmospheric conditions and enhance power tracking, a conventional
approach has been improved by integrating various methods. To optimize
the generation of electricity from solar systems, the maximum power point
tracking (MPPT) technique is employed. To overcome limitations such as
steady-state voltage oscillations and improve transient response, two
traditional MPPT methods, namely fuzzy logic controller (FLC) and perturb
and observe (P&O), have been modified. This research paper aims to
simulate and validate the step size of the proposed modified P&O and FLC
techniques within the MPPT algorithm using MATLAB/Simulink for
efficient power tracking in photovoltaic systems.
Adaptive synchronous sliding control for a robot manipulator based on neural ...IJECEIAES
Robot manipulators have become important equipment in production lines, medical fields, and transportation. Improving the quality of trajectory tracking for
robot hands is always an attractive topic in the research community. This is a
challenging problem because robot manipulators are complex nonlinear systems
and are often subject to fluctuations in loads and external disturbances. This
article proposes an adaptive synchronous sliding control scheme to improve trajectory tracking performance for a robot manipulator. The proposed controller
ensures that the positions of the joints track the desired trajectory, synchronize
the errors, and significantly reduces chattering. First, the synchronous tracking
errors and synchronous sliding surfaces are presented. Second, the synchronous
tracking error dynamics are determined. Third, a robust adaptive control law is
designed,the unknown components of the model are estimated online by the neural network, and the parameters of the switching elements are selected by fuzzy
logic. The built algorithm ensures that the tracking and approximation errors
are ultimately uniformly bounded (UUB). Finally, the effectiveness of the constructed algorithm is demonstrated through simulation and experimental results.
Simulation and experimental results show that the proposed controller is effective with small synchronous tracking errors, and the chattering phenomenon is
significantly reduced.
Remote field-programmable gate array laboratory for signal acquisition and de...IJECEIAES
A remote laboratory utilizing field-programmable gate array (FPGA) technologies enhances students’ learning experience anywhere and anytime in embedded system design. Existing remote laboratories prioritize hardware access and visual feedback for observing board behavior after programming, neglecting comprehensive debugging tools to resolve errors that require internal signal acquisition. This paper proposes a novel remote embeddedsystem design approach targeting FPGA technologies that are fully interactive via a web-based platform. Our solution provides FPGA board access and debugging capabilities beyond the visual feedback provided by existing remote laboratories. We implemented a lab module that allows users to seamlessly incorporate into their FPGA design. The module minimizes hardware resource utilization while enabling the acquisition of a large number of data samples from the signal during the experiments by adaptively compressing the signal prior to data transmission. The results demonstrate an average compression ratio of 2.90 across three benchmark signals, indicating efficient signal acquisition and effective debugging and analysis. This method allows users to acquire more data samples than conventional methods. The proposed lab allows students to remotely test and debug their designs, bridging the gap between theory and practice in embedded system design.
Detecting and resolving feature envy through automated machine learning and m...IJECEIAES
Efficiently identifying and resolving code smells enhances software project quality. This paper presents a novel solution, utilizing automated machine learning (AutoML) techniques, to detect code smells and apply move method refactoring. By evaluating code metrics before and after refactoring, we assessed its impact on coupling, complexity, and cohesion. Key contributions of this research include a unique dataset for code smell classification and the development of models using AutoGluon for optimal performance. Furthermore, the study identifies the top 20 influential features in classifying feature envy, a well-known code smell, stemming from excessive reliance on external classes. We also explored how move method refactoring addresses feature envy, revealing reduced coupling and complexity, and improved cohesion, ultimately enhancing code quality. In summary, this research offers an empirical, data-driven approach, integrating AutoML and move method refactoring to optimize software project quality. Insights gained shed light on the benefits of refactoring on code quality and the significance of specific features in detecting feature envy. Future research can expand to explore additional refactoring techniques and a broader range of code metrics, advancing software engineering practices and standards.
Smart monitoring technique for solar cell systems using internet of things ba...IJECEIAES
Rapidly and remotely monitoring and receiving the solar cell systems status parameters, solar irradiance, temperature, and humidity, are critical issues in enhancement their efficiency. Hence, in the present article an improved smart prototype of internet of things (IoT) technique based on embedded system through NodeMCU ESP8266 (ESP-12E) was carried out experimentally. Three different regions at Egypt; Luxor, Cairo, and El-Beheira cities were chosen to study their solar irradiance profile, temperature, and humidity by the proposed IoT system. The monitoring data of solar irradiance, temperature, and humidity were live visualized directly by Ubidots through hypertext transfer protocol (HTTP) protocol. The measured solar power radiation in Luxor, Cairo, and El-Beheira ranged between 216-1000, 245-958, and 187-692 W/m 2 respectively during the solar day. The accuracy and rapidity of obtaining monitoring results using the proposed IoT system made it a strong candidate for application in monitoring solar cell systems. On the other hand, the obtained solar power radiation results of the three considered regions strongly candidate Luxor and Cairo as suitable places to build up a solar cells system station rather than El-Beheira.
An efficient security framework for intrusion detection and prevention in int...IJECEIAES
Over the past few years, the internet of things (IoT) has advanced to connect billions of smart devices to improve quality of life. However, anomalies or malicious intrusions pose several security loopholes, leading to performance degradation and threat to data security in IoT operations. Thereby, IoT security systems must keep an eye on and restrict unwanted events from occurring in the IoT network. Recently, various technical solutions based on machine learning (ML) models have been derived towards identifying and restricting unwanted events in IoT. However, most ML-based approaches are prone to miss-classification due to inappropriate feature selection. Additionally, most ML approaches applied to intrusion detection and prevention consider supervised learning, which requires a large amount of labeled data to be trained. Consequently, such complex datasets are impossible to source in a large network like IoT. To address this problem, this proposed study introduces an efficient learning mechanism to strengthen the IoT security aspects. The proposed algorithm incorporates supervised and unsupervised approaches to improve the learning models for intrusion detection and mitigation. Compared with the related works, the experimental outcome shows that the model performs well in a benchmark dataset. It accomplishes an improved detection accuracy of approximately 99.21%.
Developing a smart system for infant incubators using the internet of things ...IJECEIAES
This research is developing an incubator system that integrates the internet of things and artificial intelligence to improve care for premature babies. The system workflow starts with sensors that collect data from the incubator. Then, the data is sent in real-time to the internet of things (IoT) broker eclipse mosquito using the message queue telemetry transport (MQTT) protocol version 5.0. After that, the data is stored in a database for analysis using the long short-term memory network (LSTM) method and displayed in a web application using an application programming interface (API) service. Furthermore, the experimental results produce as many as 2,880 rows of data stored in the database. The correlation coefficient between the target attribute and other attributes ranges from 0.23 to 0.48. Next, several experiments were conducted to evaluate the model-predicted value on the test data. The best results are obtained using a two-layer LSTM configuration model, each with 60 neurons and a lookback setting 6. This model produces an R 2 value of 0.934, with a root mean square error (RMSE) value of 0.015 and a mean absolute error (MAE) of 0.008. In addition, the R 2 value was also evaluated for each attribute used as input, with a result of values between 0.590 and 0.845.
A review on internet of things-based stingless bee's honey production with im...IJECEIAES
Honey is produced exclusively by honeybees and stingless bees which both are well adapted to tropical and subtropical regions such as Malaysia. Stingless bees are known for producing small amounts of honey and are known for having a unique flavor profile. Problem identified that many stingless bees collapsed due to weather, temperature and environment. It is critical to understand the relationship between the production of stingless bee honey and environmental conditions to improve honey production. Thus, this paper presents a review on stingless bee's honey production and prediction modeling. About 54 previous research has been analyzed and compared in identifying the research gaps. A framework on modeling the prediction of stingless bee honey is derived. The result presents the comparison and analysis on the internet of things (IoT) monitoring systems, honey production estimation, convolution neural networks (CNNs), and automatic identification methods on bee species. It is identified based on image detection method the top best three efficiency presents CNN is at 98.67%, densely connected convolutional networks with YOLO v3 is 97.7%, and DenseNet201 convolutional networks 99.81%. This study is significant to assist the researcher in developing a model for predicting stingless honey produced by bee's output, which is important for a stable economy and food security.
A trust based secure access control using authentication mechanism for intero...IJECEIAES
The internet of things (IoT) is a revolutionary innovation in many aspects of our society including interactions, financial activity, and global security such as the military and battlefield internet. Due to the limited energy and processing capacity of network devices, security, energy consumption, compatibility, and device heterogeneity are the long-term IoT problems. As a result, energy and security are critical for data transmission across edge and IoT networks. Existing IoT interoperability techniques need more computation time, have unreliable authentication mechanisms that break easily, lose data easily, and have low confidentiality. In this paper, a key agreement protocol-based authentication mechanism for IoT devices is offered as a solution to this issue. This system makes use of information exchange, which must be secured to prevent access by unauthorized users. Using a compact contiki/cooja simulator, the performance and design of the suggested framework are validated. The simulation findings are evaluated based on detection of malicious nodes after 60 minutes of simulation. The suggested trust method, which is based on privacy access control, reduced packet loss ratio to 0.32%, consumed 0.39% power, and had the greatest average residual energy of 0.99 mJoules at 10 nodes.
Fuzzy linear programming with the intuitionistic polygonal fuzzy numbersIJECEIAES
In real world applications, data are subject to ambiguity due to several factors; fuzzy sets and fuzzy numbers propose a great tool to model such ambiguity. In case of hesitation, the complement of a membership value in fuzzy numbers can be different from the non-membership value, in which case we can model using intuitionistic fuzzy numbers as they provide flexibility by defining both a membership and a non-membership functions. In this article, we consider the intuitionistic fuzzy linear programming problem with intuitionistic polygonal fuzzy numbers, which is a generalization of the previous polygonal fuzzy numbers found in the literature. We present a modification of the simplex method that can be used to solve any general intuitionistic fuzzy linear programming problem after approximating the problem by an intuitionistic polygonal fuzzy number with n edges. This method is given in a simple tableau formulation, and then applied on numerical examples for clarity.
The performance of artificial intelligence in prostate magnetic resonance im...IJECEIAES
Prostate cancer is the predominant form of cancer observed in men worldwide. The application of magnetic resonance imaging (MRI) as a guidance tool for conducting biopsies has been established as a reliable and well-established approach in the diagnosis of prostate cancer. The diagnostic performance of MRI-guided prostate cancer diagnosis exhibits significant heterogeneity due to the intricate and multi-step nature of the diagnostic pathway. The development of artificial intelligence (AI) models, specifically through the utilization of machine learning techniques such as deep learning, is assuming an increasingly significant role in the field of radiology. In the realm of prostate MRI, a considerable body of literature has been dedicated to the development of various AI algorithms. These algorithms have been specifically designed for tasks such as prostate segmentation, lesion identification, and classification. The overarching objective of these endeavors is to enhance diagnostic performance and foster greater agreement among different observers within MRI scans for the prostate. This review article aims to provide a concise overview of the application of AI in the field of radiology, with a specific focus on its utilization in prostate MRI.
Seizure stage detection of epileptic seizure using convolutional neural networksIJECEIAES
According to the World Health Organization (WHO), seventy million individuals worldwide suffer from epilepsy, a neurological disorder. While electroencephalography (EEG) is crucial for diagnosing epilepsy and monitoring the brain activity of epilepsy patients, it requires a specialist to examine all EEG recordings to find epileptic behavior. This procedure needs an experienced doctor, and a precise epilepsy diagnosis is crucial for appropriate treatment. To identify epileptic seizures, this study employed a convolutional neural network (CNN) based on raw scalp EEG signals to discriminate between preictal, ictal, postictal, and interictal segments. The possibility of these characteristics is explored by examining how well timedomain signals work in the detection of epileptic signals using intracranial Freiburg Hospital (FH), scalp Children's Hospital Boston-Massachusetts Institute of Technology (CHB-MIT) databases, and Temple University Hospital (TUH) EEG. To test the viability of this approach, two types of experiments were carried out. Firstly, binary class classification (preictal, ictal, postictal each versus interictal) and four-class classification (interictal versus preictal versus ictal versus postictal). The average accuracy for stage detection using CHB-MIT database was 84.4%, while the Freiburg database's time-domain signals had an accuracy of 79.7% and the highest accuracy of 94.02% for classification in the TUH EEG database when comparing interictal stage to preictal stage.
Analysis of driving style using self-organizing maps to analyze driver behaviorIJECEIAES
Modern life is strongly associated with the use of cars, but the increase in acceleration speeds and their maneuverability leads to a dangerous driving style for some drivers. In these conditions, the development of a method that allows you to track the behavior of the driver is relevant. The article provides an overview of existing methods and models for assessing the functioning of motor vehicles and driver behavior. Based on this, a combined algorithm for recognizing driving style is proposed. To do this, a set of input data was formed, including 20 descriptive features: About the environment, the driver's behavior and the characteristics of the functioning of the car, collected using OBD II. The generated data set is sent to the Kohonen network, where clustering is performed according to driving style and degree of danger. Getting the driving characteristics into a particular cluster allows you to switch to the private indicators of an individual driver and considering individual driving characteristics. The application of the method allows you to identify potentially dangerous driving styles that can prevent accidents.
Hyperspectral object classification using hybrid spectral-spatial fusion and ...IJECEIAES
Because of its spectral-spatial and temporal resolution of greater areas, hyperspectral imaging (HSI) has found widespread application in the field of object classification. The HSI is typically used to accurately determine an object's physical characteristics as well as to locate related objects with appropriate spectral fingerprints. As a result, the HSI has been extensively applied to object identification in several fields, including surveillance, agricultural monitoring, environmental research, and precision agriculture. However, because of their enormous size, objects require a lot of time to classify; for this reason, both spectral and spatial feature fusion have been completed. The existing classification strategy leads to increased misclassification, and the feature fusion method is unable to preserve semantic object inherent features; This study addresses the research difficulties by introducing a hybrid spectral-spatial fusion (HSSF) technique to minimize feature size while maintaining object intrinsic qualities; Lastly, a soft-margins kernel is proposed for multi-layer deep support vector machine (MLDSVM) to reduce misclassification. The standard Indian pines dataset is used for the experiment, and the outcome demonstrates that the HSSF-MLDSVM model performs substantially better in terms of accuracy and Kappa coefficient.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Understanding Inductive Bias in Machine LearningSUTEJAS
This presentation explores the concept of inductive bias in machine learning. It explains how algorithms come with built-in assumptions and preferences that guide the learning process. You'll learn about the different types of inductive bias and how they can impact the performance and generalizability of machine learning models.
The presentation also covers the positive and negative aspects of inductive bias, along with strategies for mitigating potential drawbacks. We'll explore examples of how bias manifests in algorithms like neural networks and decision trees.
By understanding inductive bias, you can gain valuable insights into how machine learning models work and make informed decisions when building and deploying them.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
Acoustic event characterization for service robot using convolutional networks
1. International Journal of Electrical and Computer Engineering (IJECE)
Vol. 12, No. 6, December 2022, pp. 6684∼6696
ISSN: 2088-8708, DOI: 10.11591/ijece.v12i6.pp6684-6696 ❒ 6684
Acoustic event characterization for service robot using
convolutional networks
Fernando Martı́nez, Fredy Martı́nez, César Hernández
Facultad Tecnológica, Universidad Distrital Francisco José de Caldas, Bogotá, Colombia
Article Info
Article history:
Received Oct 25, 2021
Revised May 25, 2022
Accepted Jun 21, 2022
Keywords:
Acoustic event
Convolutional neural network
Human-machine interaction
Image categorization
Learning process
ABSTRACT
This paper presents and discusses the creation of a sound event classification model
using deep learning. In the design of service robots, it is necessary to include routines
that improve the response of both the robot and the human being throughout the inter-
action. These types of tasks are critical when the robot is taking care of children, the
elderly, or people in vulnerable situations. Certain dangerous situations are difficult
to identify and assess by an autonomous system, and yet, the life of the users may
depend on these robots. Acoustic signals correspond to events that can be detected
at a great distance, are usually present in risky situations, and can be continuously
sensed without incurring privacy risks. For the creation of the model, a customized
database is structured with seven categories that allow to categorize a problem, and
eventually allow the robot to provide the necessary help. These audio signals are
processed to produce graphical representations consistent with human acoustic identi-
fication. These images are then used to train three convolutional models identified as
high-performing in this type of problem. The three models are evaluated with specific
metrics to identify the best-performing model. Finally, the results of this evaluation
are discussed and analyzed.
This is an open access article under the CC BY-SA license.
Corresponding Author:
Fredy Martı́nez
Facultad Tecnológica, Universidad Distrital Francisco José de Caldas
Carrera 7 No. 40B-53, Bogotá, Colombia
Email: fhmartinezs@udistrital.edu.co
1. INTRODUCTION
The technological development of today has made life easier for human beings, particularly for those
individuals with limitations and/or special needs. Among the first technologies designed by men to improve
their living conditions were prescription lenses and glasses. Thanks to them, a person with reduced visual
capacity was able to interact socially without restrictions. Today this technology has evolved to incorporate
ultralight and intelligent materials that adapt not only to the user but also to the conditions of the environment.
This is an example of current smart hardware, but hand in hand with this hardware has also evolved software
tools, which in the case of glasses allow not only the design but also to estimate the future performance of
the lenses and even their acceptance among users. These developments are strongly marked by their ability to
integrate with the human being, they must not only be able to solve a problem but they must also be accepted
by people [1].
Here is where service robotics research has found a broad niche with significant research problems to
solve. Today’s society has new challenges that demand the provision of specialized services, such as the care
of people (elderly and children), or special training processes (children in their homes, or adults in industrial
Journal homepage: http://ijece.iaescore.com
2. Int J Elec & Comp Eng ISSN: 2088-8708 ❒ 6685
environments) [2]. Such services require an artificial system that beyond providing content, tracks the pro-
cesses according to the particular behavior of the individual [3]. There is a lot of research on human-machine
integration systems, in some cases trying to identify emotional states from people’s faces, but this involves
processing the image in familiar environments including children, which for many is a privacy risk. A less
intrusive strategy is based on the same principle of estimating emotions and special events but from acoustic
signals, which is the area of interest of this research.
However, the question is whether it is possible for a small robot to autonomously identify acoustic
events, particularly using convolutional networks. The identification of sound events by an autonomous robot
is a desirable feature because it can improve integration in task execution, particularly in human-machine
interaction [4]–[6]. The level of interaction, and thus task success, is strongly related to the robot’s ability to
anticipate human attitudes, which is particularly true for robots that provide care to people [7], [8]. Examples
of acoustic events may include screaming people, crying babies, or an audible alarm. However, these events
are characterized by complex parameters, which, while recognizable to a person, are not recognizable to a
computerized system [9], [10]. Two cries of people turn out to be completely different, parameters such as
volume, length, frequency, or acoustic power are not enough to characterize them, moreover, the robot should
identify the group of any person, which infers the need for a learning process with a large dataset [11], [12].
Although the characterization of these complex events is complex, convolutional networks have demonstrated
great success in the categorization of other problems with similar complexity [13], [14].
Additionally, convolutional models can be trained for specific needs. The databases, and the categories
defined within them, can be adjusted to the needs of the system [15]. For example, it is possible to identify
whether a set of images belong to a jungle landscape or a beach, but the same images can also be used to
determine whether there are people or animals in them, or whether there are vehicles or houses in them [16].
A dataset can contain a lot of information, and it is up to the use and designer of the classifier to define which
features he wants to identify and use in his model. In addition, while most of the known applications focus
on image categorization, in many applications the signals of a given system have been modified into a visual
representation, which has allowed its use in multiple situations with various input parameters [17], [18].
Finally, it is possible to autonomously identify acoustic events by a small robot using convolutional
neural networks (CNNs) because convolutional models can run in real-time after they have been trained cor-
rectly [19], [20]. Service robots must operate in human environments, interacting with humans, so they must
be able to respond in real-time to immediate needs [21]–[23]. If a service robot detects a person’s scream,
it must be able to autonomously give and request help at the same instant it detects the event [24]. Even so,
robots include a large number of parallel systems that allow them not only to interact but also to guarantee their
functionality and the safety of the users [25]. Therefore, the performance demands, and therefore on hardware,
are high. Each of these systems, including the sound event identification event, must be capable of running on
the robot hardware, considering the possibility of restricted communication in cases of emergency [26]. CNNs
can create models that can be deployed on small embedded systems without excessive resource consumption
[27], [28]. The current frameworks have compatibility for their implementation on different systems, which
also facilitates their continuous updating. Thus, while neural networks are black boxes, and their performance
is strongly dependent on their training, it is possible to autonomously identify acoustic events by a small robot
using convolutional neural networks. Beyond the existing acoustic event detection (AED) systems, our appli-
cation requires in addition to high performance the ability to operate in real-time, and the possibility of running
continuously as a parallel task in a small service robot. This is precisely the gap in the current research that
we seek to solve, since the reported solutions, to the best of our knowledge, have not reached a point of devel-
opment that allows their massive use on commercial hardware (embedded processors for these robots, audio
digitization systems, and cloud processing platforms).
The structure of this paper is as follows: section 2 provides the general details of the robotic develop-
ment platform, its function, and the objective of the acoustic event characterization scheme on it. It discusses
the desired characteristics of such a scheme, and how it should be integrated with the current robot schemes.
Section 3 provides the design features of the convolutional models, all the tasks developed for their imple-
mentation and evaluation, and the corresponding analysis in each case. All the characteristics of each test are
detailed, as well as the information required for its duplication. Section 4 focuses on the results of each model
concerning the performance metrics applied on them with an unknown data set. Section 5 provides the limi-
tation of this research. Finally, section 6 presents the interpretation of the performance results of each model,
and conclusions.
Acoustic event characterization for service robot using convolutional networks (Fernando Martı́nez)
3. 6686 ❒ ISSN: 2088-8708
2. PROBLEM STATEMENT AND STATE OF THE ART
The objective of this research is to evaluate whether a model based on deep networks can correctly
identify acoustic events that may indicate high risk to users. This model is intended to be implemented on a
service robot designed for home environments, performing tasks such as the care and surveillance of children,
the elderly, or vulnerable people. Therefore, the aim is to establish the real performance of convolutional
models trained to classify acoustic events from a specific database. If the robot can identify, for example,
screams or gunshots, it is expected to be able to communicate autonomously to request immediate help. This
idea can be extended to the care of sick people, or people with disorders that require special monitoring, or
even for assistance in disaster situations.
By definition, an acoustic event corresponds to an audio segment that is identifiable by the human
being to be related to a specific context [29]. In this way, a human being can identify the proximity of a vehicle
by listening to its engine or establish the aggressivity of a dog by listening to its bark. AED corresponds to the
identification of specific parameters in the audio signal that allow classifying the event that produces the sound.
This technique has been successfully applied in acoustic surveillance applications, audio signal labeling, and
environmental sound identification [30], [31].
The analysis strategies used in AED are of two types, those applied to monophonic sounds and those
used with polyphonic sounds, in which different events appear in the same audio signal. Most of the recent
research focuses on monophonic sounds under the assumption that the predominant acoustic event in the audio
sample is identified and characterized. This predominant event corresponds to an anomaly (a rare event) and
therefore should be easily separable from the rest of the sample. This approach goes hand in hand with the
development of systems for real-life tasks, since acoustic events are usually mixed with other sounds. In
addition, a good AED must be able to isolate and identify the event of interest regardless of the environment
(other sounds in the audio), which becomes irrelevant in the identification.
From this perspective, AED systems correspond to one of the strategies of computational auditory
scene analysis (CASA). These strategies are based on the development of artificial systems from computational
tools that can similarly isolate sounds as human beings do. This type of processing is performed by identify-
ing parameters and characteristics in the audio that allow classification. The audio is divided into segments,
and then the characteristics of interest are extracted from each segment. Something similar is done in the
development of trainable models, in which these features are paired with event labels.
Machine learning techniques have brought more tools to the process, including mel-frequency cepstral
coefficients (MFCCs), and perceptual linear prediction coefficients [32], [33]. The mel-scale is of particular
importance since it allows to express of digital acoustic signals on a scale following the human perception of
sounds. It corresponds to a perceptual scale of tones defined by human observers, which allows an artificial
system to respond to stimuli close to those experienced by humans. This scale has as a reference point a
1,000 Hz tone at 40 dBs above the auditory threshold, which is leveled with a 1,000 mels tone. In addition,
above 500 Hz the frequency intervals are exponentially separated, which is perceived by the human ear as a
more linear spacing. To convert a frequency f in Hertz to a value m in mels, (1) is used.
m = 1127.01048 loge
1 +
f
700
(1)
Convolutional networks have improved the performance of automatic classification systems, particu-
larly in applications with images [34]. This is convenient since musicians and psychologists often choose to
generate a two-dimensional representation of the mel-scale. To construct these representations, tone color or
chroma is assigned, as well as a pitch, which generates an image with the audio information itself. Many deep
convolutional network models have been proposed, which have been evaluated both with public datasets and
in proprietary applications. We propose to select some of these classification models to determine the actual
performance of a real-time event identification task.
3. METHODS AND MATERIALS
Our AED system is supported on deep networks, so the training database and its pre-processing are
critical. The goal is that the system will be robust to the presence of ambient noise, so the audio samples used in
training the models must contain background noise, as expected from the real-world operation. We selected a
Int J Elec Comp Eng, Vol. 12, No. 6, December 2022: 6684–6696
4. Int J Elec Comp Eng ISSN: 2088-8708 ❒ 6687
set of public databases that meet these characteristics. Not all samples in our dataset come from a single public
database, to avoid bias we mix audios from different databases, and some were even recorded by us in the lab.
In addition, in the cases where the human voice was required, we tried to make the database gender-balanced
(males comprise 41%). This database was separated into a training set and a validation set. We used 80% of
the audios in each category for training, and the remaining 20% for validation. The selection of the training and
validation sets was completely random, but the same set was used to train each convolutional model. In total,
we selected three deep architectures according to previous laboratory performance evaluation: residual neural
network (ResNet), dense convolutional network (DenseNet), and neural architecture search network (NASNet).
The fitted models were evaluated with the same metrics on the same validation set. Figure 1 shows the details
of our framework.
Figure 1. Framework to evaluate the robustness of CNN based AED
We examined different public databases to build the dataset used in our experiments. We used the
common voice dataset to build the base category corresponding to the human voice in a natural and quiet
(normal) state. This dataset contains the voice of people between 19 and 89 years old, corresponding to reading
blog posts, books, and movies mainly [35]. We also used the UrbanSound Dataset from which we extracted the
category corresponding to the gunshot. This database is composed of labeled sounds corresponding to urban
events such as children playing, dogs barking, vehicle sirens or gunshots [36]. Another key database in the
construction of our dataset is neural information processing group GENeral sounds (NIGENS), this database
also provides isolated urban acoustic events. From NIGENS we extracted the categories for crying baby (crying
baby), burning fire (burning fire), human screams of men and women (scream), and dog barking (dog) [37].
The last database used was TUT Rare Sound Events 2017, which was developed for the DCASE challenge
2017 Task 2, and is composed of events corresponding to babies crying, gunshots, and glass breaking. All
images from this dataset were used either to complement the already defined categories or to create the new
category corresponding to glass breakage (glass) [38]. Some of these databases provide the audio files in MP3
format, others in WAV format. All files were initially unified to WAV format. In the case of MP3 files, the
FFmpeg library was used to convert them to WAV format.
Our dataset is made up of 2,100 sounds (300 per category) with durations between 2 and 40 s. The
Mel spectrogram for the audio files was calculated using the librosa 0.8.1 library. The audio signals in WAV
format were sampled as a time-series input at a sampling rate of 22,050, the magnitude of the spectrogram was
calculated and then mapped onto the Mel scale. An FFT window of 1,024 was used, and 100 samples were be-
tween successive frames. Figure 2 shows some images resulting from this conversion process, Figure 2(a)
shows an audio corresponding to the normal base category, Figure 2(b) audio from the gunshot category,
Figure 2(c) audio from the baby crying category, Figure 2(d) audio from the burning fire category, Figure 2(e)
audio from the human scream category, Figure 2(f) audio from the dog barking category, and in Figure 2(g)
audio from the glass breaking category.
Acoustic event characterization for service robot using convolutional networks (Fernando Martı́nez)
5. 6688 ❒ ISSN: 2088-8708
(a) (b)
(c) (d)
(e) (f)
(g)
Figure 2. Sample of some Mel scale images used for model training (a) normal category, (b) gunshot category,
(c) baby crying category, (d) burning fire category, (e) human scream category, (f) dog barking category, and
(g) glass breaking category
The images resulting from this process originally had a size of 432×288 pixels. For ease of testing,
this size was modified to a square structure of 256×256 pixels using OpenCV 4.1.2. For this process, the
images were initially randomly shuffled by changing the seed for each category, and then stored in folders
labeled with the name of the corresponding category. These files had different characteristics such as length
Int J Elec Comp Eng, Vol. 12, No. 6, December 2022: 6684–6696
6. Int J Elec Comp Eng ISSN: 2088-8708 ❒ 6689
(some were 1 s, but others were as long as 2 min), number of channels, and frequencies. We pre-processed
these audios to normalize the frequency (they were adjusted to the bandwidth of the human ear), unified the
number of channels to monaural, and adjusted the length of the audios from 2 to 40 s. The seven categories or
labels are:
− Category 1: normal
− Category 2: gunshot
− Category 3: cryingbaby
− Category 4: burningfire
− Category 5: scream
− Category 6: dog
− Category 7: glass
The normal category corresponds to the voice of people in a quiet state reading texts, the second
category corresponds to the sounds of firearms in different scenarios, the third is the sound of small babies
crying, the fourth corresponds to the sound of flames burning different materials, the scream category again
has the voice of men and women but producing screams, category six includes barking dogs, and category
seven corresponds to sounds produced by breaking glass. The content of the audios in the public databases
was verified and edited to ensure the content of the event as the protagonist in each file. Most of the audios
correspond to polyphonic events.
We selected three convolutional architectures for the models. The selection was made according to
their high performance on similar problems, and the small size of the models that make them suitable for
implementation on the robot. The research group had previously used these architectures in other modules
of our robotic platform to solve problems such as automatic emotion recognition and movement strategies in
unknown and dynamic environments [19], [20]. The selected architectures are ResNet, DenseNet, and NASNet.
For the ResNet architecture, we selected the ResNet-50 model with 50 depth layers for a total of 23,602,055
parameters. For the DenseNet architecture, we selected DenseNet-121 for a total of 7,044,679 parameters,
and for the NASNet architecture, we selected NASNet-Mobile for a total of 4,277,115 parameters. All models
were fit for 20 epochs under the same criteria, i.e., categorical cross-entropy was used as the loss function,
and stochastic gradient descent (SGD) was used as the optimizer. In all cases, the optimization process was
monitored by calculating both the accuracy and the mean squared error (MSE) of the model with the training
and validation samples. Table 1 summarizes the number of parameters related to each model.
Table 1. Parameters related to each model
Model ResNet 50 DenseNet 121 NASNet-Mobile
Non-trainable parameters: 53,120 83,648 36,738
Trainable parameters: 23,548,935 6,961,031 4,240,377
Total parameters: 23,602,055 7,044,679 4,277,115
In all three cases, the training was optimized according to the behavior of its accuracy with both
training and validation data (although the latter were not directly considered for the optimization process).
In the case of the ResNet network the accuracy was increased from 32.9% to 90.2% for the training data
(considering the training process, i.e., epoch 1), and from 14.8% to 85.0% for the validation data. The
DenseNet network achieved an increase from 49.8% to 90.1% in the accuracy of the training data, and from
12.6% to 81.7% for the validation data. Finally, for the NASNet network, an increase in accuracy from
56.1% to 91.9% was achieved for the training data, but 16.2% for the validation data was not achieved. All
code was developed in Python 3.7 with support for Keras 2.6.0 and Tensorflow 2.6.0.
4. RESULT AND DISCUSSION
For the evaluation of the ability of the convolutional models to identify the parameters of each event
correctly, standard machine learning metrics were used in the evaluation of classification models. These metrics
were calculated from the behavior of the models against 20% of the dataset unknown to the model and initially
separated for validation. As mentioned in the methodological development, during model training, accuracy
and MSE were tracked at each epoch for both training and validation data, the results are summarized in
Acoustic event characterization for service robot using convolutional networks (Fernando Martı́nez)
7. 6690 ❒ ISSN: 2088-8708
Figure 3, Figure 3(a) shows the performance for the ResNet model, Figure 3(b) shows the performance for the
DenseNet model, and Figure 3(c) shows the performance for the NASNet model.
(a) (b)
(c)
Figure 3. Summary of accuracy and error of the three models throughout training (a) ResNet, (b) DenseNet,
and (c) NASNet
The smallest model is NASNet, however, it is also the model with the worst generalization ability. All
three models managed to considerably reduce the error against the training data during the first five epochs;
ResNet achieved an error of 0.48 and an accuracy of 81.5%, DenseNet showed an error of 0.42 and an accuracy
of 84.2%, and NASNet an error of 0.25 with an accuracy of 88.2%. Although the best values for the train-
ing dataset were achieved by NASNet, this model suffered overfitting that prevented it from generalizing its
categorization capability to unknown data. From the fifth epoch onwards, ResNet and DenseNet considerably
reduce the error against validation data without reducing their performance on the training data, while NASNet
reduces it but without achieving good performance. At the end of training ResNet and DesnseNet achieve an
error on validation data of 0.03, while NASNet only achieves a value of 0.23. Similar behavior is observed in
the accuracy of the validation data.
The metrics used to evaluate the models are precision, Recall, F1-score, confusion matrix, and receiver
operating characteristic (ROC) curves. Precision evaluates how many elements assigned to a category really
belong to it. It corresponds to the ratio of true positives to the total number of elements identified in a category
(true positives plus false positives). The Recall metric allows establishing how many of the elements assigned
to a category by the model really belong to that category. It is calculated as the ratio between true positives
concerning the elements that actually belong to the category (true positives plus false negatives). These two
metrics can be visually summarized in the confusion matrix, since the main diagonal of this matrix corresponds
to the true positives, the elements above it correspond to the false positives, and the elements below it corre-
spond to the false negatives. To build this matrix, a table is assembled in which the rows correspond to the
real categories of the elements, and the columns to the elements classified according to the model, values with
Int J Elec Comp Eng, Vol. 12, No. 6, December 2022: 6684–6696
8. Int J Elec Comp Eng ISSN: 2088-8708 ❒ 6691
which the matrix is filled. The F1-score value corresponds to the weighted average of precision and Recall, so
it somehow summarizes these two metrics. Finally, the ROC curve corresponds to a graphical representation of
the sensitivity versus the specificity of the classifier. It uses changes in the threshold of the decision to evaluate
the ratio of true positives to false positives. This graph is constructed with a diagonal curve, and all points
above it (towards the upper left corner) are considered good classification results.
Table 2 outlines the results of the first three metrics. The values in the table are related to the previous
curves. The ResNet and DenseNet models perform much better than the NASNet model, which performed
poorly for the validation data. The only interesting value of the NASNet model corresponds to the Recall of
the sixth category, with a value of 100%. This contrasts with the values for the other metrics in all categories
but shows that the model was unable to categorize the elements of the validation group, assigning them all to a
single category, the one corresponding to dog barking. Leaving the NASNet model aside, the performance of
the first two models is excellent. In both cases, the average values of the three metrics were in the worst case at
79%. These values indicate that these models classify most of the validation values in the correct category. In
both cases, however, problems are noted in two categories, in the third category (cryingbaby) very low Recall
values are obtained, and in the seventh category (glass) very low precision values are obtained. In the first case,
it is observed that many elements of this category were classified in other categories, i.e., the crying baby is
not adequately differentiated from other sounds. In the second case, it means that many of the elements of this
category do not really belong to it, i.e. it confuses as glass breakage many of the sounds of other categories.
Table 2. Precision, Recall, and F1-score values for the three models with validation data
Category ResNet DenseNet NasNet
Precission Recall F1-Score Precission Recall F1-Score Precission Recall F1-Score
normal 1.00 1.00 1.00 1.00 0.98 0.99 00.00 00.00 00.00
gunshot 1.00 0.98 0.99 0.90 0.98 0.94 00.00 00.00 00.00
crying baby 0.88 0.19 0.31 0.79 0.20 0.32 00.00 00.00 00.00
burningfire 1.00 1.00 1.00 0.81 1.00 0.89 00.00 00.00 00.00
scream 0.96 1.00 0.98 0.98 1.00 0.99 00.00 00.00 00.00
dog 1.00 1.00 1.00 1.00 1.00 1.00 00.16 1.00 00.28
glass 0.46 0.96 0.63 0.42 0.70 0.52 00.00 00.00 00.00
Weighted avr. 0.91 0.85 0.83 0.85 0.82 0.79 00.03 00.16 00.05
The problem identified in the metrics can be observed and analyzed in the confusion matrix in Figure 4,
Figure 4(a) shows the confusion matrix for the ResNet model, Figure 4(b) shows the confusion matrix for the
DenseNet model, and Figure 4(c) shows the confusion matrix for the NASNet model. The confusion matrices
of the ResNet and DesneNet models show that in fact the problems in the third and seventh categories are
related. The plots of these matrices use color-coding that makes it easy to identify the behavior of the diagonal
of the matrix, and the classification model it represents. Dark colors show a lower amount of elements in the
box, while light colors show a high concentration (scale on the right of each matrix). This coding should show
in light colors the diagonal of a good classifier. In the ResNet and DenseNet models, this diagonal is quite well
defined for most categories, but as before, problems are observed in the third and seventh categories. The third
category looks in both cases very dark, and in the same row, in the seventh column, a high number of false
positives are observed. The matrices show that many of the sounds corresponding to baby cries are erroneously
placed in the category glass breakage. While this may be an isolated problem, the fact that both models have
the same inconsistency suggests problems in the audio quality of the training dataset. In the third matrix, the
one corresponding to the NASNet model, it is observed as before that all items were categorized in the sixth
category. This caused the audios that belonged to the category to match, but the overall performance was poor
due to the misclassification of all the others.
Since the ROC curve considers only the ratio of true positives to false positives, it is to be expected that
the behavior for the NASNet model moves close to the diagonal, while for the ResNet and DenseNet models
the curve seeks the upper left corner as shown in Figure 5, Figure 5(a) shows the ROC curve for the ResNet
model, Figure 5(b) shows the ROC curve for the DenseNet model, and Figure 5(c) shows the ROC curve for the
NASNet model. This means that the latter two models can correctly classify the elements that actually belong
to a certain category (vertical axis of the ROC curve), something equivalent to the Recall metric. In short, the
performance of the ResNet and DenseNet models is very high and similar to each other, while the NASNet
model is inadequate for the development of the event characterization system.
Acoustic event characterization for service robot using convolutional networks (Fernando Martı́nez)
9. 6692 ❒ ISSN: 2088-8708
(a)
(b)
(c)
Figure 4. Confusion matrix of the three models with validation data (a) ResNet, (b) DenseNet, and (c) NASNet
Int J Elec Comp Eng, Vol. 12, No. 6, December 2022: 6684–6696
10. Int J Elec Comp Eng ISSN: 2088-8708 ❒ 6693
(a) (b)
(c)
Figure 5. ROC curves of the three models with validation data (a) ResNet, (b) DenseNet, and (c) NASNet
5. LIMITATION
Our experiments relied on audio signals captured by commercial microphones, in some cases installed
in low-quality devices. While this approach goes hand in hand with the existing microphones in our robot, and
with the hearing ability of an average adult, it is clear that the audio quality and performance capability of
the acoustic event characterization system is increased with the use of higher-quality sensors. Future develop-
ments of our system contemplate the use of these sensors and the development of a proprietary database that
incorporates the acoustic signals captured by them.
The other limitation of our research is its narrow focus, which is primarily on domestic service robots.
While this is our field of research, and the scope of our robotic prototypes, this limited perspective also limits
the usefulness of our results in other scenarios such as applications in commercial and industrial environments.
Still, we believe that the general scheme of work, as well as the results with the evaluated convolutional models,
can serve as an initial stage for these application fields.
6. CONCLUSION
In this work, we propose the development of an autonomous system for the characterization of acoustic
events that can be implemented in small service robots, particularly dedicated to the development of domestic
tasks. In this sense, a scheme based on convolutional networks capable of being trained for a dataset specific
to the needs of the problem is proposed. A dataset is built according to seven specific events: human voice
in a normal state, gunshot, baby crying, fire in burning processes, human screams, dog barking, and glass
breaking. For the construction of these categories, the use of public databases was chosen to evaluate the initial
performance of the convolutional models. A database with 300 audios in each category was assembled, which
were randomly separated into two groups, a training group with 80% of the data, and a validation group with the
Acoustic event characterization for service robot using convolutional networks (Fernando Martı́nez)
11. 6694 ❒ ISSN: 2088-8708
remaining 20%. The audio signals were filtered to produce a homogeneous set of different parameters. To feed
the convolutional models with an input signal that reflected the characteristics identifiable by the human ear, the
entire dataset was converted to the Mel scale. The code was developed in Python with Keras and Tensorflow
support, and the processing of the audios was performed with librosa and then transformed into images with
OpenCV. The three convolutional models evaluated for the system were ResNet, DenseNet, and NASNet, these
were selected due to their high performance in similar tasks and their compact size compared to other structures.
The training was performed in all three cases under the same conditions, which involved the use of categorical
cross-entropy as loss function and SGD as optimization function. The training was performed for 20 epochs,
during which the error and accuracy of the training and validation data were calculated for model control and
adjustment. To evaluate the performance, the Precision, Recall, F1-score, confusion matrix, and ROC curve
metrics were used on the models trained with the validation data. The results provided by the metrics confirm
a high performance for the ResNet and DenseNet models (average F1-score of 83% for ResNet and 79% for
DenseNet), while they show that the NASNet model is unable to generalize the behavior to unknown data.
Some problems were observed in two categories which we intend to evaluate with our dataset captured in the
laboratory. Finally, although the ResNet model outperforms the DenseNet model in some values, the latter
is selected for the development of the prototype system due to its smaller size and memory requirements.
The research will continue adjusting the model to the conditions of the robotic platform, which implies the
reconstruction of a more complex and higher-quality dataset.
ACKNOWLEDGEMENT
This work was supported by the Universidad Distrital Francisco José de Caldas, in part through CIDC,
and partly by the Facultad Tecnológica. The views expressed in this paper are not necessarily endorsed by Uni-
versidad Distrital. The authors thank the research group ARMOS for the evaluation carried out on prototypes
of ideas and strategies.
REFERENCES
[1] J. Murphy, J. McVay, P. Mathews, D. A. Carnegie, and A. Kapur, “Expressive robotic guitars: developments
in musical robotics for chordophones,” Computer Music Journal, vol. 39, no. 1, pp. 59–73, Mar. 2015, doi:
10.1162/COMJ a 00285.
[2] R. O’Dowd, “Emerging trends and new directions in telecollaborative learning,” CALICO Journal, vol. 33, no. 3, pp.
291–310, Aug. 2016, doi: 10.1558/cj.v33i3.30747.
[3] A. Linson, C. Dobbyn, G. E. Lewis, and R. Laney, “A subsumption agent for collaborative free improvisation,”
Computer Music Journal, vol. 39, no. 4, pp. 96–115, Dec. 2015, doi: 10.1162/COMJ a 00323.
[4] E. J. Fabris, V. A. Sangalli, L. P. Soares, and M. S. Pinho, “Immersive telepresence on the operation of unmanned vehi-
cles,” International Journal of Advanced Robotic Systems, vol. 18, no. 1, Jan. 2021, doi: 10.1177/1729881420978544.
[5] X. Zhao et al., “A smart robotic walker with intelligent close-proximity interaction capabilities for elderly mobility
safety,” Frontiers in Neurorobotics, vol. 14, no. 1, pp. 1–8, Oct. 2020, doi: 10.3389/fnbot.2020.575889.
[6] C. Kertész and M. Turunen, “Common sounds in bedrooms (CSIBE) corpora for sound event recognition of domestic
robots,” Intelligent Service Robotics, vol. 11, no. 4, pp. 335–346, 2018, doi: 10.1007/s11370-018-0258-9.
[7] M. Podpora, A. Gardecki, R. Beniak, B. Klin, J. L. Vicario, and A. Kawala-Sterniuk, “Human interaction smart
subsystem-extending speech-based human-robot interaction systems with an implementation of external smart sen-
sors,” Sensors, vol. 20, no. 8, Apr. 2020, doi: 10.3390/s20082376.
[8] E. Dean-Leon, K. Ramirez-Amaro, F. Bergner, I. Dianov, and G. Cheng, “Integration of robotic technologies for
rapidly deployable robots,” IEEE Transactions on Industrial Informatics, vol. 14, no. 4, pp. 1691–1700, 2018, doi:
10.1109/TII.2017.2766096.
[9] X. Lin, X. Zou, Z. Ji, T. Huang, S. Wu, and Y. Mi, “A brain-inspired computational model for spatio-temporal
information processing,” Neural Networks, vol. 143, no. 1, pp. 74–87, Nov. 2021, doi: 10.1016/j.neunet.2021.05.015.
[10] R. M. Alsina-Pagès, R. Benocci, G. Brambilla, and G. Zambon, “Methods for noise event detection and as-
sessment of the sonic environment by the harmonica index,” Applied Sciences, vol. 11, no. 17, Aug. 2021, doi:
10.3390/app11178031.
[11] R. Löwe, J. Böhm, D. G. Jensen, J. Leandro, and S. H. Rasmussen, “U-FLOOD-Topographic deep learn-
ing for predicting urban pluvial flood water depth,” Journal of Hydrology, vol. 603, pp. 1–6, 2021, doi:
10.1016/j.jhydrol.2021.126898.
[12] T. C. Pham, C. M. Luong, V. D. Hoang, and A. Doucet, “AI outperformed every dermatologist in dermoscopic
melanoma diagnosis, using an optimized deep-CNN architecture with custom mini-batch logic and loss function,”
Int J Elec Comp Eng, Vol. 12, No. 6, December 2022: 6684–6696
12. Int J Elec Comp Eng ISSN: 2088-8708 ❒ 6695
Scientific Reports, vol. 11, no. 1, pp. 1–12, 2021, doi: 10.1038/s41598-021-96707-8.
[13] Y. Xu and M. Vaziri-Pashkam, “Limits to visual representational correspondence between convolutional neural net-
works and the human brain,” Nature Communications, vol. 12, no. 1, 2021, doi: 10.1038/s41467-021-22244-7.
[14] J. Karhade, S. K. Ghosh, P. Gajbhiye, R. K. Tripathy, and U. R. Acharya, “Multichannel multiscale two-stage con-
volutional neural network for the detection and localization of myocardial infarction using vectorcardiogram signal,”
Applied Sciences, vol. 11, no. 17, 2021, doi: 10.3390/app11177920.
[15] Z. Nabulsi et al., “Deep learning for distinguishing normal versus abnormal chest radiographs and generaliza-
tion to two unseen diseases tuberculosis and COVID-19,” Scientific Reports, vol. 11, no. 1, pp. 1–10, 2021, doi:
10.1038/s41598-021-93967-2.
[16] A. De Cesarei, S. Cavicchi, G. Cristadoro, and M. Lippi, “Do humans and deep convolutional neural networks use
visual information similarly for the categorization of natural scenes?,” Cognitive Science, vol. 45, no. 6, pp. 1–14,
2021, doi: 10.1111/cogs.13009.
[17] T. Hachaj, Ł. Bibrzycki, and M. Piekarczyk, “Recognition of cosmic ray images obtained from cmos sensors used
in mobile phones by approximation of uncertain class assignment with deep convolutional neural network,” Sensors,
vol. 21, no. 6, pp. 1–16, 2021, doi: 10.3390/s21061963.
[18] D. Gibert, C. Mateu, J. Planes, and R. Vicens, “Using convolutional neural networks for classification of malware
represented as images,” Journal of Computer Virology and Hacking Techniques, vol. 15, no. 1, pp. 15–28, 2019, doi:
10.1007/s11416-018-0323-0.
[19] F. Martinez, C. Hernández, and A. Rendón, “Identifier of human emotions based on convolutional neural network
for assistant robot,” TELKOMNIKA (Telecommunication Computing Electronics and Control), vol. 18, no. 3, pp.
1499-1504, Jun. 2020, doi: 10.12928/telkomnika.v18i3.14777.
[20] F. Martı́nez, C. Penagos, and L. Pacheco, “Deep regression models for local interaction in multi-agent robot tasks,”
in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes
in Bioinformatics), vol. 10942, 2018, pp. 66–73, doi: 10.1007/978-3-319-93818-9 7.
[21] M. Arduengo, C. Torras, and L. Sentis, “Robust and adaptive door operation with a mobile robot,” Intelligent Service
Robotics, vol. 14, no. 3, pp. 409–425, 2021, doi: 10.1007/s11370-021-00366-7.
[22] J. C. Molina-Molina, M. Salhaoui, A. Guerrero-González, and M. Arioua, “Autonomous marine robot based on
ai recognition for permanent surveillance in marine protected areas,” Sensors, vol. 21, no. 8, pp. 1–28, 2021, doi:
10.3390/s21082664.
[23] K. Zheng, F. Wu, and X. Chen, “Laser-based people detection and obstacle avoidance for a hospital transport robot,”
Sensors, vol. 21, no. 3, pp. 1–24, 2021, doi: 10.3390/s21030961.
[24] J. Castañeda and Y. Salguero, “Adjustment of visual identification algorithm for use in stand-alone robot navigation
applications,” Tekhnê, vol. 14, no. 1, pp. 73–86, 2017.
[25] A. Moreno and D. F. Páez, “Performance evaluation of ROS on the Raspberry Pi platform as OS for small robots,”
Tekhnê, vol. 14, no. 1, pp. 61–72, 2017.
[26] A. Mehrotra et al., “Accessible maker-based approaches to educational robotics in online learning,” IEEE Access,
vol. 9, no. 9471866, pp. 96877–96889, 2021, doi: 10.1109/ACCESS.2021.3094158.
[27] Q. Ji, C. Dai, C. Hou, and X. Li, “Real-time embedded object detection and tracking system in Zynq SoC,” EURASIP
Journal on Image and Video Processing, vol. 2021, no. 1, Dec. 2021, doi: 10.1186/s13640-021-00561-7.
[28] G. H. Kim, E. S. Sung, and K. W. Nam, “Automated laryngeal mass detection algorithm for home-based self-
screening test based on convolutional neural network,” BioMedical Engineering Online, vol. 20, no. 1, pp. 1–10,
2021, doi: 10.1186/s12938-021-00886-4.
[29] K. Imoto, “Introduction to acoustic event and scene analysis,” Acoustical Science and Technology, vol. 39, no. 3, pp.
182–188, 2018, doi: 10.1250/ast.39.182.
[30] E. Lieskovska, M. Jakubec, and R. Jarina, “Acoustic surveillance system for children’s emotion detection,” in
42nd
International Conference on Telecommunications and Signal Processing (TSP), Jul. 2019, pp. 525–528, doi:
10.1109/TSP.2019.8768884.
[31] S. S. Sethi et al., “Characterizing soundscapes across diverse ecosystems using a universal acoustic feature
set,” in Proceedings of the National Academy of Sciences, vol. 117, no. 29, pp. 17049–17055, Jul. 2020, doi:
10.1073/pnas.2004702117.
[32] I. D. S. Miranda, A. H. Diacon, and T. R. Niesler, “A comparative study of features for acoustic cough detection
using deep architectures,” in 41st
Annual International Conference of the IEEE Engineering in Medicine and Biology
Society (EMBC), Jul. 2019, pp. 2601–2605, doi: 10.1109/EMBC.2019.8856412.
[33] K. Feroze and A. R. Maud, “Sound event detection in real life audio using perceptual linear predictive feature with
neural network,” in 15th International Bhurban Conference on Applied Sciences and Technology (IBCAST), Jan.
2018, pp. 377–382, doi: 10.1109/IBCAST.2018.8312252.
[34] F. Martinez, F. Martı́nez, and E. Jacinto, “Performance evaluation of the NASNet convolutional network in the auto-
matic identification of COVID-19,” International Journal on Advanced Science, Engineering and Information Tech-
nology, vol. 10, no. 2, Apr. 2020, doi: 10.18517/ijaseit.10.2.11446.
Acoustic event characterization for service robot using convolutional networks (Fernando Martı́nez)
13. 6696 ❒ ISSN: 2088-8708
[35] R. Ardila et al., “Common voice: a massively-multilingual speech corpus,” arXiv preprint arXiv:1912.06670, 2019.
[36] J. Salamon, C. Jacoby, and J. P. Bello, “A dataset and taxonomy for urban sound research,” in 22nd
ACM International
Conference on Multimedia, Nov. 2014, pp. 1041–1044, doi: 10.1145/2647868.2655045.
[37] I. Trowitzsch, J. Taghia, Y. Kashef, and K. Obermayer, “The NIGENS general sound events database,” arXiv preprint
arXiv:1902.0831, pp. 1–5, Feb. 2019.
[38] A. Mesaros et al., “DCASE 2017 Challenge setup : Tasks, datasets and baseline system,” in Detection and Classifi-
cation of Acoustic Scenes and Events 2017 Workshop, 2017, pp. 85–92.
BIOGRAPHIES OF AUTHORS
Fernando Martı́nez is a doctoral researcher at the Universidad Distrital Francisco José
de Caldas focusing on the development of navigation strategies for autonomous vehicles using hi-
erarchical control schemes. In 2009 he completed his M.Sc. degree in Computer and Electronics
Engineering at Universidad de Los Andes, Colombia. He is a researcher of the ARMOS research
group (Modern Architectures for Power Systems) supporting the lines of electronic instrumentation,
control and robotics. He can be contacted at email: fmartinezs@udistrital.edu.co.
Fredy Martı́nez is a professor of control, intelligent systems, and robotics at the Universi-
dad Distrital Francisco José de Caldas (Colombia) and director of the ARMOS research group (Mod-
ern Architectures for Power Systems). His research interests are control schemes for autonomous
robots, mathematical modeling, electronic instrumentation, pattern recognition, and multi-agent sys-
tems. Martinez holds a Ph.D. in Computer and Systems Engineering from the Universidad Nacional
de Colombia. He can be contacted at email: fhmartinezs@udistrital.edu.co.
César Hernández is a professor of telecommunications, digital signal processing, and
electronics at the Universidad Distrital Francisco José de Caldas (Colombia) and director of the
SIREC research group (Systems and Cognitive Networks). His research interests are telecommu-
nications, assistive technology, tele informatics, and advanced digital systems. Hernández holds a
Ph.D. in Computer and Systems Engineering from the Universidad Nacional de Colombia. He can
be contacted at email: cahernandezs@udistrital.edu.co.
Int J Elec Comp Eng, Vol. 12, No. 6, December 2022: 6684–6696