Using a camera to monitor an object or a group of objects over time is the process of object detection. It can be used for a variety of things, including security and surveillance, video communication, traffic light detection (TLD), object detection from compressed video in public places. In recent times, object tracking has become a popular topic in computer science particularly, the data science community, thanks to the usage of deep learning (DL) in artificial intelligence (AI). DL which convolutional neural network (CNN) as one of its techniques usually used two-stage detection methods in TLD. Despite all successes recorded in TLD through the use of two-stage detection methods, there is no study that has analyzed these methods in experimental research, studying the strength and witnesses by the researchers. Based on the needs this study analyses the applications of DL techniques in TLD. We implemented object detection for TLD using 5 two-stage detection methods with the traffic light dataset using a Jupyter notebook and the sklearn libraries. We present the achievements of two-stage detection methods in TLD, going by standard performance metrics used, FASTER-CNN was the best in detection accuracy, F1-score, precision and recall with 0.89, 0.93, 0.83 and 0.90 respectively.
Accident vehicle types classification: a comparative study between different...nooriasukmaningtyas
Classifying and finding type of individual vehicles within an accident image are considered difficult problems. This research concentrates on accurately classifying and recognizing vehicle accidents in question. The aim to provide a comparative analysis of vehicle accidents. A number of network topologies are tested to arrive at convincing results and a variety of matrices are used in the evaluation process to identify the best networks. The best two networks are used with faster recurrent convolution neural network (Faster RCNN) and you only look once (YOLO) to determine which network will identifiably detect the location and type of the vehicle. In addition, two datasets are used in this research. In consequence, experiment results show that MobileNetV2 and ResNet50 have accomplished higher accuracy compared to the rest of the models, with 89.11% and 88.45% for the GAI dataset as well as 88.72% and 89.69% for KAI dataset, respectively. The findings reveal that the ResNet50 base network for YOLO achieved higher accuracy than MobileNetV2 for YOLO, ResNet50 for Faster RCNN with 83%, 81%, and 79% for GAI dataset and 79%, 78% and 74% for KAI dataset.
Through the generalization of deep learning, the research community has addressed critical challenges in
the network security domain, like malware identification and anomaly detection. However, they have yet to
discuss deploying them on Internet of Things (IoT) devices for day-to-day operations. IoT devices are often
limited in memory and processing power, rendering the compute-intensive deep learning environment
unusable. This research proposes a way to overcome this barrier by bypassing feature engineering in the
deep learning pipeline and using raw packet data as input. We introduce a feature- engineering-less
machine learning (ML) process to perform malware detection on IoT devices. Our proposed model,”
Feature engineering-less ML (FEL-ML),” is a lighter-weight detection algorithm that expends no extra
computations on “engineered” features. It effectively accelerates the low-powered IoT edge. It is trained
on unprocessed byte-streams of packets. Aside from providing better results, it is quicker than traditional
feature-based methods. FEL-ML facilitates resource-sensitive network traffic security with the added
benefit of eliminating the significant investment by subject matter experts in feature engineering.
EFFICIENT ATTACK DETECTION IN IOT DEVICES USING FEATURE ENGINEERING-LESS MACH...ijcsit
Through the generalization of deep learning, the research community has addressed critical challenges in
the network security domain, like malware identification and anomaly detection. However, they have yet to
discuss deploying them on Internet of Things (IoT) devices for day-to-day operations. IoT devices are often
limited in memory and processing power, rendering the compute-intensive deep learning environment
unusable. This research proposes a way to overcome this barrier by bypassing feature engineering in the
deep learning pipeline and using raw packet data as input. We introduce a feature- engineering-less
machine learning (ML) process to perform malware detection on IoT devices. Our proposed model,”
Feature engineering-less ML (FEL-ML),” is a lighter-weight detection algorithm that expends no extra
computations on “engineered” features. It effectively accelerates the low-powered IoT edge. It is trained
on unprocessed byte-streams of packets. Aside from providing better results, it is quicker than traditional
feature-based methods. FEL-ML facilitates resource-sensitive network traffic security with the added
benefit of eliminating the significant investment by subject matter experts in feature engineering.
Satellite and Land Cover Image Classification using Deep Learningijtsrd
Satellite imagery is very significant for many applications including disaster response, law enforcement and environmental monitoring. These applications require the manual identification of objects and facilities in the imagery. Because the geographic area to be covered are great and the analysts available to conduct the searches are few, automation is required. The traditional object detection and classification algorithms are too inaccurate, takes a lot of time and unreliable to solve the problem. Deep learning is a family of machine learning algorithms that can be used for the automation of such tasks. It has achieved success in image classification by using convolutional neural networks. The problem of object and facility classification in satellite imagery is considered. The system is developed by using various facilities like Tensor Flow, XAMPP, FLASK and other various deep learning libraries. Roshni Rajendran | Liji Samuel "Satellite and Land Cover Image Classification using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-5 , August 2020, URL: https://www.ijtsrd.com/papers/ijtsrd32912.pdf Paper Url :https://www.ijtsrd.com/computer-science/other/32912/satellite-and-land-cover-image-classification-using-deep-learning/roshni-rajendran
End-to-end deep auto-encoder for segmenting a moving object with limited tra...IJECEIAES
Deep learning-based approaches have been widely used in various applications, including segmentation and classification. However, a large amount of data is required to train such techniques. Indeed, in the surveillance video domain, there are few accessible data due to acquisition and experiment complexity. In this paper, we propose an end-to-end deep auto-encoder system for object segmenting from surveillance videos. Our main purpose is to enhance the process of distinguishing the foreground object when only limited data are available. To this end, we propose two approaches based on transfer learning and multi-depth auto-encoders to avoid over-fitting by combining classical data augmentation and principal component analysis (PCA) techniques to improve the quality of training data. Our approach achieves good results outperforming other popular models, which used the same principle of training with limited data. In addition, a detailed explanation of these techniques and some recommendations are provided. Our methodology constitutes a useful strategy for increasing samples in the deep learning domain and can be applied to improve segmentation accuracy. We believe that our strategy has a considerable interest in various applications such as medical and biological fields, especially in the early stages of experiments where there are few samples.
Accident vehicle types classification: a comparative study between different...nooriasukmaningtyas
Classifying and finding type of individual vehicles within an accident image are considered difficult problems. This research concentrates on accurately classifying and recognizing vehicle accidents in question. The aim to provide a comparative analysis of vehicle accidents. A number of network topologies are tested to arrive at convincing results and a variety of matrices are used in the evaluation process to identify the best networks. The best two networks are used with faster recurrent convolution neural network (Faster RCNN) and you only look once (YOLO) to determine which network will identifiably detect the location and type of the vehicle. In addition, two datasets are used in this research. In consequence, experiment results show that MobileNetV2 and ResNet50 have accomplished higher accuracy compared to the rest of the models, with 89.11% and 88.45% for the GAI dataset as well as 88.72% and 89.69% for KAI dataset, respectively. The findings reveal that the ResNet50 base network for YOLO achieved higher accuracy than MobileNetV2 for YOLO, ResNet50 for Faster RCNN with 83%, 81%, and 79% for GAI dataset and 79%, 78% and 74% for KAI dataset.
Through the generalization of deep learning, the research community has addressed critical challenges in
the network security domain, like malware identification and anomaly detection. However, they have yet to
discuss deploying them on Internet of Things (IoT) devices for day-to-day operations. IoT devices are often
limited in memory and processing power, rendering the compute-intensive deep learning environment
unusable. This research proposes a way to overcome this barrier by bypassing feature engineering in the
deep learning pipeline and using raw packet data as input. We introduce a feature- engineering-less
machine learning (ML) process to perform malware detection on IoT devices. Our proposed model,”
Feature engineering-less ML (FEL-ML),” is a lighter-weight detection algorithm that expends no extra
computations on “engineered” features. It effectively accelerates the low-powered IoT edge. It is trained
on unprocessed byte-streams of packets. Aside from providing better results, it is quicker than traditional
feature-based methods. FEL-ML facilitates resource-sensitive network traffic security with the added
benefit of eliminating the significant investment by subject matter experts in feature engineering.
EFFICIENT ATTACK DETECTION IN IOT DEVICES USING FEATURE ENGINEERING-LESS MACH...ijcsit
Through the generalization of deep learning, the research community has addressed critical challenges in
the network security domain, like malware identification and anomaly detection. However, they have yet to
discuss deploying them on Internet of Things (IoT) devices for day-to-day operations. IoT devices are often
limited in memory and processing power, rendering the compute-intensive deep learning environment
unusable. This research proposes a way to overcome this barrier by bypassing feature engineering in the
deep learning pipeline and using raw packet data as input. We introduce a feature- engineering-less
machine learning (ML) process to perform malware detection on IoT devices. Our proposed model,”
Feature engineering-less ML (FEL-ML),” is a lighter-weight detection algorithm that expends no extra
computations on “engineered” features. It effectively accelerates the low-powered IoT edge. It is trained
on unprocessed byte-streams of packets. Aside from providing better results, it is quicker than traditional
feature-based methods. FEL-ML facilitates resource-sensitive network traffic security with the added
benefit of eliminating the significant investment by subject matter experts in feature engineering.
Satellite and Land Cover Image Classification using Deep Learningijtsrd
Satellite imagery is very significant for many applications including disaster response, law enforcement and environmental monitoring. These applications require the manual identification of objects and facilities in the imagery. Because the geographic area to be covered are great and the analysts available to conduct the searches are few, automation is required. The traditional object detection and classification algorithms are too inaccurate, takes a lot of time and unreliable to solve the problem. Deep learning is a family of machine learning algorithms that can be used for the automation of such tasks. It has achieved success in image classification by using convolutional neural networks. The problem of object and facility classification in satellite imagery is considered. The system is developed by using various facilities like Tensor Flow, XAMPP, FLASK and other various deep learning libraries. Roshni Rajendran | Liji Samuel "Satellite and Land Cover Image Classification using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-5 , August 2020, URL: https://www.ijtsrd.com/papers/ijtsrd32912.pdf Paper Url :https://www.ijtsrd.com/computer-science/other/32912/satellite-and-land-cover-image-classification-using-deep-learning/roshni-rajendran
End-to-end deep auto-encoder for segmenting a moving object with limited tra...IJECEIAES
Deep learning-based approaches have been widely used in various applications, including segmentation and classification. However, a large amount of data is required to train such techniques. Indeed, in the surveillance video domain, there are few accessible data due to acquisition and experiment complexity. In this paper, we propose an end-to-end deep auto-encoder system for object segmenting from surveillance videos. Our main purpose is to enhance the process of distinguishing the foreground object when only limited data are available. To this end, we propose two approaches based on transfer learning and multi-depth auto-encoders to avoid over-fitting by combining classical data augmentation and principal component analysis (PCA) techniques to improve the quality of training data. Our approach achieves good results outperforming other popular models, which used the same principle of training with limited data. In addition, a detailed explanation of these techniques and some recommendations are provided. Our methodology constitutes a useful strategy for increasing samples in the deep learning domain and can be applied to improve segmentation accuracy. We believe that our strategy has a considerable interest in various applications such as medical and biological fields, especially in the early stages of experiments where there are few samples.
Social distance and face mask detector system exploiting transfer learningIJECEIAES
As time advances, the use of deep learning-based object detection algorithms has also evolved leading to developments of new human-computer interactions, facilitating an exploration of various domains. Considering the automated process of detection, systems suitable for detecting violations are developed. One such applications is the social distancing and face mask detectors to control air-borne diseases. The objective of this research is to deploy transfer learning on object detection models for spotting violations in face masks and physical distance rules in real-time. The common drawbacks of existing models are low accuracy and inability to detect in real-time. The MobileNetV2 object detection model and YOLOv3 model with Euclidean distance measure have been used for detection of face mask and physical distancing. A proactive transfer learning approach is used to perform the functionality of face mask classification on the patterns obtained from the social distance detector model. On implementing the application on various surveillance footage, it was observed that the system could classify masked and unmasked faces and if social distancing was maintained or not with accuracies 99% and 94% respectively. The models exhibited high accuracy on testing and the system can be infused with the existing internet protocol (IP) cameras or surveillance systems for real-time surveillance of face masks and physical distancing rules effectively.
Real Time Object Detection System with YOLO and CNN Models: A ReviewSpringer
The field of artificial intelligence is built on object detection techniques. YOU ONLY LOOK
ONCE (YOLO) algorithm and it's more evolved versions are briefly described in this research survey. This
survey is all about YOLO and convolution neural networks (CNN) in the direction of real time object detection.
YOLO does generalized object representation more effectively without precision losses than other object
detection models. CNN architecture models have the ability to eliminate highlights and identify objects in any
given image. When implemented appropriately, CNN models can address issues like deformity diagnosis,
creating educational or instructive application, etc. This article reached at number of observations and
perspective findings through the analysis. Also it provides support for the focused visual information and
feature extraction in the financial and other industries, highlights the method of target detection and feature
selection, and briefly describes the development process of yolo algorithm
Traffic Congestion Detection Using Deep Learningijtsrd
Despite the huge amount of traffic surveillance videos and images have been accumulated in the daily monitoring, deep learning approaches have been underutilized in the application of traffic intelligent management and control. Traffic images, including various illumination, weather conditions, and vast scenarios are considered and preprocessed to set up a proper training dataset. In order to detect traffic congestion, a network structure is proposed based on residual learning to be pre trained and fine tuned. The network is then transferred to the traffic application and retrained with self established training dataset to generate the Traffic Net. The accuracy of Traffic Net to classify congested and uncongested road states reaches 99 for the validation dataset and 95 for the testing dataset. The trained model is stored in cloud storage for easy access for application from anywhere. The proposed Traffic Net can be used by a regional detection of traffic congestion on a large scale surveillance system. The effectiveness and efficiencies are magnificently demonstrated with quick detection in the high accuracy in the case study. The experimental trial could extend its successful application to traffic surveillance system and has potential enhancement for intelligent transport system in future. Anusha C | Dr. J. Bhuvana "Traffic Congestion Detection Using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-2 , February 2022, URL: https://www.ijtsrd.com/papers/ijtsrd49401.pdf Paper URL: https://www.ijtsrd.com/computer-science/other/49401/traffic-congestion-detection-using-deep-learning/anusha-c
Satellite Image Classification with Deep Learning Surveyijtsrd
Satellite imagery is important for many applications including disaster response, law enforcement and environmental monitoring etc. These applications require the manual identification of objects in the imagery. Because the geographic area to be covered is very large and the analysts available to conduct the searches are few, thus an automation is required. Yet traditional object detection and classification algorithms are too inaccurate and unreliable to solve the problem. Deep learning is a part of broader family of machine learning methods that have shown promise for the automation of such tasks. It has achieved success in image understanding by means that of convolutional neural networks. The problem of object and facility recognition in satellite imagery is considered. The system consists of an ensemble of convolutional neural networks and additional neural networks that integrate satellite metadata with image features. Roshni Rajendran | Liji Samuel ""Satellite Image Classification with Deep Learning: Survey"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-2 , February 2020,
URL: https://www.ijtsrd.com/papers/ijtsrd30031.pdf
Paper Url : https://www.ijtsrd.com/engineering/computer-engineering/30031/satellite-image-classification-with-deep-learning-survey/roshni-rajendran
Vehicle detection and classification are essential for advanced driver assistance systems (ADAS) and even traffic camera surveillance. Yet, it is challenging due to complex backgrounds, varying illumination intensities, occlusions, vehicle size, and type variations. This paper aims to apply you only look once (YOLO) since it has been proven to produce high object detection and classification accuracy. There are various versions of YOLO, and their performances differ. An investigation on the detection and classification performance of YOLOv3, YOLOv4, and YOLOv5 has been conducted. The training images were from common objects in context (COCO) and open image, two publicly available datasets. The testing input images were captured on a few highways in two main cities in Malaysia, namely Shah Alam and Kuala Lumpur. These images were captured using a mobile phone camera with different backgrounds during the day and night, representing different illuminations and varying types and sizes of vehicles. The accuracy and speed of detecting and classifying cars, trucks, buses, motorcycles, and bicycles have been evaluated. The experimental results show that YOLOv5 detects vehicles more accurately but slower than its predecessors, namely YOLOv4 and YOLOv3. Future work includes experimenting with newer versions of YOLO.
Vehicle Speed Estimation using Haar Classifier Algorithmijtsrd
An efficient traffic management system is needed in all kinds of roads, such as off roads, highways, etc... Though several laws and speed controller has been attached to the vehicles, Speed limit may vary from road to road. Still Traffic management system faces different kinds of challenges everyday and its being a research area though number of proposals has been identified. Many numbers of methods has been proposed in computer Vision and machine learning approaches for object tracking. In this paper vehicles are identified and detected using a videos that taken from surveillance camera. The objective of the present work is to identification of the vehicles is done by using Computer vision technique and detection of vehicles using Haar cascade classifier. Detecting the vehicles using machine learning and estimating speed is tough but beneficial task. For the past few tears Convolution Neural Network CNN has been widely used in computer vision for vehicle detection and identification. This method manages to track multiple objects at real time using dlibs. P. Devi Mahalakshmi | Dr. M. Babu "Vehicle Speed Estimation using Haar Classifier Algorithm" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-1 , December 2019, URL: https://www.ijtsrd.com/papers/ijtsrd29482.pdfPaper URL: https://www.ijtsrd.com/computer-science/other/29482/vehicle-speed-estimation-using-haar-classifier-algorithm/p-devi-mahalakshmi
Convolutional neural network with binary moth flame optimization for emotion ...IAESIJAI
Electroencephalograph (EEG) signals have the ability of real-time reflecting brain activities. Utilizing the EEG signal for analyzing human emotional states is a common study. The EEG signals of the emotions aren’t distinctive and it is different from one person to another as every one of them has different emotional responses to same stimuli. Which is why, the signals of the EEG are subject dependent and proven to be effective for the subject dependent detection of the Emotions. For the purpose of achieving enhanced accuracy and high true positive rate, the suggested system proposed a binary moth flame optimization (BMFO) algorithm for the process of feature selection and convolutional neural networks (CNNs) for classifications. In this proposal, optimum features are chosen with the use of accuracy as objective function. Ultimately, optimally chosen features are classified after that with the use of a CNN for the purpose of discriminating different emotion states.
A novel ensemble model for detecting fake newsIAESIJAI
Due the growing proliferation of fake news over the past couple of years, our objective in this paper is to propose an ensemble model for the automatic classification of article news as being either real or fake. For this purpose, we opt for a blending technique that combines three models, namely bidirectional long short-term memory (Bi-LSTM), stochastic gradient descent classifier and ridge classifier. The implementation of the proposed model (i.e. BI-LSR) on real world datasets, has shown outstanding results. In fact, it achieved an accuracy score of 99.16%. Accordingly, this ensemble learning has proven to do perform better than individual conventional machine learning and deep learning models as well as many ensemble learning approaches cited in the literature.
More Related Content
Similar to Performance investigation of two-stage detection techniques using traffic light detection dataset
Social distance and face mask detector system exploiting transfer learningIJECEIAES
As time advances, the use of deep learning-based object detection algorithms has also evolved leading to developments of new human-computer interactions, facilitating an exploration of various domains. Considering the automated process of detection, systems suitable for detecting violations are developed. One such applications is the social distancing and face mask detectors to control air-borne diseases. The objective of this research is to deploy transfer learning on object detection models for spotting violations in face masks and physical distance rules in real-time. The common drawbacks of existing models are low accuracy and inability to detect in real-time. The MobileNetV2 object detection model and YOLOv3 model with Euclidean distance measure have been used for detection of face mask and physical distancing. A proactive transfer learning approach is used to perform the functionality of face mask classification on the patterns obtained from the social distance detector model. On implementing the application on various surveillance footage, it was observed that the system could classify masked and unmasked faces and if social distancing was maintained or not with accuracies 99% and 94% respectively. The models exhibited high accuracy on testing and the system can be infused with the existing internet protocol (IP) cameras or surveillance systems for real-time surveillance of face masks and physical distancing rules effectively.
Real Time Object Detection System with YOLO and CNN Models: A ReviewSpringer
The field of artificial intelligence is built on object detection techniques. YOU ONLY LOOK
ONCE (YOLO) algorithm and it's more evolved versions are briefly described in this research survey. This
survey is all about YOLO and convolution neural networks (CNN) in the direction of real time object detection.
YOLO does generalized object representation more effectively without precision losses than other object
detection models. CNN architecture models have the ability to eliminate highlights and identify objects in any
given image. When implemented appropriately, CNN models can address issues like deformity diagnosis,
creating educational or instructive application, etc. This article reached at number of observations and
perspective findings through the analysis. Also it provides support for the focused visual information and
feature extraction in the financial and other industries, highlights the method of target detection and feature
selection, and briefly describes the development process of yolo algorithm
Traffic Congestion Detection Using Deep Learningijtsrd
Despite the huge amount of traffic surveillance videos and images have been accumulated in the daily monitoring, deep learning approaches have been underutilized in the application of traffic intelligent management and control. Traffic images, including various illumination, weather conditions, and vast scenarios are considered and preprocessed to set up a proper training dataset. In order to detect traffic congestion, a network structure is proposed based on residual learning to be pre trained and fine tuned. The network is then transferred to the traffic application and retrained with self established training dataset to generate the Traffic Net. The accuracy of Traffic Net to classify congested and uncongested road states reaches 99 for the validation dataset and 95 for the testing dataset. The trained model is stored in cloud storage for easy access for application from anywhere. The proposed Traffic Net can be used by a regional detection of traffic congestion on a large scale surveillance system. The effectiveness and efficiencies are magnificently demonstrated with quick detection in the high accuracy in the case study. The experimental trial could extend its successful application to traffic surveillance system and has potential enhancement for intelligent transport system in future. Anusha C | Dr. J. Bhuvana "Traffic Congestion Detection Using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-2 , February 2022, URL: https://www.ijtsrd.com/papers/ijtsrd49401.pdf Paper URL: https://www.ijtsrd.com/computer-science/other/49401/traffic-congestion-detection-using-deep-learning/anusha-c
Satellite Image Classification with Deep Learning Surveyijtsrd
Satellite imagery is important for many applications including disaster response, law enforcement and environmental monitoring etc. These applications require the manual identification of objects in the imagery. Because the geographic area to be covered is very large and the analysts available to conduct the searches are few, thus an automation is required. Yet traditional object detection and classification algorithms are too inaccurate and unreliable to solve the problem. Deep learning is a part of broader family of machine learning methods that have shown promise for the automation of such tasks. It has achieved success in image understanding by means that of convolutional neural networks. The problem of object and facility recognition in satellite imagery is considered. The system consists of an ensemble of convolutional neural networks and additional neural networks that integrate satellite metadata with image features. Roshni Rajendran | Liji Samuel ""Satellite Image Classification with Deep Learning: Survey"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-2 , February 2020,
URL: https://www.ijtsrd.com/papers/ijtsrd30031.pdf
Paper Url : https://www.ijtsrd.com/engineering/computer-engineering/30031/satellite-image-classification-with-deep-learning-survey/roshni-rajendran
Vehicle detection and classification are essential for advanced driver assistance systems (ADAS) and even traffic camera surveillance. Yet, it is challenging due to complex backgrounds, varying illumination intensities, occlusions, vehicle size, and type variations. This paper aims to apply you only look once (YOLO) since it has been proven to produce high object detection and classification accuracy. There are various versions of YOLO, and their performances differ. An investigation on the detection and classification performance of YOLOv3, YOLOv4, and YOLOv5 has been conducted. The training images were from common objects in context (COCO) and open image, two publicly available datasets. The testing input images were captured on a few highways in two main cities in Malaysia, namely Shah Alam and Kuala Lumpur. These images were captured using a mobile phone camera with different backgrounds during the day and night, representing different illuminations and varying types and sizes of vehicles. The accuracy and speed of detecting and classifying cars, trucks, buses, motorcycles, and bicycles have been evaluated. The experimental results show that YOLOv5 detects vehicles more accurately but slower than its predecessors, namely YOLOv4 and YOLOv3. Future work includes experimenting with newer versions of YOLO.
Vehicle Speed Estimation using Haar Classifier Algorithmijtsrd
An efficient traffic management system is needed in all kinds of roads, such as off roads, highways, etc... Though several laws and speed controller has been attached to the vehicles, Speed limit may vary from road to road. Still Traffic management system faces different kinds of challenges everyday and its being a research area though number of proposals has been identified. Many numbers of methods has been proposed in computer Vision and machine learning approaches for object tracking. In this paper vehicles are identified and detected using a videos that taken from surveillance camera. The objective of the present work is to identification of the vehicles is done by using Computer vision technique and detection of vehicles using Haar cascade classifier. Detecting the vehicles using machine learning and estimating speed is tough but beneficial task. For the past few tears Convolution Neural Network CNN has been widely used in computer vision for vehicle detection and identification. This method manages to track multiple objects at real time using dlibs. P. Devi Mahalakshmi | Dr. M. Babu "Vehicle Speed Estimation using Haar Classifier Algorithm" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-1 , December 2019, URL: https://www.ijtsrd.com/papers/ijtsrd29482.pdfPaper URL: https://www.ijtsrd.com/computer-science/other/29482/vehicle-speed-estimation-using-haar-classifier-algorithm/p-devi-mahalakshmi
Convolutional neural network with binary moth flame optimization for emotion ...IAESIJAI
Electroencephalograph (EEG) signals have the ability of real-time reflecting brain activities. Utilizing the EEG signal for analyzing human emotional states is a common study. The EEG signals of the emotions aren’t distinctive and it is different from one person to another as every one of them has different emotional responses to same stimuli. Which is why, the signals of the EEG are subject dependent and proven to be effective for the subject dependent detection of the Emotions. For the purpose of achieving enhanced accuracy and high true positive rate, the suggested system proposed a binary moth flame optimization (BMFO) algorithm for the process of feature selection and convolutional neural networks (CNNs) for classifications. In this proposal, optimum features are chosen with the use of accuracy as objective function. Ultimately, optimally chosen features are classified after that with the use of a CNN for the purpose of discriminating different emotion states.
A novel ensemble model for detecting fake newsIAESIJAI
Due the growing proliferation of fake news over the past couple of years, our objective in this paper is to propose an ensemble model for the automatic classification of article news as being either real or fake. For this purpose, we opt for a blending technique that combines three models, namely bidirectional long short-term memory (Bi-LSTM), stochastic gradient descent classifier and ridge classifier. The implementation of the proposed model (i.e. BI-LSR) on real world datasets, has shown outstanding results. In fact, it achieved an accuracy score of 99.16%. Accordingly, this ensemble learning has proven to do perform better than individual conventional machine learning and deep learning models as well as many ensemble learning approaches cited in the literature.
K-centroid convergence clustering identification in one-label per type for di...IAESIJAI
Disease prediction is a high demand field which requires significant support from machine learning (ML) to enhance the result efficiency. The research works on application of K-means clustering supervised classification in disease prediction where each class only has one labeled data. The K-centroid convergence clustering identification (KC3 I) system is based on semi-K-means clustering but only requires single labeled data per class for the training process with the training dataset to update the centroid. The KC3 I model also includes a dictionary box to index all the input centroids before and after the updating process. Each centroid matches with a corresponding label inside this box. After the training process, each time the input features arrive, the trained centroid will put them to its cluster depending on the Euclidean distance, then convert them into the specific class name, which is coherent to that centroid index. Two validation stages were carried out and accomplished the expectation in terms of precision, recall, F1-score, and absolute accuracy. The last part demonstrates the possibility of feature reduction by selecting the most crucial feature with the extra tree classifier method. Total data are fed into the KC3 I system with the most important features and remain the same accuracy.
Plant leaf detection through machine learning based image classification appr...IAESIJAI
Since maize is a staple diet for people, especially vegetarians and vegans, maize leaf disease has a significant influence here on the food industry including maize crop productivity. Therefore, it should be understood that maize quality must be optimal; yet, to do so, maize must be safeguarded from several illnesses. As a result, there is a great demand for such an automated system that can identify the condition early on and take the appropriate action. Early disease identification is crucial, but it also poses a major obstacle. As a result, in this research project, we adopt the fundamental k-nearest neighbor (KNN) model and concentrate on building and developing the enhanced k-nearest neighbor (EKNN) model. EKNN aids in identifying several classes of disease. To gather discriminative, boundary, pattern, and structurally linked information, additional high-quality fine and coarse features are generated. This information is then used in the classification process. The classification algorithm offers high-quality gradient-based features. Additionally, the proposed model is assessed using the Plant-Village dataset, and a comparison with many standard classification models using various metrics is also done.
Backbone search for object detection for applications in intrusion warning sy...IAESIJAI
In this work, we propose a novel backbone search method for object detection for applications in intrusion warning systems. The goal is to find a compact model for use in embedded thermal imaging cameras widely used in intrusion warning systems. The proposed method is based on faster region-based convolutional neural network (Faster R-CNN) because it can detect small objects. Inspired by EfficientNet, the sought-after backbone architecture is obtained by finding the most suitable width scale for the base backbone (ResNet50). The evaluation metrics are mean average precision (mAP), number of parameters, and number of multiply–accumulate operations (MACs). The experimental results showed that the proposed method is effective in building a lightweight neural network for the task of object detection. The obtained model can keep the predefined mAP while minimizing the number of parameters and computational resources. All experiments are executed elaborately on the person detection in intrusion warning systems (PDIWS) dataset.
Deep learning method for lung cancer identification and classificationIAESIJAI
Lung cancer (LC) is calming many lives and is becoming a serious cause of concern. The detection of LC at an early stage assists the chances of recovery. Accuracy of detection of LC at an early stage can be improved with the help of a convolutional neural network (CNN) based deep learning approach. In this paper, we present two methodologies for Lung cancer detection (LCD) applied on Lung image database consortium (LIDC) and image database resource initiative (IDRI) data sets. Classification of these LC images is carried out using support vector machine (SVM), and deep CNN. The CNN is trained with i) multiple batches and ii) single batch for LC image classification as non cancer and cancer image. All these methods are being implemented in MATLAB. The accuracy of classification obtained by SVM is 65%, whereas deep CNN produced detection accuracy of 80% and 100% respectively for multiple and single batch training. The novelty of our experimentation is near 100% classification accuracy obtained by our deep CNN model when tested on 25 Lung computed tomography (CT) test images each of size 512×512 pixels in less than 20 iterations as compared to the research work carried out by other researchers using cropped LC nodule images.
Optically processed Kannada script realization with Siamese neural network modelIAESIJAI
Optical character recognition (OCR) is a technology that allows computers to recognize and extract text from images or scanned documents. It is commonly used to convert printed or handwritten text into machine-readable format. This Study presents an OCR system on Kannada Characters based on siamese neural network (SNN). Here the SNN, a Deep neural network which comprises of two identical convolutional neural network (CNN) compare the script and ranks based on the dissimilarity. When lesser dissimilarity score is identified, prediction is done as character match. In this work the authors use 5 classes of Kannada characters which were initially preprocessed using grey scaling and convert it to pgm format. This is directly input into the Deep convolutional network which is learnt from matching and non-matching image between the CNN with contrastive loss function in Siamese architecture. The Proposed OCR system uses very less time and gives more accurate results as compared to the regular CNN. The model can become a powerful tool for identification, particularly in situations where there is a high degree of variation in writing styles or limited training data is available.
Embedded artificial intelligence system using deep learning and raspberrypi f...IAESIJAI
Melanoma is a kind of skin cancer that originates in melanocytes responsible for producing melanin, it can be a severe and potentially deadly form of cancer because it can metastasize to other regions of the body if not detected and treated early. To facilitate this process, Recently, various computer-assisted low-cost, reliable, and accurate diagnostic systems have been proposed based on artificial intelligence (AI) algorithms, particularly deep learning techniques. This work proposed an innovative and intelligent system that combines the internet of things (IoT) with a Raspberry Pi connected to a camera and a deep learning model based on the deep convolutional neural network (CNN) algorithm for real-time detection and classification of melanoma cancer lesions. The key stages of our model before serializing to the Raspberry Pi: Firstly, the preprocessing part contains data cleaning, data transformation (normalization), and data augmentation to reduce overfitting when training. Then, the deep CNN algorithm is used to extract the features part. Finally, the classification part with applied Sigmoid Activation Function. The experimental results indicate the efficiency of our proposed classification system as we achieved an accuracy rate of 92%, a precision of 91%, a sensitivity of 91%, and an area under the curve- receiver operating characteristics (AUC-ROC) of 0.9133.
Deep learning based biometric authentication using electrocardiogram and irisIAESIJAI
Authentication systems play an important role in wide range of applications. The traditional token certificate and password-based authentication systems are now replaced by biometric authentication systems. Generally, these authentication systems are based on the data obtained from face, iris, electrocardiogram (ECG), fingerprint and palm print. But these types of models are unimodal authentication, which suffer from accuracy and reliability issues. In this regard, multimodal biometric authentication systems have gained huge attention to develop the robust authentication systems. Moreover, the current development in deep learning schemes have proliferated to develop more robust architecture to overcome the issues of tradition machine learning based authentication systems. In this work, we have adopted ECG and iris data and trained the obtained features with the help of hybrid convolutional neural network- long short-term memory (CNN-LSTM) model. In ECG, R peak detection is considered as an important aspect for feature extraction and morphological features are extracted. Similarly, gabor-wavelet, gray level co-occurrence matrix (GLCM), gray level difference matrix (GLDM) and principal component analysis (PCA) based feature extraction methods are applied on iris data. The final feature vector is obtained from MIT-BIH and IIT Delhi Iris dataset which is trained and tested by using CNN-LSTM. The experimental analysis shows that the proposed approach achieves average accuracy, precision, and F1-core as 0.985, 0.962 and 0.975, respectively.
Hybrid channel and spatial attention-UNet for skin lesion segmentationIAESIJAI
Melanoma is a type of skin cancer which has affected many lives globally. The American Cancer Society research has suggested that it a serious type of skin cancer and lead to mortality but it is almost 100% curable if it is detected and treated in its early stages. Currently automated computer vision-based schemes are widely adopted but these systems suffer from poor segmentation accuracy. To overcome these issue, deep learning (DL) has become the promising solution which performs extensive training for pattern learning and provide better classification accuracy. However, skin lesion segmentation is affected due to skin hair, unclear boundaries, pigmentation, and mole. To overcome this issue, we adopt UNet based deep learning scheme and incorporated attention mechanism which considers low level statistics and high-level statistics combined with feedback and skip connection module. This helps to obtain the robust features without neglecting the channel information. Further, we use channel attention, spatial attention modulation to achieve the final segmentation. The proposed DL based scheme is instigated on publically available dataset and experimental investigation shows that the proposed Hybrid Attention UNet approach achieves average performance as 0.9715, 0.9962, 0.9710.
Photoplethysmogram signal reconstruction through integrated compression sensi...IAESIJAI
The transmission of photoplethysmogram (PPG) signals in real-time is extremely challenging and facilitates the use of an internet of things (IoT) environment for healthcare- monitoring. This paper proposes an approach for PPG signal reconstruction through integrated compression sensing and basis function aware shallow learning (CSBSL). Integrated-CSBSL approach for combined compression of PPG signals via multiple channels thereby improving the reconstruction accuracy for the PPG signals essential in healthcare monitoring. An optimal basis function aware shallow learning procedure is employed on PPG signals with prior initialization; this is further fine-tuned by utilizing the knowledge of various other channels, which exploit the further sparsity of the PPG signals. The proposed method for learning combined with PPG signals retains the knowledge of spatial and temporal correlation. The proposed Integrated-CSBSL approach consists of two steps, in the first step the shallow learning based on basis function is carried out through training the PPG signals. The proposed method is evaluated using multichannel PPG signal reconstruction, which potentially benefits clinical applications through PPG monitoring and diagnosis.
Speaker identification under noisy conditions using hybrid convolutional neur...IAESIJAI
Speaker identification is biometrics that classifies or identifies a person from other speakers based on speech characteristics. Recently, deep learning models outperformed conventional machine learning models in speaker identification. Spectrograms of the speech have been used as input in deep learning-based speaker identification using clean speech. However, the performance of speaker identification systems gets degraded under noisy conditions. Cochleograms have shown better results than spectrograms in deep learning-based speaker recognition under noisy and mismatched conditions. Moreover, hybrid convolutional neural network (CNN) and recurrent neural network (RNN) variants have shown better performance than CNN or RNN variants in recent studies. However, there is no attempt conducted to use a hybrid CNN and enhanced RNN variants in speaker identification using cochleogram input to enhance the performance under noisy and mismatched conditions. In this study, a speaker identification using hybrid CNN and the gated recurrent unit (GRU) is proposed for noisy conditions using cochleogram input. VoxCeleb1 audio dataset with real-world noises, white Gaussian noises (WGN) and without additive noises were employed for experiments. The experiment results and the comparison with existing works show that the proposed model performs better than other models in this study and existing works.
Multi-channel microseismic signals classification with convolutional neural n...IAESIJAI
Identifying and classifying microseismic signals is essential to warn of mines’ dangers. Deep learning has replaced traditional methods, but labor-intensive manual identification and varying deep learning outcomes pose challenges. This paper proposes a transfer learning-based convolutional neural network (CNN) method called microseismic signals-convolutional neural network (MS-CNN) to automatically recognize and classify microseismic events and blasts. The model was instructed on a limited sample of data to obtain an optimal weight model for microseismic waveform recognition and classification. A comparative analysis was performed with an existing CNN model and classical image classification models such as AlexNet, GoogLeNet, and ResNet50. The outcomes demonstrate that the MS-CNN model achieved the best recognition and classification effect (99.6% accuracy) in the shortest time (0.31 s to identify 277 images in the test set). Thus, the MS-CNN model can efficiently recognize and classify microseismic events and blasts in practical engineering applications, improving the recognition timeliness of microseismic signals and further enhancing the accuracy of event classification.
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...IAESIJAI
Efficient and accurate coronavirus disease (COVID-19) surveillance necessitates robust identification of individuals wearing face masks. This research introduces the sophisticated face mask dataset (SFMD), a comprehensive compilation of high-quality face mask images enriched with detailed annotations on mask types, fits, and usage patterns. Leveraging cutting-edge deep learning models—EfficientNet-B2, ResNet50, and MobileNet-V2—, we compare SFMD against two established benchmarks: the real-world masked face dataset (RMFD) and the masked face recognition dataset (MFRD). Across all models, SFMD consistently outperforms RMFD and MFRD in key metrics, including accuracy, precision, recall, and F1 score. Additionally, our study demonstrates the dataset's capability to cultivate robust models resilient to intricate scenarios like low-light conditions and facial occlusions due to accessories or facial hair.
Transfer learning for epilepsy detection using spectrogram imagesIAESIJAI
Epilepsy stands out as one of the common neurological diseases. The neural activity of the brain is observed using electroencephalography (EEG). Manual inspection of EEG brain signals is a slow and arduous process, which puts heavy load on neurologists and affects their performance. The aim of this study is to find the best result of classification using the transfer learning model that automatically identify the epileptic and the normal activity, to classify EEG signals by using images of spectrogram which represents the percentage of energy for each coefficient of the continuous wavelet. Dataset includes the EEG signals recorded at monitoring unit of epilepsy used in this study to presents an application of transfer learning by comparing three models Alexnet, visual geometry group (VGG19) and residual neural network ResNet using different combinations with seven different classifiers. This study tested the models and reached a different value of accuracy and other metrics used to judge their performances, and as a result the best combination has been achieved with ResNet combined with support vector machine (SVM) classifier that classified EEG signals with a high success rate using multiple performance metrics such as 97.22% accuracy and 2.78% the value of the error rate.
Deep neural network for lateral control of self-driving cars in urban environ...IAESIJAI
The exponential growth of the automotive industry clearly indicates that self-driving cars are the future of transportation. However, their biggest challenge lies in lateral control, particularly in urban bottlenecking environments, where disturbances and obstacles are abundant. In these situations, the ego vehicle has to follow its own trajectory while rapidly correcting deviation errors without colliding with other nearby vehicles. Various research efforts have focused on developing lateral control approaches, but these methods remain limited in terms of response speed and control accuracy. This paper presents a control strategy using a deep neural network (DNN) controller to effectively keep the car on the centerline of its trajectory and adapt to disturbances arising from deviations or trajectory curvature. The controller focuses on minimizing deviation errors. The Matlab/Simulink software is used for designing and training the DNN. Finally, simulation results confirm that the suggested controller has several advantages in terms of precision, with lateral deviation remaining below 0.65 meters, and rapidity, with a response time of 0.7 seconds, compared to traditional controllers in solving lateral control.
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...IAESIJAI
Recently, cardiovascular diseases (CVDs) have become a rapidly growing problem in the world, especially in developing countries. The latter are facing a lifestyle change that introduces new risk factors for heart disease, that requires a particular and urgent interest. Besides, cardiomegaly is a sign of cardiovascular diseases that refers to various conditions; it is associated with the heart enlargement that can be either transient or permanent depending on certain conditions. Furthermore, cardiomegaly is visible on any imaging test including Chest X-Radiation (X-Ray) images; which are one of the most common tools used by Cardiologists to detect and diagnose many diseases. In this paper, we propose an innovative deep learning (DL) model based on an attention module and MobileNet architecture to recognize Cardiomegaly patients using the popular Chest X-Ray8 dataset. Actually, the attention module captures the spatial relationship between the relevant regions in Chest X-Ray images. The experimental results show that the proposed model achieved interesting results with an accuracy rate of 81% which makes it suitable for detecting cardiomegaly disease.
Efficient commodity price forecasting using long short-term memory modelIAESIJAI
Predicting commodity prices, particularly food prices, is a significant concern for various stakeholders, especially in regions that are highly sensitive to commodity price volatility. Historically, many machine learning models like autoregressive integrated moving average (ARIMA) and support vector machine (SVM) have been suggested to overcome the forecasting task. These models struggle to capture the multifaceted and dynamic factors influencing these prices. Recently, deep learning approaches have demonstrated considerable promise in handling complex forecasting tasks. This paper presents a novel long short-term memory (LSTM) network-based model for commodity price forecasting. The model uses five essential commodities namely bread, meat, milk, oil, and petrol. The proposed model focuses on advanced feature engineering which involves moving averages, price volatility, and past prices. The results reveal that our model outperforms traditional methods as it achieves 0.14, 3.04%, and 98.2% for root mean square error (RMSE), mean absolute percentage error (MAPE), and R-squared (R2 ), respectively. In addition to the simplicity of the model, which consists of an LSTM single-cell architecture that reduced the training time to a few minutes instead of hours. This paper contributes to the economic literature on price prediction using advanced deep learning techniques as well as provides practical implications for managing commodity price instability globally.
1-dimensional convolutional neural networks for predicting sudden cardiacIAESIJAI
Sudden cardiac arrest (SCA) is a serious heart problem that occurs without symptoms or warning. SCA causes high mortality. Therefore, it is important to estimate the incidence of SCA. Current methods for predicting ventricular fibrillation (VF) episodes require monitoring patients over time, resulting in no complications. New technologies, especially machine learning, are gaining popularity due to the benefits they provide. However, most existing systems rely on manual processes, which can lead to inefficiencies in disseminating patient information. On the other hand, existing deep learning methods rely on large data sets that are not publicly available. In this study, we propose a deep learning method based on one-dimensional convolutional neural networks to learn to use discrete fourier transform (DFT) features in raw electrocardiogram (ECG) signals. The results showed that our method was able to accurately predict the onset of SCA with an accuracy of 96% approximately 90 minutes before it occurred. Predictions can save many lives. That is, optimized deep learning models can outperform manual models in analyzing long-term signals.
A deep learning-based approach for early detection of disease in sugarcane pl...IAESIJAI
In many regions of the nation, agriculture serves as the primary industry. The farming environment now faces a number of challenges to farmers. One of the major concerns, and the focus of this research, is disease prediction. A methodology is suggested to automate a process for identifying disease in plant growth and warning farmers in advance so they can take appropriate action. Disease in crop plants has an impact on agricultural production. In this work, a novel DenseNet-support vector machine: explainable artificial intelligence (DNet-SVM: XAI) interpretation that combines a DenseNet with support vector machine (SVM) and local interpretable model-agnostic explanation (LIME) interpretation has been proposed. DNet-SVM: XAI was created by a series of modifications to DenseNet201, including the addition of a support vector machine (SVM) classifier. Prior to using SVM to identify if an image is healthy or un-healthy, images are first feature extracted using a convolution network called DenseNet. In addition to offering a likely explanation for the prediction, the reasoning is carried out utilizing the visual cue produced by the LIME. In light of this, the proposed approach, when paired with its determined interpretability and precision, may successfully assist farmers in the detection of infected plants and recommendation of pesticide for the identified disease.
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Performance investigation of two-stage detection techniques using traffic light detection dataset
1. IAES International Journal of Artificial Intelligence (IJ-AI)
Vol. 12, No. 4, December 2023, pp. 1909~1919
ISSN: 2252-8938, DOI: 10.11591/ijai.v12.i4.pp1909-1919 1909
Journal homepage: http://ijai.iaescore.com
Performance investigation of two-stage detection techniques
using traffic light detection dataset
Sunday Adeola Ajagbe1,4
, Adekanmi Adeyinka Adegun2
, Ahmed Babajide Olanrewaju3
, John Babalola
Oladosu4
, Matthew Olusegun Adigun5
1
Department of Computer Engineering, First Technical University, Ibadan, Nigeria
2
School of Mathematics, Statistics and Computer Science, University of Kwazulu-Natal, South Africa
3
Department of Computer Science, University of Ibadan, Ibadan, Nigeria
4
Department of Computer Engineering, Ladoke Akintola University of Technology LAUTECH, Ogbomoso, Nigeria
5
Department of Information Technology, Cape Peninsula University of Technology, Cape Town, South Africa
Article Info ABSTRACT
Article history:
Received Feb 7, 2023
Revised May 5, 2023
Accepted May 7, 2023
Using a camera to monitor an object or a group of objects over time is the
process of object detection. It can be used for a variety of things, including
security and surveillance, video communication, traffic light detection (TLD),
object detection from compressed video in public places. In recent times,
object tracking has become a popular topic in computer science particularly,
the data science community, thanks to the usage of deep learning (DL) in
artificial intelligence (AI). DL which convolutional neural network (CNN) as
one of its techniques usually used two-stage detection methods in TLD.
Despite all successes recorded in TLD through the use of two-stage detection
methods, there is no study that has analyzed these methods in experimental
research, studying the strength and witnesses by the researchers. Based on the
needs this study analyses the applications of DL techniques in TLD. We
implemented object detection for TLD using 5 two-stage detection methods
with the traffic light dataset using a Jupyter notebook and the sklearn libraries.
We present the achievements of two-stage detection methods in TLD, going
by standard performance metrics used, FASTER-CNN was the best in
detection accuracy, F1-score, precision and recall with 0.89, 0.93, 0.83 and
0.90 respectively.
Keywords:
Artificial intelligence
Convolutional neural network
(CNN)
Deep learning
Faster-CNN
Object detection
Two-stage detection
This is an open access article under the CC BY-SA license.
Corresponding Author:
Sunday Adeola AJAGBE
Department of Computer Engineering, First Technical University, Ibadan
Km. 15, Lagos Express Way, Ibadan, Nigeria
Email: Sunday.ajagbe@tech-u.edu.ng
1. INTRODUCTION
The goal of deep learning (DL), a subfield of machine learning (ML) and artificial intelligence (AI)
is to recreate and imitate in computers the architecture through which the brain conducts vision. DL has many
advantages in image processing and computer vision generally, these advantages include; there is no manual
setting of a learning rate, the automatic setting of a learning rate helps the accuracy of DL-based models, over
gradient descent, minimal computation is referred, suitable for both local and distributed environments and
robust in the face of large gradients, noise, and architecture selection [1]. Using a camera, object tracking
involves following a specific object or set of objects over time. It is an intelligence-based system because a
non-living things object is performing the roles that are meant to perform by living things objects. The
surveillance cameras have a wide range of applications, including security and surveillance, lane detection
(LD), pedestrian detection (PD), traffic light detection (TLD), traffic sign detection (TSD), and vehicle
2. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 4, December 2023: 1909-1919
1910
detection (VD), object detection from complete video (ODFCV), object detection from compressed video
public areas such as airports and subway stations [2], [3]. Computer vision (CV) has been particularly
successful in the areas of object detection and tracking. High-level, intricate abstractions are extracted as data
representations by DL models. Robots, self-driving cars, scene interpretation, and video surveillance are a few
examples of how DL transforms object detection and tracking into an intelligence system [4].
The capacity to recognize and track significant things in near real-time is highly wanted by
manufacturers who are striving toward streamlining manufacturing in today's competitive production
environment. It was almost impossible to keep track of every object in a complicated manufacturing plant by
hand; thus, an automation system with that capacity is desperately needed [5]. In a related study, [6] opined
that sharing the roads with human drivers, autonomous terrestrial vehicles must be able to detect traffic lights
and recognize their current states, which has always been an issue. The majority of the time, human drivers
were able to recognize the appropriate traffic lights. Integration of recognition with past maps has a frequent
technique for autonomous cars to deal with this problem. Although, for the purpose of identifying and detecting
traffic lights, an extra solution is necessary. DL algorithms have demonstrated excellent performance and
generalization power in a variety of issues, including traffic congestion.
Methods for recognizing traffic lights that use onboard sensors in vehicles have been actively
researched [7]. The advancement in image processing methods that have been largely employed to sequence
images acquired by in-vehicle cameras in many manners seems helpful. The successes might be because of
their superior categorization performance, and learning-based approaches have grown in popularity [8].
Meanwhile, due to the presence of many disturbance variables in outdoor contexts, like incomplete light forms,
dark light conditions, and partial occlusion, detection accuracy remains inadequate [9]. These issues are
difficult to solve using computer vision and image processing methods, the difficulties require an investigation.
Although vision-based autonomous driving has shown great promises, there has still the issue of analyzing the
complex traffic scenario using the data collected. Recently, autonomous driving has been broken down into
multiple tasks utilizing various models, such as object detection and intention identification [10].
Although DL models have very powerful object detectors, especially the generic object detector (one-
stage and two-stage detection technique), the effectiveness of the DL-based models in TLD is crucial to the
users (most importantly, the motorist, traffic operators, and pedestrians) for the upcoming vehicle
identification, which is a critical task for self-driving cars. DL-based models are not generalized to new and
unseen traffic light scenarios, including different types of traffic lights and intersections Therefore, analysis of
the usage of DL models in TLD to detect the environment, and appraise the applications of the known DL
models in the TLD projects have not been duly attended to in the academic community. Based on this
requirement, this study has the following objectives; i) Carry out an experimental comparison of convolutional
neural network (CNN) two-stage algorithms for TLD; ii) Evaluation of the two-stage detection methods
implemented in (iii) using standard evaluation metrics (detection accuracy (DA), F1-score, precision and recall
in TLD. The remaining parts of this paper are organized in the following ways: Section 2 is the review of the
related works. In section 3, the traffic light detection algorithm using pre-trained CNN is explained, and in
section 4 analysis of the experimental results and discussion are compiled. The authors conclude the study in
section 5.
2. RELATED STUDIES
Salari et al. (2022) [11] provided a detailed analysis of datasets in the highly researched areas of object
recognition. About 160 datasets were examined using statistics and descriptions. In addition, the paper provided
a discussion of the metrics frequently used for evaluation in the CV community, together with an overview of
the most well-known object recognition benchmarks and contests. Li et al. (2020) [12] enumerated detection
tracking and classification of the traffic light that was implemented in DL. The traffic jam affects business
activities as well as other social lives of people living in the urban centers. It also affects the role of social
workers and security operatives in saving lives and properties. These issues and other related issues informed
the TLD system that is priority-based [13]. It's become essential as numerous government and private
organizations gather enormous amounts of domain-specific data, which can offer insightful data on topics like
marketing, national intelligence, cyber security, and fraud detection.
Ornek et al. (2021) [14] demonstrated the application of the deep neural network (DNN) method,
(a DL-based method) in face images in order to classify them into the mask and non-mask classes using the
last convolutional layers of DNN. The ResNet-18 DNN was chosen, and the technique was trained on 18,600
balanced facial images from two classes and tested with 4540 non-training face images. When the study was
asked why an image was classified into the mask class, it responded by indicating the location where the mask
image was present. This was another intelligence-based system showcasing object detection means and DL
technique. For counting persons in films, In´acio et al. (2021) [15] presented a DL method that paid particular
3. Int J Artif Intell ISSN: 2252-8938
Performance investigation of two-stage detection techniques using traffic … (Sunday Adeola Ajagbe)
1911
attention to gender and age information. Counting individuals in movies is an important task with applications
in surveillance, commerce, health care, and many other areas. Gender and age were extracted from faces found
in films, the system created uses customized Deep Neural Networks (DNN). The study project makes an effort
to manage occlusions, prevent double-counting, and lessen the detrimental effects of background information.
The hardest part of the work was identifying and tracking faces in angles and handling occlusions.
With their work, Zhang et al. (2019) [16] enhanced the functionality of the selective refinement
network (SRN) algorithm. The research project combined existing methods, such as a decoupled classification
module, to create an SRN face detection. On the benchmark WIDER FACE dataset for face detection, the
system performed superbly. Based on the three assessment measures utilized, the resulting approach performed
admirably, particularly on the hard subset that contains a sizable number of small faces. This was another
landmark achievement of DL-based application, it gave further confidence in to TLD, an innovation that aids
smart city development. Also, high-performance face identification, according to [17], [18] was a difficult
problem to solve, especially when there were many little faces. The researchers found that the current face
detection algorithms prioritize a high recall rate while neglecting the problem of too many false positives. They
think there is room for improvement in recall effectiveness and accuracy of location, particularly in terms of
lowering the proportion of false positives at high recall rates and optimizing the position of the bounding box.
To reduce false positives and increase the accuracy of the location at the same time, they suggested a
revolutionary single-shot face detection termed an SRN. To generate a more diverse receptive field and aid in
data collection in extreme positions, a model with a receptive field improvement was designed.
Another line of study that was similar was done by [19] was the creation of the repulsion loss for
pedestrian detection, which significantly improved detection performance, particularly in circumstances with
crowds. The underlying justification for the repulsion loss was that repulsion-by-surrounding can be incredibly
helpful and attraction-by-target loss alone may not be sufficient to train an optimal detection. To implement
the repulsion energy, they offered two repLoss forms. On the two large datasets, City Persons and Caltech,
they have the highest reported performance. Sun et al. (2018) [20] offered a new DL-based face detection
strategy that achieved an improved detection performance through a well-known FDDB face detection
benchmark evaluation. To enhance the Faster Region-based Convolutional Neural Network (RCNN)
architecture, they integrated a number of strategies, including meticulous calibration of crucial parameters. The
WIDER FACE dataset was initially used to train the RCNN algorithm of CNN. The pre-trained model was
later put to the test using the data set in order to provide hard negatives. In the following round of training, the
hard negatives were input into the network to produce fewer false positives. The CNN network's backbone,
VGG-16, served as the installation and training platform for the DL approach. The encouraging results
demonstrate the efficacy of the suggested DL-based system for face detection as an intelligence-based system.
The method of zero watermarking was used by [21] to authenticate medical images. The image is
transformed into four bands using the Discrete Wavelet Transform method. Following the discrete wavelet
transform (DWT), each block was handled with SVD to remove selected features, and a master share was
created by conducting a logical operation between the extracted unique feature and the watermark. The
technique produces superior experimental findings and can be used in the field of telemedicine to safeguard
shared photos from unauthorized use. It was another breakthrough in the medical field.
Zhou et al. (2017) [22] demonstrates the importance of DL technology applications and the influence
of dataset for DL technique through the use of the quicker R-CNN on new image datasets of a football game,
which identifies four (4) categories of objects, and the corner flag. Experiments indicate that DL technology
was an efficient method for moving a human-made feature that relied on the drive of experience to learning
that relied on the drive of data. The challenge was in the quality of the data fed, they recommend the use of
some synthetic data in order to increase the amount of data in future research. Huge data is the foundation for
DL’s success, just as large data was the fuel for DL's rocket. A DNN fusion architecture for fast and reliable
pedestrian detection was introduced by [23]. For speed, the proposed network fusion algorithm allows many
networks to be processed simultaneously. As an object detector, a single shot Deep convolutional neural
network (DCNN) was all potential pedestrian candidates through training of various sizes and occlusions. Also,
for further improvement of these pedestrian candidates, multiple DNNs were utilized in tandem. Additionally,
as a reinforcement to pedestrian detection, they present a method for incorporating a pixel-by-pixel network
fusion architecture that incorporates the semantic segmentation network. The technique performs better than
most of the existing techniques while also outperformed other in terms of speed.
Angelova et al. (2015) [24] designed an intelligence-based system for real-time pedestrian detection
architecture using a cascade of DCNN. The tiny model was used to reject a huge DNN was used to classify the
hard proposals of a large number of easy negatives were identified. Girshick (2017) [25] introduced an object
detection models named the Fast RCNN. The Fast R-CNN improved on the advantages and disadvantaged of
Spatial pyramid pooling networks (SPPnet) and Region-based Convolutional R-CNN. It efficiently classifies
objects with improved speed and accuracy using DCNNs.
4. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 4, December 2023: 1909-1919
1912
Sermanet et al., (2013) [26] proposed another learning model for pedestrian detection, an area in traffic
zones. Unlike common ways where the low-level features were manually designed, the proposed technique
learns all features at each and every level of a hierarchy. By combining high- and low-resolution features in
the methods and learning features from the input's color channels, they improved the convolutional filter bank
method, which they had previously used. Using the INRIA dataset, the method demonstrated that these
improvements offer obvious performance advantages. On most measures of all publicly accessible datasets,
the generated framework offers a competitive result. Karim et al. (2019) [7] suggested Mask RCNN technique
for the segmentation and zones in plants, the technique was trained via transfer learning, which started with a
neural network (NN) that had been pre-trained using the Microsoft Common Objects in Context (MS-COCO)
dataset then fine-tuned it with a small number of annotated photos images. The Mask RCNN technique was
then tweaked to produce consistent video detection outcomes, which was accomplished by employing a two-
staged detection threshold and analyzing the temporal coherence information of discovered items. The system
now includes an object tracking capability for detecting object misplacement. Analyzing a sample of video
footages confirmed the usefulness and efficiency of the suggested approach. This has not been employed or
experimented using TL dataset.
Ajagbe et al. (2021) [27] implemented multi-classification of alzheimer disease using magnetic
resonance images (MRI) and DCNN techniques, the research particularly used CNN and transfer learning
techniques Visual Geometry Group (VGG-16 and VGG-19). The VGG-19 outperformed other techniques used
in the study based on the results obtained from the four (4) different evaluation metrics in the research. It is
observed from the reviewed literature that there are limited studies that accounted for the applications of DL
in object detection with specific attention to TLD. Studies that outlined that of DL-based models that is efficient
in TLD are rarely found in scholarly databases. Research investigating TDL based on CNN two-stage detection
methods is limited to the best of researchers’ knowledge and based on the requirement and contribution of this
method. Hence, this study was informed.
3. METHOD
This section reports our DL experimental approach towards TLD on Representative two-stage
detectors, a DL framework. Representative two-stage detection technique comprising three CNN methods,
namely; FASTER-CNN (FASTER-CNN), FAST- CNN (FAST-CNN), R-CNN, Region-based fully
convolutional network (R-FCN) and Spatial Pyramid Pooling networks (SPPNet). Basically, the experimental
approach in this chapter entails: 1) TLD data collection, 2) pre-processing of the dataset, 3) Data partitioning,
4) Feature section 5) DL-based TLD Analysis, and 6) Evaluation. Figure 1 presents the architecture for this
study named DL-based (two-stage detection methods) analysis using LaRA Traffic Lights Recognition.
Algorithm 1 was proposed for the investigation of two-stage detectors in a TLD environment and thus
presented.
Algorithm 1: Investigation of Two-stage Detectors in TLD Algorithm
Input: LaRA TL dataset; Training and testing, testing; 0
Output: TLD result with forecast and accomplishment, DA, F1-score, precision, recall and RT in TLD.
1. Dataset = LaRA TL dataset
2. Train = Train FASTER-CNN, FAST-CNN, R-CNN, R-FCN and SPPNet using LaRA dataset
3. analyzeNetwork (FaceNet)
4. TrainingOption:
OptimizationAlgorithm = rmsprop
InitialLearningRate = 0.00002
MaxEpochs = 25
MiniBatchSize = 32,
5. Load Dataset
6. Resize Dataset to [227,227, 3]
7. Dataset Split [TrainingData (80%), TestingData (20%)]
8. Train Load Dataset
9. Test Trained models/* using TestingData */
10. Return TestResults
11. TestResults = Validation DA, F1-score, precision and recall
15. If TestResults = Satisfactory, then
(1) Save the Trained models and TestResults /* for transfer
learning purpose */
(2) End
16. Else
17. Adjust the TrainingOption, then
18. Repeat the process
5. Int J Artif Intell ISSN: 2252-8938
Performance investigation of two-stage detection techniques using traffic … (Sunday Adeola Ajagbe)
1913
Figure 1. Architecture of two-stage detection technique based on TLD
3.1. Data collection
The pre-trained dataset used for a benchmark of the experiment was acquired online and used by many
scholar including [28]. The LaRA traffic lights recognition public benchmarks [28] video data was collected.
There are 11,179 frames in total, with a resolution of 640*480. This dataset includes a sequence file and a
ground truth file containing 9,168 instances of traffic lights classified into four categories (ambiguous, green,
red, and orange). Labelled traffic lights in LaRA are 5 pixels larger.
3.1.1. Pre-processing of TLD dataset
The TLD dataset was automatically resized, cropped and unwanted noises were removed from the
background of the data. The data pre-processing was done using the instrumentality of the python programming
environment. The data were also flattened, and converted to greyscale. This was done to gain better and good
insight of the dataset for analysis and better performance of the models.
3.1.2. Selection of features
After the preprocessing of the TLD dataset that terminated at the conversion of the images to grayscale
images, suitable features were selected for the experiment. The feature vectors were partitioned into a 20: 80
percent test-train split ratio of the datasets. This was done to gain better and good insight of the dataset for
analysis and better performance of the models. All the irrelevant and unwanted features were removed.
3.1.3. Representative two-stage detectors
The generic object detector is a CNN framework that uses for object mainstreaming using a DL-based
approach, they are typically divided into two categories; one-stage and two-stage methods. While FASTER-
CNN, FAST-CNN, R-CNN, R-FCN and SPPNet are examples of two-stage detectors, you only look once
(YOLO), Single-Shot object Detection (SSD), YOLOv2 and YOLOv3 are popular examples of one-stage.
Detection in two-stage method is based on region proposals. Proposals in R-CNN and Fast R-CNN were
generated from images selected from the original image, whereas proposals in Faster R-CNN are also generated
directly from feature maps. Classification and bounding box regression was applied in two stages to each region
proposal to achieve Faster R-CNN. In this section, the typical two-stage detection series are introduced in
detail, focusing on network structure and performance. Other representative two-stage detectors that will be
discussed and used include spatial pyramid pooling [29] and region-based fully convolutional network [30].
R-CNN: Girshick et al. (2014) [31] developed R-CNN as a breakthrough study in the development
of applying DL-based methods to detection. In contrast to traditional methods that use a sliding window, R-
CNN obtains proposals using the selective search (SS) technique [32]. CNN and support vector machine
(SVM) were responsible for feature extraction and classification, respectively. Such a significant improvement
demonstrated the significant benefit of the region proposal with CNN. Meanwhile, this ground-breaking
framework was not without flaws [25] (1) multi-step training with time-consuming steps; and (2) slow object
detection. SPPNet provided a good solution to the second problem.
6. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 4, December 2023: 1909-1919
1914
SPPNet: In R-CNN, there were 2k proposals, and each proposal was fed to a CNN and an SVM
classifier separately. The feature extraction and detection were repeated 2,000 times for each image, which was
the main cause of the slow detection. 2,000 proposed were also generated in SPPNet [30]. Meanwhile, CNN
was used to extract the entire feature map of the original image, and the feature map of each proposal is
generated by mapping. Consequently, feature maps are only computed once. Although the input length for the
FC layer is fixed, the feature maps for the various approaches are not all the same size. An SPP [33] layer was
added before the FC layers. In conclusion, using SPPNet greatly accelerates object detection when compared
to R-CNN, but issues such as multi-stage training remain [25]. Furthermore, when fine-tuning the network, the
CONV layers are fixed, and only the FC layer is fine-tuned, on the other hand, very deep networks'
expressiveness was constrained. Fast R-CNN, was an efficient and quick object detection method that Girshick
(2015) [25] presented to address these two issues.
Fast R-CNN: The feature extraction sections of various proposals are also shared. Pooling by RoI
was used to generate proposals of the same size. As previously stated, the training phase of SPPNet and R-
CNN are a multi-stage pipeline. Softmax classifier is used instead of SVM to perform classification in Fast R-
CNN. Furthermore, the loss function, which includes regression task loss, makes the entire training process
end-to-end. To summarize, Fast R-CNN simplifies the training and testing of the entire network. However, SS
for region extraction is time-consuming, accounting for a significant portion of the running time, and cannot
meet the requirements of a real-time application. Because SS was used to extract region proposals first, in fact,
Fast R-CNN does not achieve true end-to-end [34].
Faster R-CNN: It requires only 0.32 s to feed-forward an image; the bottleneck is that generating
proposals takes 2 s. Ren et al. (2015) [35] proposed a faster R-CNN that replaced SS with RPN. In comparison
to SS, which slides on the original image, RPN employs a similar sliding policy but on smaller feature maps,
resulting in fewer proposals. End-to-end training was achieved by combining RPN with the entire network,
and the test time for Faster R-CNN was 0.2 s, which was close to real-time.
R—FCN: After RoI pooling, quicker for each proposal, R-CNN must carry out two-branch prediction
independently. Long et al. 2015 [36] designed R-FCN to allow almost all computing to be shared in order to
increase speed in the network. Following the shared CONV layers, the position-sensitive score maps are
generated using one more CONV layer. When performing RoI pooling, each proposal was divided into 3 * 3
grids, and R-FCN assigns each grid to a different channel of the score maps. To generate a final feature map,
average pooling is applied to each grid.
In summary, R-CNN, SPPNet, Fast R-CNN, Faster R-CNN, and R-FCN methods evolving increase
the amount of image-level calculations in a network over time while decreasing the proportion of regional-
level calculations. Almost all R-CNN calculations were at the regional level, whereas almost all R-FCN
calculations were at the image level. Therefore, this research is extending the frontier of research on the DL-
based CNN two-stage detection methods to be studied using the TLD dataset.
3.1.4. Implementation environment
The implementation environment for DL-based analysis of the TLD was carried out on the Anaconda
platform using a Jupyter notebook and the sklearn libraries [37], [38]. All experiments were carried out on an
Intel® coreTM i5-7200 Pentium Windows computer with 8GB RAM and an Intel® coreTM i5-7200 CPU
running at 2.50 GHz to 2.70 GHz. In this section, an analysis of DL works in TLD was implemented according
to Representative two-stage detection methods. Unlike the one-stage detection methods, these methods are
advanced and evolving relatively used in object detection projects. It also allows direct use of the framework
with or without changes.
3.1.5. Evaluation method
The analysis of TLD in this chapter was based on five metrics, they are detection accuracy (DA), F1-
score, precision and recall. The five (CNN) two-stage detection methods used for TLD analysis were;
FASTER-CNN, FAST-CNN, R-CNN, R-FCN and SPPNet are based on the five metrics. The DL-based TLD
analysis in this research is experimentally based, the four (4) distinct state-of-the-art evaluation metrics have
been selected to examine the strength and weaknesses of these CNN methods techniques. The metrics of
evaluation in the two-stage detection DL-based methods for TLD investigation presents in this sub-section:
𝑫𝒆𝒕𝒆𝒄𝒕𝒊𝒐𝒏 𝑨𝒄𝒄𝒖𝒓𝒂𝒄𝒚 =
𝑻𝒑+𝑻𝒏
𝑻𝒑+𝑻𝒏+𝑭𝒑+𝑭𝒏
(1)
𝑷𝒓𝒆𝒄𝒊𝒔𝒊𝒐𝒏 =
𝑻𝒑
𝑻𝒑+𝑭𝒑
(2)
7. Int J Artif Intell ISSN: 2252-8938
Performance investigation of two-stage detection techniques using traffic … (Sunday Adeola Ajagbe)
1915
𝑹𝒆𝒄𝒂𝒍𝒍 =
𝑻𝒑
𝑻𝒑+𝑭𝒏
(3)
𝑭𝟏 − 𝑺𝒄𝒐𝒓𝒆 = 𝟐 ×
𝑷𝒓𝒆𝒄𝒊𝒔𝒊𝒐𝒏×𝑹𝒆𝒄𝒂𝒍𝒍
𝑷𝒓𝒆𝒄𝒊𝒔𝒊𝒐𝒏+𝑹𝒆𝒄𝒂𝒍𝒍
(4)
4. RESULTS AND DISCUSSION
After a successful experimental investigation of two-stage detectors models LaRA dataset (traffic light
dataset). The results and discussion of the experimental analysis of TLD in this study are presented in this
section. The section presents the detail analysis of the five standards metrics for the evaluation of TLD in this
study viz-a-viz; detection accuracy (DA), F1-score, precision and recall. The DL-based two-stage detection
methods evaluated and discussed are FASTER-CNN, FAST-CNN, R-CNN, R-FCN and SPPNet.
4.1. Result
The result of the evaluation was based on the state-of-the-art performance metric in DL and AI
generally, the metrics are detection accuracy (DA), precision, F1-score and recall [39], all the experimental
parameters were kept on default for the five methods used. The suitable and pre-trained dataset meant for TLD
from LaRA was used [28]. As discussed earlier in Sections 3, the comparison of the five different two-stage
TLD detection algorithms' outcomes on the LaRA dataset were presented. The values for all other performance
metrics range from between 0 and 1. The closer the value of the metrics to 1, the better the method. Table 1
summarizes the Performance Analysis of the two-stage detection methods on the TLD Dataset in this research.
Table 1. Performance Analysis of the two-stage detection methods on TLD Dataset
Methods R-CNN SPPNet Fast-CNN Faster-CNN R-FCN
Detection Accuracy (DA) 0.86 0.73 0.88 0.89 0.79
F1-Score 0.88 0.84 0.89 0.93 0.85
Precision 0.79 0.75 0.83 0.83 0.79
Recall 0.81 0.75 0.84 0.90 0.81
4.2. Discussion
4.2.1. Detection accuracy
Figure 2 presents the detection accuracy (DA) of the two-stage detection methods using the TLD
dataset in this research; FASTER-CNN, FAST-CNN, R-CNN, R-FCN and SPPNet with 0.89, 0.88, 0.86, 0.79
and 0.73 respectively. Since the method that has a rate closer to 1 is the best. Hence, FASTER-CNN is better
than the others.
Figure 2. Two-stage TLD detection accuracy
4.2.2. F1-score
Figure 3 presents the F1-score of the two-stage detection methods using the TLD dataset in this
research; FASTER-CNN, FAST-CNN, R-CNN, R-FCN and SPPNet with 0.93, 0.89, 0.88, 0.85 and 0.84
respectively. Since the method that has a rate closer to 1 is the best. Hence, FASTER-CNN is better than the
others.
8. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 4, December 2023: 1909-1919
1916
Figure 3. Two-stage TLD F1-score
4.2.3. Precision
Figure 4 presents the precision of the two-stage detection methods using the TLD dataset in this
research; FASTER-CNN, FAST-CNN, R-CNN, R-FCN and SPPNet with 0.83, 0.83, 0.79, 0.79 and 0.75
respectively. Since the method that has a rate closer to 1 is the best. Hence, FASTER-CNN and FAST-CNN
are better than the others.
Figure 4. Two-stage TLD precision
4.2.4. Recall
Figure 5 presents the recall report of the two-stage detection methods using the TLD dataset in this
research; FASTER-CNN, FAST-CNN, R-CNN, R-FCN and SPPNet with 0.9, 0.84, 0.81, 0.81 and 0.75
respectively. Since the method that has a rate closer to 1 is the best. Hence, FASTER-CNN is better than the
others.
Figure 5. Two-stage TLD recall
9. Int J Artif Intell ISSN: 2252-8938
Performance investigation of two-stage detection techniques using traffic … (Sunday Adeola Ajagbe)
1917
Summarily, among the four standard and state-of-the-art performance metrics used for the evaluation
of the five distinct two-stage detection methods that have been judged to be more popular in object detection
and tracking, FASTER-CNN performed best in all. This was followed by FAST-CNN which was second in
four metrics and alongside FASTER-CNN, FAST-CNNout performed others techniques in precision.
5. CONCLUSION
The DL-based method is a hot research field, its roles in object detection particularly TLD booming
due to its cutting-edge solutions in autonomous driving and implementation of the smart city. The powerful
CNN is one of the DL-based methods that made DL-based research more popular. This chapter has
exhaustively reviewed object detection applications viz-a-viz architecture, techniques, and DL-CNN in TLD.
We further implemented object detection using DL/CNN two-stage detection methods in a TLD environment.
FASTER-CNN was the best method based on the standard evaluation metrics used. This chapter addresses
some of the concerns in TLD using DL methods which is one of the popular AI methods in TLD, we also
classified detection according to different applications works such as pedestrian detection, lane detection and
so forth. To the extent that we are aware, this is the first experimental study that analyzed DL-based models
focusing on two-stage detection in the TLD environment. Considering the performance improvement of two-
stage detection methods in this study, this will also bring advantages of DL/CNN in TLD to both researchers
and users.
ACKNOWLEDGEMENTS
Authors acknowledge the support of the Cape Peninsula University of Technology, South Africa and
the First Technical University, Ibadan, Nigeria for their support on this project.
REFERENCES
[1] S. Ojha and S. Sakhare, “Image processing techniques for object tracking in video surveillance- A survey,” Jan. 2015.
doi: 10.1109/PERVASIVE.2015.7087180.
[2] R. Verschae and J. Ruiz-del-Solar, “Object detection: Current and future directions,” Frontiers Robotics AI, vol. 2, no. NOV, Nov.
2015, doi: 10.3389/frobt.2015.00029.
[3] S. A. Ajagbe and I. Bamimore, “Design and implementation of smart home nodes for security using radio frequency modules,”
International Journal of Digital Signals and Smart Systems, vol. 4, no. 4, p. 286, 2020, doi: 10.1504/ijdsss.2020.10032471.
[4] T. P. Olalere and O. D. Adeniji, “An artificial intelligent video assistant invigilator to curb examination malpractice,” 2021.
[5] J. Davis and M. Goadrich, “The relationship between precision-recall and ROC curves,” in ACM International Conference
Proceeding Series, 2006, vol. 148, pp. 233–240. doi: 10.1145/1143844.1143874.
[6] M. Enzweiler and D. M. Gavrila, “Monocular pedestrian detection: Survey and experiments,” IEEE Transactions on Pattern
Analysis and Machine Intelligence, vol. 31, no. 12, pp. 2179–2195, Dec. 2009, doi: 10.1109/TPAMI.2008.260.
[7] M. M. Karim, D. Doell, R. Lingard, Z. Yin, M. C. Leu, and R. Qin, “A region-based deep learning algorithm for detecting and
tracking objects in manufacturing plants,” Procedia Manufacturing, vol. 39, pp. 168–177, 2019, doi: 10.1016/j.promfg.2020.01.289.
[8] L. C. Possatti et al., “Traffic light recognition using deep learning and prior maps for autonomous cars,” in Proceedings of the
International Joint Conference on Neural Networks, Jul. 2019, vol. 2019-July. doi: 10.1109/IJCNN.2019.8851927.
[9] M. B. Jensen, M. P. Philipsen, A. Møgelmose, T. B. Moeslund, and M. M. Trivedi, “Vision for looking at traffic lights: issues,
survey, and perspectives,” IEEE Transactions on Intelligent Transportation Systems, vol. 17, no. 7, pp. 1800–1815, Jul. 2016, doi:
10.1109/TITS.2015.2509509.
[10] V. Ramanishka, Y. T. Chen, T. Misu, and K. Saenko, “Toward driving scene understanding: a dataset for learning driver behavior
and causal reasoning,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition,
Jun. 2018, pp. 7699–7707. doi: 10.1109/CVPR.2018.00803.
[11] A. Salari, A. Djavadifar, X. Liu, and H. Najjaran, “Object recognition datasets and challenges: A review,” Neurocomputing,
vol. 495, pp. 129–152, Jul. 2022, doi: 10.1016/j.neucom.2022.01.022.
[12] Y. Li et al., “A deep learning-based hybrid framework for object detection and recognition in autonomous driving,” IEEE Access,
vol. 8, pp. 194228–194239, 2020, doi: 10.1109/ACCESS.2020.3033289.
[13] A. Mousavian, D. Anguelov, J. Košecká, and J. Flynn, “3D bounding box estimation using deep learning and geometry,” in
Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Jul. 2017, vol. 2017-January,
pp. 5632–5640. doi: 10.1109/CVPR.2017.597.
[14] A. H. Ornek, M. Celik, H. T. R. Center, and M. Ceylan, “Explainable artificial intelligence: How face masks are detected via deep
neural networks,” International Journal of Innovative Science and Research Technology, vol. 6, no. 9, pp. 1104–1112, 2021.
[15] A. de S. Inácio, R. H. Ramos, and H. S. Lopes, “Deep learning for people counting in videos by age and gender,” in Anais do 15.
Congresso Brasileiro de Inteligência Computacional, Jan. 2021, pp. 1–6. doi: 10.21528/cbic2021-53.
[16] S. Zhang et al., “Improved selective refinement network for face detection,” Jan. 2019, Accessed: May 24, 2023. [Online].
Available: http://arxiv.org/abs/1901.06651
[17] C. Chi, S. Zhang, J. Xing, Z. Lei, S. Z. Li, and X. Zou, “Selective refinement network for high performance face detection,” 33rd
AAAI Conference on Artificial Intelligence, AAAI 2019, 31st Innovative Applications of Artificial Intelligence Conference, IAAI
2019 and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, vol. 33, no. 01, pp. 8231–8238,
Jul. 2019, doi: 10.1609/aaai.v33i01.33018231.
[18] J. B. Awotunde et al., “An improved machine learnings diagnosis technique for COVID-19 pandemic using chest x-ray images,”
in Communications in Computer and Information Science, vol. 1455 CCIS, Springer International Publishing, 2021, pp. 319–330.
doi: 10.1007/978-3-030-89654-6_23.
10. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 4, December 2023: 1909-1919
1918
[19] X. Wang, T. Xiao, Y. Jiang, S. Shao, J. Sun, and C. Shen, “Repulsion loss: Detecting pedestrians in a crowd,” in Proceedings of
the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2018, pp. 7774–7783.
doi: 10.1109/CVPR.2018.00811.
[20] X. Sun, P. Wu, and S. C. H. Hoi, “Face detection using deep learning: An improved faster RCNN approach,” Neurocomputing,
vol. 299, pp. 42–50, Jul. 2018, doi: 10.1016/j.neucom.2018.03.030.
[21] S. Sinha, A. Singh, R. Gupta, and S. Singh, “Authentication and tamper detection in tele-medicine using zero watermarking,”
Procedia Computer Science, vol. 132, pp. 557–562, 2018, doi: 10.1016/j.procs.2018.05.009.
[22] X. Zhou, W. Gong, W. Fu, and F. Du, “Application of deep learning in object detection,” in Proceedings - 16th IEEE/ACIS
International Conference on Computer and Information Science, ICIS 2017, May 2017, pp. 631–634.
doi: 10.1109/ICIS.2017.7960069.
[23] X. Du, M. El-Khamy, J. Lee, and L. Davis, “Fused DNN: A deep neural network fusion approach to fast and robust pedestrian
detection,” in Proceedings - 2017 IEEE Winter Conference on Applications of Computer Vision, WACV 2017, Mar. 2017,
pp. 953–961. doi: 10.1109/WACV.2017.111.
[24] A. Angelova, A. Krizhevsky, V. Vanhoucke, A. Ogale, and D. Ferguson, “Real-time pedestrian detection with deep network
cascades,” in Procedings of the British Machine Vision Conference 2015, 2015, pp. 32.1-32.12. doi: 10.5244/c.29.32.
[25] R. Girshick, “Fast R-CNN,” in Proceedings of the IEEE International Conference on Computer Vision, Dec. 2015, vol. 2015
International Conference on Computer Vision, ICCV 2015, pp. 1440–1448. doi: 10.1109/ICCV.2015.169.
[26] P. Sermanet, K. Kavukcuoglu, S. Chintala, and Y. Lecun, “Pedestrian detection with unsupervised multi-stage feature learning,” in
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2013, pp. 3626–3633.
doi: 10.1109/CVPR.2013.465.
[27] S. A. Ajagbe, K. A. Amuda, M. A. Oladipupo, O. F. AFE, and K. I. Okesola, “Multi-classification of alzheimer disease on magnetic
resonance images (MRI) using deep convolutional neural network (DCNN) approaches,” International Journal of Advanced
Computer Research, vol. 11, no. 53, pp. 51–60, Mar. 2021, doi: 10.19101/ijacr.2021.1152001.
[28] Q. Wang, Q. Zhang, X. Liang, Y. Wang, C. Zhou, and V. I. Mikulovich, “Traffic lights detection and recognition method based on
the improved yolov4 algorithm,” Sensors, vol. 22, no. 1, p. 200, Dec. 2022, doi: 10.3390/s22010200.
[29] J. Redmon and A. Farhadi, “YOLO9000: Better, faster, stronger,” in Proceedings - 30th IEEE Conference on Computer Vision and
Pattern Recognition, CVPR 2017, Jul. 2017, vol. 2017-Janua, pp. 6517–6525. doi: 10.1109/CVPR.2017.690.
[30] K. He, X. Zhang, S. Ren, and J. Sun, “Spatial pyramid pooling in deep convolutional networks for visual recognition,” IEEE
Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no. 9, pp. 1904–1916, Sep. 2015,
doi: 10.1109/TPAMI.2015.2389824.
[31] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic
segmentation,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2014,
pp. 580–587. doi: 10.1109/CVPR.2014.81.
[32] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Region-based convolutional networks for accurate object detection and
segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 1, pp. 142–158, Jan. 2016,
doi: 10.1109/TPAMI.2015.2437384.
[33] S. Lazebnik, C. Schmid, and J. Ponce, “Beyond bags of features: Spatial pyramid matching for recognizing natural scene
categories,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2006, vol. 2,
pp. 2169–2178. doi: 10.1109/CVPR.2006.68.
[34] A. Boukerche and Z. Hou, “Object detection using deep learning methods in traffic scenarios,” ACM Computing Surveys, vol. 54,
no. 2, pp. 1–35, Mar. 2021, doi: 10.1145/3434398.
[35] S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal networks,” IEEE
Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, pp. 1137–1149, Jun. 2017,
doi: 10.1109/TPAMI.2016.2577031.
[36] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE
Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2015, vol. 07-12-June-2015, pp. 431–440.
doi: 10.1109/CVPR.2015.7298965.
[37] R. Rajmohan et al., “G-Sep: A deep learning algorithm for detection of long-term sepsis using bidirectional gated recurrent unit,”
International Journal of Uncertainty, Fuzziness and Knowlege-Based Systems, vol. 30, no. Supp01, pp. 1–29, May 2022,
doi: 10.1142/S0218488522400013.
[38] S. Devadharshini, R. Kalaipriya, R. Rajmohan, M. Pavithra, and T. Ananthkumar, “Performance investigation of hybrid YOLO-
VGG16 based ship detection framework using SAR images,” Jul. 2020. doi: 10.1109/ICSCAN49426.2020.9262440.
[39] S. Patel and A. Patel, “Deep leaning architectures and its applications a survey,” International Journal of Computer Sciences and
Engineering, vol. 6, no. 6, pp. 1177–1183, Jun. 2018, doi: 10.26438/ijcse/v6i6.11771183.
BIOGRAPHIES OF AUTHORS
Sunday Adeola Ajagbe is a Ph.D. candidate at the Department of Computer
Engineering, Ladoke Akintola University of Technology (LAUTECH), Ogbomoso,
Nigeria and a Lecturer, a First Technical University, Ibadan, Nigeria. He obtained MSc and
BSc in Information Technology and Communication Technology respectively at the
National Open University of Nigeria (NOUN). His specialization includes Artificial
Intelligence (AI), Natural language processing (NLP), Information Security,
Communication, and Internet of Things (IoT). He has many publications to his credit in
reputable academic databases. He can be contacted at email: Sunday.ajagbe@tech-u.edu.ng
11. Int J Artif Intell ISSN: 2252-8938
Performance investigation of two-stage detection techniques using traffic … (Sunday Adeola Ajagbe)
1919
Adekanmi Adeyinka Adegun received the B.Tech., M.Sc., and Ph.D. degrees
in computer science. He has close to ten years lecturing experience in Universities. He has
also co-supervised M.Sc. and Ph.D. candidates in ML fields. He has published extensively
in several artificial intelligence and computer vision-related accredited journals and
international and national conference proceedings. His main research interests include
artificial intelligence, computer vision, image processing, machine learning, medical image
analysis, pattern recognition, and NLP. He currently serves as a reviewer for some machine
learning and computer vision-related journals. He can be contacted at email:
adegunadekanmi@gmail.com
Ahmed Babajide Olanrewaju is a System Analyst at University of Ibadan,
his interest in the area of NLP with special focus on social media, women assess to health
services. He has been the lead organiser of the AI community in Ibadan which provides
opportunity for women and youth through organising local content resources to aid AI-
based learning and research. He can be contacted at email: a.olanrewaju@ui.edu.ng
Prof. John Babalola Oladosu is a Professor of Computer Engineering and the
current Head of the Department, Computer Engineering, LAUTECH, Ogbomoso, Nigeria.
He is licensed by The Council Regulating Engineering in Nigeria (COREN) as a
professional Computer Engineer and a member of the International Association of
Engineers (IAENG). He has over 20 years of Teaching and Research experience in the
university system. He can be contacted at email: jboladosu@lautech.edu.ng
Prof. Matthew Olusegun Adigun retired in 2020 as a Senior Professor of
Computer Science at the University of Zululand. He obtained his doctorate degree in 1989
from Obafemi Awolowo University, Nigeria; having previously received both Masters in
Computer Science (1984) and a Combined Honours degree in Computer Science and
Economics (1979) from the same University (when it was known as University of Ife,
Nigeria). A very active researcher in Software Engineering of the Wireless Internet, he has
published widely in the specialised areas of reusability, software product lines, and the
engineering of on-demand grid computing-based applications in Mobile Computing,
Mobile Internet and ad hoc Mobile Clouds. Recently, his interest in the Wireless Internet
has extended to Wireless. He has received both research and teaching recognitions for
raising the flag of Excellence in Historically Disadvantaged South African Universities as
well as being awarded a 2020 SAICSIT Pioneer of the Year in the Computing Discipline.
Currently, he works as a Temporary Senior Professor at the University of Zululand to
pursue his recent interest in AI-enabled Pandemic Response and Preparedness. He can be
contacted at email: profmatthewo@gmail.com