IRJET - Direct Me-Nevigation for Blind PeopleIRJET Journal
This document describes a system for direct navigation assistance for blind people using object detection and audio cues. It uses a convolutional neural network model called You Only Look Once (YOLO) to perform real-time object detection on camera images and then describes the detected objects and their locations to the blind user using 3D spatialized sound. The system aims to allow blind users to independently navigate environments by audibly identifying surrounding objects. It analyzes previous works on sensory substitution and assistive technologies for the blind, as well as research on using 3D sound for navigation assistance. The document outlines the object detection methods used, including YOLO and anchor boxes to improve accuracy at detecting multiple objects within each image grid.
This document summarizes an academic paper that proposes a method for incrementally training object detection models to classify unseen object classes in real-time. It begins by providing background on object detection techniques like YOLO and SSD that can perform detection in a single pass. The paper aims to improve these single-shot detectors through incremental learning to classify new object classes without retraining the entire model from scratch. It conducted experiments on YOLO and VGG16 to investigate how well they can classify objects from unseen classes and whether their performance is affected by factors like background, bounding box size, or network architecture. The goal is to develop a more robust object detection method that can easily adapt to new classes of objects in real-time applications.
This document provides an overview of recent developments in object detection using AI robots. It explores several deep learning-based object detection techniques and their advantages over traditional computer vision methods. The paper discusses how object detection is used in robotics applications like grasping, manipulating, and navigation. It also presents the results of experiments conducted to evaluate an object detection system using a robot equipped with cameras and sensors. The system uses a Fast R-CNN algorithm combined with a Kalman filter for real-time object detection and tracking.
OBJECT AND MOTION DEDUCTION SYSTEM USING YOLO IN DEEP LEARNING.pptxJishnu G L
YOLO (You Only Look Once) is a deep learning-based object detection algorithm. Unlike traditional methods that scan an image multiple times, YOLO divides the image into a grid and detects objects in one pass, making it fast and efficient. It predicts both the location (bounding boxes) and class of objects. While primarily used for object detection, YOLO can be combined with tracking techniques for motion deduction in videos. Its speed and accuracy have made it a popular choice in real-time applications.
The goal of the project is to run an object detection algorithm on every frame of a video, thus allowing the algorithm to detect all the objects in it, including but not limited to people, vehicles, animals etc. Object recognition and detection play a crucial role in computer vision and automated driving systems. We aim to design a system that does not compromise on performance or accuracy and provides real time solutions. With the importance of computer vision growing with each passing day, models that deliver high performance results are all the more dominant. Exponential growth in computing power as well as growing popularity in deep learning led to a stark increase in high performance algorithms that solve real world problems. Our model can be taken a step further, allowing the user the flexibility to detect only the objects that are needed at the moment despite being trained on a larger dataset. P. Rajeshwari | P. Abhishek | P. Srikanth | T. Vinod ""Object Detection: An Overview"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-3 , April 2019, URL: https://www.ijtsrd.com/papers/ijtsrd23422.pdf
Paper URL: https://www.ijtsrd.com/computer-science/artificial-intelligence/23422/object-detection-an-overview/p-rajeshwari
auto-assistance system for visually impaired personshahsamkit73
The World Health Organization (WHO) reported that there are 285 million visually-impaired people worldwide. Among these individuals, there are 39 million who are totally blind. There have been several systems designed to support visually-impaired people and to improve the quality of their lives. One of the most difficult activities that must be conducted by visually impaired is indoor navigation. In indoor environment, visually impaired should be aware of obstacles in front of them and be able to avoid it. The use of powered wheelchairs with high transportability and obstacle avoidance intelligence is one of the great steps towards the integration of physically disabled and mentally handicapped people. The disable person will not be able to visualize the object so this Auto-assistance system may suffice the requirement. Auto-Assistance System operating in dynamic environments need to sense its surrounding environment and adapt the control signal in real time to avoid collisions and protect the users. Auto-Assistance System that assist or replace user control could be developed to serve for these users, utilizing systems and algorithms from Auto-Assistance robots. This system could be used to assist disable in their mobility by warning of obstacles. The system could be used in indoor environment like hospital, public garden area. So, we are designing an Auto-assistance system which will help the visually impaired person to work independently. In this system we would be detecting the obstruction in the path of visually impaired person using USB Camera & help them to avoid the collisions.
GitHub Link: https://github.com/shahsamkit73/Auto-Assistance-System-for-visually-impaired
IRJET- Real-Time Object Detection using Deep Learning: A SurveyIRJET Journal
This document summarizes recent advances in real-time object detection using deep learning. It first provides an overview of object detection and deep learning. It then reviews popular object detection models including CNNs, R-CNNs, Fast R-CNN, Faster R-CNN, YOLO, and SSD. The document proposes modifications to existing models to improve small object detection accuracy. Specifically, it proposes using Darknet-53 with feature map upsampling and concatenation at multiple scales to detect objects of different sizes. It also describes using k-means clustering to select anchor boxes tailored to each detection scale.
IRJET - Direct Me-Nevigation for Blind PeopleIRJET Journal
This document describes a system for direct navigation assistance for blind people using object detection and audio cues. It uses a convolutional neural network model called You Only Look Once (YOLO) to perform real-time object detection on camera images and then describes the detected objects and their locations to the blind user using 3D spatialized sound. The system aims to allow blind users to independently navigate environments by audibly identifying surrounding objects. It analyzes previous works on sensory substitution and assistive technologies for the blind, as well as research on using 3D sound for navigation assistance. The document outlines the object detection methods used, including YOLO and anchor boxes to improve accuracy at detecting multiple objects within each image grid.
This document summarizes an academic paper that proposes a method for incrementally training object detection models to classify unseen object classes in real-time. It begins by providing background on object detection techniques like YOLO and SSD that can perform detection in a single pass. The paper aims to improve these single-shot detectors through incremental learning to classify new object classes without retraining the entire model from scratch. It conducted experiments on YOLO and VGG16 to investigate how well they can classify objects from unseen classes and whether their performance is affected by factors like background, bounding box size, or network architecture. The goal is to develop a more robust object detection method that can easily adapt to new classes of objects in real-time applications.
This document provides an overview of recent developments in object detection using AI robots. It explores several deep learning-based object detection techniques and their advantages over traditional computer vision methods. The paper discusses how object detection is used in robotics applications like grasping, manipulating, and navigation. It also presents the results of experiments conducted to evaluate an object detection system using a robot equipped with cameras and sensors. The system uses a Fast R-CNN algorithm combined with a Kalman filter for real-time object detection and tracking.
OBJECT AND MOTION DEDUCTION SYSTEM USING YOLO IN DEEP LEARNING.pptxJishnu G L
YOLO (You Only Look Once) is a deep learning-based object detection algorithm. Unlike traditional methods that scan an image multiple times, YOLO divides the image into a grid and detects objects in one pass, making it fast and efficient. It predicts both the location (bounding boxes) and class of objects. While primarily used for object detection, YOLO can be combined with tracking techniques for motion deduction in videos. Its speed and accuracy have made it a popular choice in real-time applications.
The goal of the project is to run an object detection algorithm on every frame of a video, thus allowing the algorithm to detect all the objects in it, including but not limited to people, vehicles, animals etc. Object recognition and detection play a crucial role in computer vision and automated driving systems. We aim to design a system that does not compromise on performance or accuracy and provides real time solutions. With the importance of computer vision growing with each passing day, models that deliver high performance results are all the more dominant. Exponential growth in computing power as well as growing popularity in deep learning led to a stark increase in high performance algorithms that solve real world problems. Our model can be taken a step further, allowing the user the flexibility to detect only the objects that are needed at the moment despite being trained on a larger dataset. P. Rajeshwari | P. Abhishek | P. Srikanth | T. Vinod ""Object Detection: An Overview"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-3 , April 2019, URL: https://www.ijtsrd.com/papers/ijtsrd23422.pdf
Paper URL: https://www.ijtsrd.com/computer-science/artificial-intelligence/23422/object-detection-an-overview/p-rajeshwari
auto-assistance system for visually impaired personshahsamkit73
The World Health Organization (WHO) reported that there are 285 million visually-impaired people worldwide. Among these individuals, there are 39 million who are totally blind. There have been several systems designed to support visually-impaired people and to improve the quality of their lives. One of the most difficult activities that must be conducted by visually impaired is indoor navigation. In indoor environment, visually impaired should be aware of obstacles in front of them and be able to avoid it. The use of powered wheelchairs with high transportability and obstacle avoidance intelligence is one of the great steps towards the integration of physically disabled and mentally handicapped people. The disable person will not be able to visualize the object so this Auto-assistance system may suffice the requirement. Auto-Assistance System operating in dynamic environments need to sense its surrounding environment and adapt the control signal in real time to avoid collisions and protect the users. Auto-Assistance System that assist or replace user control could be developed to serve for these users, utilizing systems and algorithms from Auto-Assistance robots. This system could be used to assist disable in their mobility by warning of obstacles. The system could be used in indoor environment like hospital, public garden area. So, we are designing an Auto-assistance system which will help the visually impaired person to work independently. In this system we would be detecting the obstruction in the path of visually impaired person using USB Camera & help them to avoid the collisions.
GitHub Link: https://github.com/shahsamkit73/Auto-Assistance-System-for-visually-impaired
IRJET- Real-Time Object Detection using Deep Learning: A SurveyIRJET Journal
This document summarizes recent advances in real-time object detection using deep learning. It first provides an overview of object detection and deep learning. It then reviews popular object detection models including CNNs, R-CNNs, Fast R-CNN, Faster R-CNN, YOLO, and SSD. The document proposes modifications to existing models to improve small object detection accuracy. Specifically, it proposes using Darknet-53 with feature map upsampling and concatenation at multiple scales to detect objects of different sizes. It also describes using k-means clustering to select anchor boxes tailored to each detection scale.
IRJET - Real Time Object Detection using YOLOv3IRJET Journal
The document describes using the YOLO (You Only Look Once) algorithm for real-time object detection. YOLO uses a single neural network to predict bounding boxes and class probabilities for the entire image simultaneously. This allows it to detect multiple objects faster than algorithms that require region proposals or sliding windows. The authors trained a YOLO model to detect bottles, cars, and mobiles using 6000 iterations. On their test dataset, the model achieved a mean average precision of 98.14%, intersection over union of 83.19%, and F1-score of 0.94, demonstrating accurate real-time object detection.
This document discusses the development of a face mask detection system using YOLOv4. The system uses a deep learning model with YOLOv4 to detect faces in real-time video and determine if each person is wearing a mask or not. It is trained on images of faces with and without masks. The model uses CSPDarknet53 as the backbone network and PANet for feature aggregation. It is implemented with OpenCV and a Python GUI for a user interface. The goal is to help enforce mask mandates and alert authorities if too many people in an area are not wearing masks.
Development of wearable object detection system & blind stick for visuall...Arkadev Kundu
It is a wearable device. It has a camera, and it detects all living and non living object. This module detects moving object also. It is made with raspberry pi 3, and a camera. One headphone connect with raspberry pi. When this module detects items, it gave a sound output through headphone. Hence the blind man know that item, which is in-front of him or her. We made it in very low budget, and it is very helpful for visually challenged people. And the Blind stick help him to detect obstacles.
Object detection is a computer technology related to computer vision and image processing that deals with detecting instances of semantic objects of a certain class (such as humans, buildings, or cars) in digital images and videos.
A Small Helping Hand from me to my Engineering collegues and my other friends in need of Object Detection
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorQUESTJOURNAL
ABSTRACT:Tracking of moving objects that is called video tracking is used for measuring motion parameters and obtaining a visual record of the moving objects, it is an important area of application in image processing. In general there are two different approaches to obtain object tracking: the first is Recognition-based Tracking, and the second is the Motion-based Tracking. Video tracking system raises a wide possibility in today’s society. This system is used in various applications such as military, security, monitoring, robotic, and nowadays in dayto-day applications. However the video tracking systems still have many open problems and various research activities in a video tracking system are explores. This paper presents an algorithm for video tracking of any moving targets with the uses of contour based detection technique that depends on the sobel operator. The proposed system is suitable for indoor and outdoor applications. Our approach has the advantage of extending the applicability of tracking system and also, as presented here improves the performance of the tracker making feasible high frame rate video tracking. The goal of the tracking system is to analyze the video frames and estimate the position of a part of the input video frame (usually a moving object), our approach can detect, tracked any object more than one object and calculate the position of the moving objects. Therefore, the aim of this paper is to construct a motion tracking system for moving objects. Where, at the end of this paper, the detail outcome and result are discussed using experiments results of the proposed technique
Detection of a user-defined object in an image using feature extraction- Trai...IRJET Journal
The document proposes a method for detecting user-defined objects in images using feature extraction and training. The method combines contour detection, edge detection, k-means clustering, color identification, and image segmentation. It uses an original "source" object image to train the system to recognize and identify the target object in other images based on a feature set. The key steps include pre-processing images, extracting features like contours and edges, using k-means clustering to identify colors, and analyzing color and shape features to detect matching objects. The results demonstrate the ability to accurately detect target objects against complex backgrounds.
Fast Human Detection in Surveillance VideoIOSR Journals
This document proposes a fast and efficient algorithm for tracking humans in indoor surveillance videos. It uses HOG features and correlation-based methods. Candidate frames are selected using strips along the borders to detect new objects entering the frame. HOG features are extracted only on candidate frames to reduce computation time. A correlation-based method then tracks humans between frames using the highest correlated sample window as the tracked window. Experimental results showed over 86% detection rates on test videos totalling over 6 hours, demonstrating the algorithm can detect and track humans in real-time surveillance applications.
Road signs detection using voila jone's algorithm with the help of opencvMohdSalim34
This document provides an introduction and overview of a project to develop an automatic road sign detection system using the Viola-Jones object detection framework. It discusses the motivation for the project to address safety concerns from drivers missing road signs. The document outlines the contributions of the project, which are to train a classifier using OpenCV to detect German road signs in images by implementing the Viola-Jones algorithm. It also provides details on the Viola-Jones algorithm, which combines Haar features, integral images, AdaBoost testing, and cascading classifiers to rapidly detect objects in real-time.
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...IJECEIAES
The World Health Organization (WHO) reported in 2019 that at least 2.2 billion people were visual-impairment or blindness. The main problem of living for visually impaired people have been facing difficulties in moving even indoor or outdoor situations. Therefore, their lives are not safe and harmful. In this paper, we proposed an assistive application model based on deep learning: YOLOv3 with a Darknet-53 base network for visually impaired people on a smartphone. The Pascal VOC2007 and Pascal VOC2012 were used for the training set and used Pascal VOC2007 test set for validation. The assistive model was installed on a smartphone with an eSpeak synthesizer which generates the audio output to the user. The experimental result showed a high speed and also high detection accuracy. The proposed application with the help of technology will be an effective way to assist visually impaired people to interact with the surrounding environment in their daily life.
The document provides an introduction to computer vision. It discusses key topics including:
- What computer vision is and why it is useful. It uses mathematical and computational tools to extract information from images and improve human vision.
- Some basic concepts in computer vision including digital images, sampling, noise removal, segmentation, and feature extraction techniques.
- Where computer vision is used such as healthcare, autonomous vehicles, augmented/virtual reality, industry, social media, security, agriculture, and fashion.
- A brief history of computer vision including classical approaches and the revolution enabled by advances in artificial intelligence and deep learning.
Real Time Object Detection System with YOLO and CNN Models: A ReviewSpringer
The field of artificial intelligence is built on object detection techniques. YOU ONLY LOOK
ONCE (YOLO) algorithm and it's more evolved versions are briefly described in this research survey. This
survey is all about YOLO and convolution neural networks (CNN) in the direction of real time object detection.
YOLO does generalized object representation more effectively without precision losses than other object
detection models. CNN architecture models have the ability to eliminate highlights and identify objects in any
given image. When implemented appropriately, CNN models can address issues like deformity diagnosis,
creating educational or instructive application, etc. This article reached at number of observations and
perspective findings through the analysis. Also it provides support for the focused visual information and
feature extraction in the financial and other industries, highlights the method of target detection and feature
selection, and briefly describes the development process of yolo algorithm
1. The document discusses Faster YOLO, a new object detection method that is both accurate and faster than previous methods like YOLO. Faster YOLO divides the image into grids and has each grid predict bounding box coordinates and labels for the objects within it, improving speed over YOLO.
2. Faster YOLO addresses the issue of duplicate predictions that YOLO has by suppressing lower probability bounding boxes using non-maximal suppression. Real-world applications of object detection algorithms like Faster YOLO and YOLO include autonomous vehicles, security, healthcare, agriculture, and manufacturing.
3. While great progress has been made in object detection, human-level performance is
DEEP LEARNING APPROACH FOR EVENT MONITORING SYSTEMIJMIT JOURNAL
With an increasing number of extreme events and complexity, more alarms are being used to monitor
control rooms. Operators in the control rooms need to monitor and analyze these alarms to take suitable
actions to ensure the system’s stability and security. Security is the biggest concern in the modern world. It
is important to have a rigid surveillance that should guarantee protection from any sought of hazard.
Considering security, Closed Circuit TV (CCTV) cameras are being utilized for reconnaissance, but these
CCTV cameras require a person for supervision. As a human being, there can be a possibility to be tired
off in supervision at any point of time. So, we need a system to detect automatically. Thus, we came up with
a solution using YOLO V5. We have taken a data set and used robo-flow framework to enhance the existing
images into numerous variations where it will create a copy of grey scale image, a copy of its rotation and
a copy of its blurred version which will be used to get an enlarged data set. This work mainly focuses on
providing a secure environment using CCTV live footage as a source to detect the weapons. Using YOLO
algorithm, it divides an image from the video into grid system and each grid detects an object within itself
In Comparison with other object detection algorithms, YOLO proposes the use of an end-to-end neural network that makes predictions of bounding boxes and class probabilities all at once.
A Critical Survey on Detection of Object and Tracking of Object With differen...Editor IJMTER
Basically object detection and object tracking are two important and challenging aspects in
many computer vision applications like surveillance system, vehicle navigation, autonomous robot
navigation, compression of video etc. Object detection is first low level important task for any video
surveillance application. To detection of moving object is a challenging task. Tracking is required in
higher level applications that required the location and shape of object. There are three key steps in
video analysis: detection of interesting moving objects, tracking of such objects from frame to frame,
and analysis of object tracks to recognize their behavior. Object detection and tracking especially for
human and vehicle is currently most active research topic. A lot of research has been undergoing
ranging from applications to noble algorithms. The main objective of this paper is to review (survey)
of various moving object detection and object tracking methodologies.
The document describes a social distancing detection system that uses the YOLO object detection algorithm and COCO dataset to detect people in video frames and estimate distances between them. It draws bounding boxes around detected people, with violations of the default distance threshold shown in red and non-violations in green. The number of violations and alert messages are displayed on screen to help users maintain safe distances. The system was tested on prerecorded video and images and was able to accurately detect violations and maintain social distancing.
A Survey on Approaches for Object Trackingjournal ijrtem
This document summarizes various approaches for object tracking in video sequences. It discusses common object detection methods like temporal differencing, optical flow and background subtraction. For object representation, it describes shape-based, motion-based, color-based and texture-based approaches. For object tracking, it analyzes target representation and localization methods like blob tracking and mean shift, as well as filtering and data association approaches like the Kalman filter and particle filter. It provides comparisons between Kalman and particle filters, and between contour tracking and visual feature matching methods. The document concludes that further research could improve computational efficiency and decrease tracking time for diverse video content.
A Smart Assistance for Visually ImpairedIRJET Journal
The document describes a smart assistance system developed for visually impaired people. The system uses technologies like object detection, speech recognition and text-to-speech to help visually impaired users recognize objects, search Wikipedia and send emails through voice commands. Specifically, it uses a YOLO v3 deep learning model trained on the COCO dataset to detect objects in images captured by a camera. It also integrates APIs to search Wikipedia and uses Python libraries for speech recognition, text-to-speech and email functionality. The system was developed in Python and aims to provide more independence to visually impaired users in their daily lives.
Detection and Tracking of Objects: A Detailed StudyIJEACS
Detecting and tracking objects are the most widespread and challenging tasks that a surveillance system must achieve to determine expressive events and activities, and automatically interpret and recover video content. An object can be a queue of people, a human, a head or a face. The goal of this article is to state the Detecting and tracking methods, classify them into different categories, and identify new trends, we introduce main trends and provide method to give a perception to fundamental ideas as well as to show their limitations in the object detection and tracking for more effective video analytics.
This document discusses object identification using convolutional neural networks and the YOLO detection algorithm. It begins with an introduction to neural networks and their history. It then discusses datasets used to train object detection models. The document describes experiments conducted using the YOLO detector on different sized images to evaluate performance. Processing speed and objects detected were compared between the CPU and GPU. The YOLO detector was then tested on a set of 500 images, and its performance metrics were reported.
ACEP Magazine edition 4th launched on 05.06.2024Rahul
This document provides information about the third edition of the magazine "Sthapatya" published by the Association of Civil Engineers (Practicing) Aurangabad. It includes messages from current and past presidents of ACEP, memories and photos from past ACEP events, information on life time achievement awards given by ACEP, and a technical article on concrete maintenance, repairs and strengthening. The document highlights activities of ACEP and provides a technical educational article for members.
More Related Content
Similar to ppt - Copy for projects will help you further
IRJET - Real Time Object Detection using YOLOv3IRJET Journal
The document describes using the YOLO (You Only Look Once) algorithm for real-time object detection. YOLO uses a single neural network to predict bounding boxes and class probabilities for the entire image simultaneously. This allows it to detect multiple objects faster than algorithms that require region proposals or sliding windows. The authors trained a YOLO model to detect bottles, cars, and mobiles using 6000 iterations. On their test dataset, the model achieved a mean average precision of 98.14%, intersection over union of 83.19%, and F1-score of 0.94, demonstrating accurate real-time object detection.
This document discusses the development of a face mask detection system using YOLOv4. The system uses a deep learning model with YOLOv4 to detect faces in real-time video and determine if each person is wearing a mask or not. It is trained on images of faces with and without masks. The model uses CSPDarknet53 as the backbone network and PANet for feature aggregation. It is implemented with OpenCV and a Python GUI for a user interface. The goal is to help enforce mask mandates and alert authorities if too many people in an area are not wearing masks.
Development of wearable object detection system & blind stick for visuall...Arkadev Kundu
It is a wearable device. It has a camera, and it detects all living and non living object. This module detects moving object also. It is made with raspberry pi 3, and a camera. One headphone connect with raspberry pi. When this module detects items, it gave a sound output through headphone. Hence the blind man know that item, which is in-front of him or her. We made it in very low budget, and it is very helpful for visually challenged people. And the Blind stick help him to detect obstacles.
Object detection is a computer technology related to computer vision and image processing that deals with detecting instances of semantic objects of a certain class (such as humans, buildings, or cars) in digital images and videos.
A Small Helping Hand from me to my Engineering collegues and my other friends in need of Object Detection
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorQUESTJOURNAL
ABSTRACT:Tracking of moving objects that is called video tracking is used for measuring motion parameters and obtaining a visual record of the moving objects, it is an important area of application in image processing. In general there are two different approaches to obtain object tracking: the first is Recognition-based Tracking, and the second is the Motion-based Tracking. Video tracking system raises a wide possibility in today’s society. This system is used in various applications such as military, security, monitoring, robotic, and nowadays in dayto-day applications. However the video tracking systems still have many open problems and various research activities in a video tracking system are explores. This paper presents an algorithm for video tracking of any moving targets with the uses of contour based detection technique that depends on the sobel operator. The proposed system is suitable for indoor and outdoor applications. Our approach has the advantage of extending the applicability of tracking system and also, as presented here improves the performance of the tracker making feasible high frame rate video tracking. The goal of the tracking system is to analyze the video frames and estimate the position of a part of the input video frame (usually a moving object), our approach can detect, tracked any object more than one object and calculate the position of the moving objects. Therefore, the aim of this paper is to construct a motion tracking system for moving objects. Where, at the end of this paper, the detail outcome and result are discussed using experiments results of the proposed technique
Detection of a user-defined object in an image using feature extraction- Trai...IRJET Journal
The document proposes a method for detecting user-defined objects in images using feature extraction and training. The method combines contour detection, edge detection, k-means clustering, color identification, and image segmentation. It uses an original "source" object image to train the system to recognize and identify the target object in other images based on a feature set. The key steps include pre-processing images, extracting features like contours and edges, using k-means clustering to identify colors, and analyzing color and shape features to detect matching objects. The results demonstrate the ability to accurately detect target objects against complex backgrounds.
Fast Human Detection in Surveillance VideoIOSR Journals
This document proposes a fast and efficient algorithm for tracking humans in indoor surveillance videos. It uses HOG features and correlation-based methods. Candidate frames are selected using strips along the borders to detect new objects entering the frame. HOG features are extracted only on candidate frames to reduce computation time. A correlation-based method then tracks humans between frames using the highest correlated sample window as the tracked window. Experimental results showed over 86% detection rates on test videos totalling over 6 hours, demonstrating the algorithm can detect and track humans in real-time surveillance applications.
Road signs detection using voila jone's algorithm with the help of opencvMohdSalim34
This document provides an introduction and overview of a project to develop an automatic road sign detection system using the Viola-Jones object detection framework. It discusses the motivation for the project to address safety concerns from drivers missing road signs. The document outlines the contributions of the project, which are to train a classifier using OpenCV to detect German road signs in images by implementing the Viola-Jones algorithm. It also provides details on the Viola-Jones algorithm, which combines Haar features, integral images, AdaBoost testing, and cascading classifiers to rapidly detect objects in real-time.
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...IJECEIAES
The World Health Organization (WHO) reported in 2019 that at least 2.2 billion people were visual-impairment or blindness. The main problem of living for visually impaired people have been facing difficulties in moving even indoor or outdoor situations. Therefore, their lives are not safe and harmful. In this paper, we proposed an assistive application model based on deep learning: YOLOv3 with a Darknet-53 base network for visually impaired people on a smartphone. The Pascal VOC2007 and Pascal VOC2012 were used for the training set and used Pascal VOC2007 test set for validation. The assistive model was installed on a smartphone with an eSpeak synthesizer which generates the audio output to the user. The experimental result showed a high speed and also high detection accuracy. The proposed application with the help of technology will be an effective way to assist visually impaired people to interact with the surrounding environment in their daily life.
The document provides an introduction to computer vision. It discusses key topics including:
- What computer vision is and why it is useful. It uses mathematical and computational tools to extract information from images and improve human vision.
- Some basic concepts in computer vision including digital images, sampling, noise removal, segmentation, and feature extraction techniques.
- Where computer vision is used such as healthcare, autonomous vehicles, augmented/virtual reality, industry, social media, security, agriculture, and fashion.
- A brief history of computer vision including classical approaches and the revolution enabled by advances in artificial intelligence and deep learning.
Real Time Object Detection System with YOLO and CNN Models: A ReviewSpringer
The field of artificial intelligence is built on object detection techniques. YOU ONLY LOOK
ONCE (YOLO) algorithm and it's more evolved versions are briefly described in this research survey. This
survey is all about YOLO and convolution neural networks (CNN) in the direction of real time object detection.
YOLO does generalized object representation more effectively without precision losses than other object
detection models. CNN architecture models have the ability to eliminate highlights and identify objects in any
given image. When implemented appropriately, CNN models can address issues like deformity diagnosis,
creating educational or instructive application, etc. This article reached at number of observations and
perspective findings through the analysis. Also it provides support for the focused visual information and
feature extraction in the financial and other industries, highlights the method of target detection and feature
selection, and briefly describes the development process of yolo algorithm
1. The document discusses Faster YOLO, a new object detection method that is both accurate and faster than previous methods like YOLO. Faster YOLO divides the image into grids and has each grid predict bounding box coordinates and labels for the objects within it, improving speed over YOLO.
2. Faster YOLO addresses the issue of duplicate predictions that YOLO has by suppressing lower probability bounding boxes using non-maximal suppression. Real-world applications of object detection algorithms like Faster YOLO and YOLO include autonomous vehicles, security, healthcare, agriculture, and manufacturing.
3. While great progress has been made in object detection, human-level performance is
DEEP LEARNING APPROACH FOR EVENT MONITORING SYSTEMIJMIT JOURNAL
With an increasing number of extreme events and complexity, more alarms are being used to monitor
control rooms. Operators in the control rooms need to monitor and analyze these alarms to take suitable
actions to ensure the system’s stability and security. Security is the biggest concern in the modern world. It
is important to have a rigid surveillance that should guarantee protection from any sought of hazard.
Considering security, Closed Circuit TV (CCTV) cameras are being utilized for reconnaissance, but these
CCTV cameras require a person for supervision. As a human being, there can be a possibility to be tired
off in supervision at any point of time. So, we need a system to detect automatically. Thus, we came up with
a solution using YOLO V5. We have taken a data set and used robo-flow framework to enhance the existing
images into numerous variations where it will create a copy of grey scale image, a copy of its rotation and
a copy of its blurred version which will be used to get an enlarged data set. This work mainly focuses on
providing a secure environment using CCTV live footage as a source to detect the weapons. Using YOLO
algorithm, it divides an image from the video into grid system and each grid detects an object within itself
In Comparison with other object detection algorithms, YOLO proposes the use of an end-to-end neural network that makes predictions of bounding boxes and class probabilities all at once.
A Critical Survey on Detection of Object and Tracking of Object With differen...Editor IJMTER
Basically object detection and object tracking are two important and challenging aspects in
many computer vision applications like surveillance system, vehicle navigation, autonomous robot
navigation, compression of video etc. Object detection is first low level important task for any video
surveillance application. To detection of moving object is a challenging task. Tracking is required in
higher level applications that required the location and shape of object. There are three key steps in
video analysis: detection of interesting moving objects, tracking of such objects from frame to frame,
and analysis of object tracks to recognize their behavior. Object detection and tracking especially for
human and vehicle is currently most active research topic. A lot of research has been undergoing
ranging from applications to noble algorithms. The main objective of this paper is to review (survey)
of various moving object detection and object tracking methodologies.
The document describes a social distancing detection system that uses the YOLO object detection algorithm and COCO dataset to detect people in video frames and estimate distances between them. It draws bounding boxes around detected people, with violations of the default distance threshold shown in red and non-violations in green. The number of violations and alert messages are displayed on screen to help users maintain safe distances. The system was tested on prerecorded video and images and was able to accurately detect violations and maintain social distancing.
A Survey on Approaches for Object Trackingjournal ijrtem
This document summarizes various approaches for object tracking in video sequences. It discusses common object detection methods like temporal differencing, optical flow and background subtraction. For object representation, it describes shape-based, motion-based, color-based and texture-based approaches. For object tracking, it analyzes target representation and localization methods like blob tracking and mean shift, as well as filtering and data association approaches like the Kalman filter and particle filter. It provides comparisons between Kalman and particle filters, and between contour tracking and visual feature matching methods. The document concludes that further research could improve computational efficiency and decrease tracking time for diverse video content.
A Smart Assistance for Visually ImpairedIRJET Journal
The document describes a smart assistance system developed for visually impaired people. The system uses technologies like object detection, speech recognition and text-to-speech to help visually impaired users recognize objects, search Wikipedia and send emails through voice commands. Specifically, it uses a YOLO v3 deep learning model trained on the COCO dataset to detect objects in images captured by a camera. It also integrates APIs to search Wikipedia and uses Python libraries for speech recognition, text-to-speech and email functionality. The system was developed in Python and aims to provide more independence to visually impaired users in their daily lives.
Detection and Tracking of Objects: A Detailed StudyIJEACS
Detecting and tracking objects are the most widespread and challenging tasks that a surveillance system must achieve to determine expressive events and activities, and automatically interpret and recover video content. An object can be a queue of people, a human, a head or a face. The goal of this article is to state the Detecting and tracking methods, classify them into different categories, and identify new trends, we introduce main trends and provide method to give a perception to fundamental ideas as well as to show their limitations in the object detection and tracking for more effective video analytics.
This document discusses object identification using convolutional neural networks and the YOLO detection algorithm. It begins with an introduction to neural networks and their history. It then discusses datasets used to train object detection models. The document describes experiments conducted using the YOLO detector on different sized images to evaluate performance. Processing speed and objects detected were compared between the CPU and GPU. The YOLO detector was then tested on a set of 500 images, and its performance metrics were reported.
Similar to ppt - Copy for projects will help you further (20)
ACEP Magazine edition 4th launched on 05.06.2024Rahul
This document provides information about the third edition of the magazine "Sthapatya" published by the Association of Civil Engineers (Practicing) Aurangabad. It includes messages from current and past presidents of ACEP, memories and photos from past ACEP events, information on life time achievement awards given by ACEP, and a technical article on concrete maintenance, repairs and strengthening. The document highlights activities of ACEP and provides a technical educational article for members.
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesChristina Lin
Traditionally, dealing with real-time data pipelines has involved significant overhead, even for straightforward tasks like data transformation or masking. However, in this talk, we’ll venture into the dynamic realm of WebAssembly (WASM) and discover how it can revolutionize the creation of stateless streaming pipelines within a Kafka (Redpanda) broker. These pipelines are adept at managing low-latency, high-data-volume scenarios.
Comparative analysis between traditional aquaponics and reconstructed aquapon...bijceesjournal
The aquaponic system of planting is a method that does not require soil usage. It is a method that only needs water, fish, lava rocks (a substitute for soil), and plants. Aquaponic systems are sustainable and environmentally friendly. Its use not only helps to plant in small spaces but also helps reduce artificial chemical use and minimizes excess water use, as aquaponics consumes 90% less water than soil-based gardening. The study applied a descriptive and experimental design to assess and compare conventional and reconstructed aquaponic methods for reproducing tomatoes. The researchers created an observation checklist to determine the significant factors of the study. The study aims to determine the significant difference between traditional aquaponics and reconstructed aquaponics systems propagating tomatoes in terms of height, weight, girth, and number of fruits. The reconstructed aquaponics system’s higher growth yield results in a much more nourished crop than the traditional aquaponics system. It is superior in its number of fruits, height, weight, and girth measurement. Moreover, the reconstructed aquaponics system is proven to eliminate all the hindrances present in the traditional aquaponics system, which are overcrowding of fish, algae growth, pest problems, contaminated water, and dead fish.
Presentation of IEEE Slovenia CIS (Computational Intelligence Society) Chapte...University of Maribor
Slides from talk presenting:
Aleš Zamuda: Presentation of IEEE Slovenia CIS (Computational Intelligence Society) Chapter and Networking.
Presentation at IcETRAN 2024 session:
"Inter-Society Networking Panel GRSS/MTT-S/CIS
Panel Session: Promoting Connection and Cooperation"
IEEE Slovenia GRSS
IEEE Serbia and Montenegro MTT-S
IEEE Slovenia CIS
11TH INTERNATIONAL CONFERENCE ON ELECTRICAL, ELECTRONIC AND COMPUTING ENGINEERING
3-6 June 2024, Niš, Serbia
Understanding Inductive Bias in Machine LearningSUTEJAS
This presentation explores the concept of inductive bias in machine learning. It explains how algorithms come with built-in assumptions and preferences that guide the learning process. You'll learn about the different types of inductive bias and how they can impact the performance and generalizability of machine learning models.
The presentation also covers the positive and negative aspects of inductive bias, along with strategies for mitigating potential drawbacks. We'll explore examples of how bias manifests in algorithms like neural networks and decision trees.
By understanding inductive bias, you can gain valuable insights into how machine learning models work and make informed decisions when building and deploying them.
A SYSTEMATIC RISK ASSESSMENT APPROACH FOR SECURING THE SMART IRRIGATION SYSTEMSIJNSA Journal
The smart irrigation system represents an innovative approach to optimize water usage in agricultural and landscaping practices. The integration of cutting-edge technologies, including sensors, actuators, and data analysis, empowers this system to provide accurate monitoring and control of irrigation processes by leveraging real-time environmental conditions. The main objective of a smart irrigation system is to optimize water efficiency, minimize expenses, and foster the adoption of sustainable water management methods. This paper conducts a systematic risk assessment by exploring the key components/assets and their functionalities in the smart irrigation system. The crucial role of sensors in gathering data on soil moisture, weather patterns, and plant well-being is emphasized in this system. These sensors enable intelligent decision-making in irrigation scheduling and water distribution, leading to enhanced water efficiency and sustainable water management practices. Actuators enable automated control of irrigation devices, ensuring precise and targeted water delivery to plants. Additionally, the paper addresses the potential threat and vulnerabilities associated with smart irrigation systems. It discusses limitations of the system, such as power constraints and computational capabilities, and calculates the potential security risks. The paper suggests possible risk treatment methods for effective secure system operation. In conclusion, the paper emphasizes the significant benefits of implementing smart irrigation systems, including improved water conservation, increased crop yield, and reduced environmental impact. Additionally, based on the security analysis conducted, the paper recommends the implementation of countermeasures and security approaches to address vulnerabilities and ensure the integrity and reliability of the system. By incorporating these measures, smart irrigation technology can revolutionize water management practices in agriculture, promoting sustainability, resource efficiency, and safeguarding against potential security threats.
Embedded machine learning-based road conditions and driving behavior monitoringIJECEIAES
Car accident rates have increased in recent years, resulting in losses in human lives, properties, and other financial costs. An embedded machine learning-based system is developed to address this critical issue. The system can monitor road conditions, detect driving patterns, and identify aggressive driving behaviors. The system is based on neural networks trained on a comprehensive dataset of driving events, driving styles, and road conditions. The system effectively detects potential risks and helps mitigate the frequency and impact of accidents. The primary goal is to ensure the safety of drivers and vehicles. Collecting data involved gathering information on three key road events: normal street and normal drive, speed bumps, circular yellow speed bumps, and three aggressive driving actions: sudden start, sudden stop, and sudden entry. The gathered data is processed and analyzed using a machine learning system designed for limited power and memory devices. The developed system resulted in 91.9% accuracy, 93.6% precision, and 92% recall. The achieved inference time on an Arduino Nano 33 BLE Sense with a 32-bit CPU running at 64 MHz is 34 ms and requires 2.6 kB peak RAM and 139.9 kB program flash memory, making it suitable for resource-constrained embedded systems.
TIME DIVISION MULTIPLEXING TECHNIQUE FOR COMMUNICATION SYSTEMHODECEDSIET
Time Division Multiplexing (TDM) is a method of transmitting multiple signals over a single communication channel by dividing the signal into many segments, each having a very short duration of time. These time slots are then allocated to different data streams, allowing multiple signals to share the same transmission medium efficiently. TDM is widely used in telecommunications and data communication systems.
### How TDM Works
1. **Time Slots Allocation**: The core principle of TDM is to assign distinct time slots to each signal. During each time slot, the respective signal is transmitted, and then the process repeats cyclically. For example, if there are four signals to be transmitted, the TDM cycle will divide time into four slots, each assigned to one signal.
2. **Synchronization**: Synchronization is crucial in TDM systems to ensure that the signals are correctly aligned with their respective time slots. Both the transmitter and receiver must be synchronized to avoid any overlap or loss of data. This synchronization is typically maintained by a clock signal that ensures time slots are accurately aligned.
3. **Frame Structure**: TDM data is organized into frames, where each frame consists of a set of time slots. Each frame is repeated at regular intervals, ensuring continuous transmission of data streams. The frame structure helps in managing the data streams and maintaining the synchronization between the transmitter and receiver.
4. **Multiplexer and Demultiplexer**: At the transmitting end, a multiplexer combines multiple input signals into a single composite signal by assigning each signal to a specific time slot. At the receiving end, a demultiplexer separates the composite signal back into individual signals based on their respective time slots.
### Types of TDM
1. **Synchronous TDM**: In synchronous TDM, time slots are pre-assigned to each signal, regardless of whether the signal has data to transmit or not. This can lead to inefficiencies if some time slots remain empty due to the absence of data.
2. **Asynchronous TDM (or Statistical TDM)**: Asynchronous TDM addresses the inefficiencies of synchronous TDM by allocating time slots dynamically based on the presence of data. Time slots are assigned only when there is data to transmit, which optimizes the use of the communication channel.
### Applications of TDM
- **Telecommunications**: TDM is extensively used in telecommunication systems, such as in T1 and E1 lines, where multiple telephone calls are transmitted over a single line by assigning each call to a specific time slot.
- **Digital Audio and Video Broadcasting**: TDM is used in broadcasting systems to transmit multiple audio or video streams over a single channel, ensuring efficient use of bandwidth.
- **Computer Networks**: TDM is used in network protocols and systems to manage the transmission of data from multiple sources over a single network medium.
### Advantages of TDM
- **Efficient Use of Bandwidth**: TDM all
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTjpsjournal1
The rivalry between prominent international actors for dominance over Central Asia's hydrocarbon
reserves and the ancient silk trade route, along with China's diplomatic endeavours in the area, has been
referred to as the "New Great Game." This research centres on the power struggle, considering
geopolitical, geostrategic, and geoeconomic variables. Topics including trade, political hegemony, oil
politics, and conventional and nontraditional security are all explored and explained by the researcher.
Using Mackinder's Heartland, Spykman Rimland, and Hegemonic Stability theories, examines China's role
in Central Asia. This study adheres to the empirical epistemological method and has taken care of
objectivity. This study analyze primary and secondary research documents critically to elaborate role of
china’s geo economic outreach in central Asian countries and its future prospect. China is thriving in trade,
pipeline politics, and winning states, according to this study, thanks to important instruments like the
Shanghai Cooperation Organisation and the Belt and Road Economic Initiative. According to this study,
China is seeing significant success in commerce, pipeline politics, and gaining influence on other
governments. This success may be attributed to the effective utilisation of key tools such as the Shanghai
Cooperation Organisation and the Belt and Road Economic Initiative.
Advanced control scheme of doubly fed induction generator for wind turbine us...IJECEIAES
This paper describes a speed control device for generating electrical energy on an electricity network based on the doubly fed induction generator (DFIG) used for wind power conversion systems. At first, a double-fed induction generator model was constructed. A control law is formulated to govern the flow of energy between the stator of a DFIG and the energy network using three types of controllers: proportional integral (PI), sliding mode controller (SMC) and second order sliding mode controller (SOSMC). Their different results in terms of power reference tracking, reaction to unexpected speed fluctuations, sensitivity to perturbations, and resilience against machine parameter alterations are compared. MATLAB/Simulink was used to conduct the simulations for the preceding study. Multiple simulations have shown very satisfying results, and the investigations demonstrate the efficacy and power-enhancing capabilities of the suggested control system.
Eric Nizeyimana's document 2006 from gicumbi to ttc nyamata handball play
ppt - Copy for projects will help you further
1. Major Project Presentation
on
REAL TIME OBJECT RECOGNITION FOR VISUALLY
IMPAIRED PEOPLE
Mahatma Gandhi Mission’s College Of
Engineering & Technology
A-09, Sector 62, Noida, Uttar Pradesh 201301
Submitted by:
Vikas Kumar Pandey Akshay kumar Hariom
Roll No.:1900950310011 Roll no:1900950310002 Roll no: 190950310006
2. Content
Introduction
Problems faced by blind peoples
Literature review
Objective
Block diagram
Yolo algorithm
Block diagram of yolo algorithm
Object detection
Database used
Methodology
Flow chart
Hardware used
Advantages of yolo algorithm
Survey
Advantages
Conclusion
Future work
Reference
3. Introduction
The World Health Organization (WHO) had a survey over around 7889 million people.
The statistics showed that among the population under consideration while survey, 253
millions were visually impaired.[4]
There are many visually impaired people facing many problems in our society.
The device developed can detect the objects in the user's surroundings.
This is a model has been proposed which makes the visually impaired people detect
objects in his surroundings. The output of the system is in audio form that can be easily
understandable for a blind user.
4. Problem faced by blind people
Visually Impaired People confront many problems in recognizing the
objects.
Blind people can’t recognize the objects next to them without touching
them.
This is developed to detect the objects in the user's surroundings.
It will also solve the problem of keeping a walking stick.
5. Literature review
1. “The authors in(Seema et al ) suggested using a smart system that guides a
blind person in 2016[1]”
• The system detects the obstacles that could not be detected by his/her cane.
However, the proposed system was designed to protect the blind from the area
near to his/her head.
Problem statement - The buzzer and vibrator were used and employed as output
modes to a user. This is useful for obstacles detection only at a head level without
recognizing the type of obstacles.
6. Contd.
2. “A modification of several systems used in visual recognition was proposed
in 2014.[2]”
• The authors used fast-feature pyramids and provided findings on general object
detection systems. The results showed that the proposed scheme can be strictly
used for wide-spectrum images.
Problem statement - It does not succeed for narrow-spectrum images. Hence,
their work cannot be used as efficient general objects detection.
7. Contd.
3. “In (Nazli Mohajeri et al, 2011) the authors suggested a two-camera
system to capture photos”.[3]
• However, the proposed system was only tested under three conditions and for
three objects. Specific obstacles that have distances from cameras of about 70 cm
were detected.
Problem statement - The results showed some range of error. Blind helping
systems need to cover more cases with efficient and satisfied results.
8. Objective
This project aims to relieve some of their problems using assistive technology.
Simply it is the technique of real time stationary object recognition.
To make visually impaired people self independent.
To provide a device for detection of objects.
Our main aim is, an object recognition function with device should be able to detect
certain items from the camera and return an audio output to announce what it is. In
order to recognize object, machine learning has to be involved.
10. METHODOLOGY
The steps of a currency recognition system based on image processing are as follows
–
Image capturing
Image Acquisition
Object detection
YOLO algorithm
Prediction
12. Capturing image
Capturing of image is done by camera
module for that purpose the objects
captured in real time and stationary also.
13. Image Acquisition
The image is captured by digital camera
as RGB image and is converted to Gray
scale version by intensity equation 1.
I = (R+G+B)/3
14. RESIDUAL BLOCKS
The image is divided into various grids. Each grid has
a dimension of S x S.
It uses the dimensions of 3 x 3, 13 x 13 and 19 x 19.
There are many grid cells of equal dimension. Every
grid cell will detect objects that appear within it.
15. LOCALIZATION
The term 'localization' refers to where the
object in the image is present. In YOLO object
detection we classify image with localization
i.e., a supervised learning algorithm is trained
to not only predict class but also the bounding
box around the object in image.
Classification + localization = object detection
16. BOUNDING BOXES
A bounding box is an outline that highlights an
object in an image.
Every bounding box in the image consists of
the following attributes:
• Bounding box center (bx, by)
• Height (bh)
• Width (bw)
• Class (for example, person, car, traffic light,
etc.). This is represented by the letter c.
(bw)
(bh) .
(bx, by)
17. BOUNDING BOXES - CONT...
Each 13x13 cell detects objects in the input image
via its specified number of bounding boxes 13x13.
In YOLO v4, each cell has 3 bounding boxes. So
the total number of bounding boxes using 13x13
feature map would be.
(13x13)x3 = 507 bounding boxes.
The remaining bounding boxes are discarded as
they don't localize the objects in the picture.
18. TARGET LABEL Y
Target label y for this supervised learning task is
explained as:
Y is a vector containing Pc, Bx, By, Bh, Bw, CI,..., Ch
Pc is the probability of presence of particular class in
the grid cell. Pc >=0 and <=1. (i.e., Pc=0) means that
object is not found. Pc>I means 100% probability that
object is present.
(Bx, By) defines the mid-point of object and (Bh, Bw)
defines the height and width of bounding box.
Also, if Pc > 0 then there will be n number of C which
represents the classes of objects present in the image.
19. Intersection over union (IOU)
(Intersection over Union) is a term used to
describe the extent of overlap of two boxes. The
greater the region of overlap, the greater the
IOU.
IOU is mainly used in applications related to
object detection, where we train a model to
output a box that fits perfectly around an object.
IOU is also used in non max suppression
algorithm.
𝑰𝑶𝑼 = 𝑰𝑵𝑻𝑬𝑹𝑺𝑬𝑪𝑻𝑰𝑶𝑵 𝑨𝑹𝑬𝑨 𝑶𝑭 𝑶𝑽𝑬𝑹𝑳𝑨𝑷
UNION
20. NMS- NONMAX SUPRESSION
To select the best bounding box, from the multiple predicted bounding
boxes, an algorithm called Non-Max Suppression is used to
"suppress" the less likely bounding boxes and keep only the best one.
21. Dataset
Coco dataset – COCO dataset, meaning “Common Objects In Context”.
It is a large-scale image dataset containing 328,000 images of everyday objects
and humans.
The dataset contains annotations of deep learning models to recognize, label, and
describe objects.
COCO provides the following types of annotations:
• Object detection
• Captioning
22. Object detection
Object detection is a phenomenon in computer vision that
involves the detection of various objects in digital images or
videos.
Some of the objects detected include people, cars, chairs, stones,
buildings, and animals.
It identify the object in a specific image.
Establish the exact location of the object within the image.
23. Contd:
Object detection consists of various approaches such as fast R-CNN,
Retina-Net, and Sliding Window detection but none of the
aforementioned methods can detect object in one single run. So there
comes another efficient and faster algorithm called YOLO algorithm.
24. Sr no ALGORITHM ADVANTAGE DISADVANTAGE
1 RESNET • solve degradation problem by shortcuts
• skip connections.
• RESNETs are that for a deeper network the detection
of errors becomes difficult.
2 R-CNN • very accurate at image recognition and
classification
• They fail to encode the position and orientation of
objects.
3 FAST R-CNN • save time compared to traditional algorithms like
Selective Search.
• It still uses the Selective Search Algorithm which is
slow and a time-consuming process.
4 SSD • SSD makes more predictions.
• It has better coverage on location, scale, and
aspect ratios.
• Shallow layers in a neural network may not generate
enough high level features to do prediction for small
objects.
5 YOLO
• Allows real time object detection.
• System trains in single go.
• More efficient and fast.
• Struggles to detect close objects because each grid
can propose only 2 bounding boxes.
EXISTING ALGORITHM
25. The YOLOv4 performance was evaluated based on previous YOLO versions (YOLOv3
and YOLOv2)as baselines.
The new YOLOv4 shows the best speed-to-accuracy balance compared to state-of-the-art
object detectors.
In general, YOLOv4 surpasses all previous object detectors in terms of both speed and
accuracy, ranging from 5 FPS to as much as 160 FPS.
The YOLO v4 algorithm achieves the highest accuracy among all other real-time object
detection models – while achieving 30 FPS or higher using a GPU.
ALGORITHM SELECTION
26. YOLO algorithm
YOLO is an abbreviation for the term 'You Only Look Once’.
Created by Joseph Redmon, Santosh Divvala, Ross Girshick and Ali Farhadi.
YOLO algorithm detects and recognizes various objects in the picture.
Object detection in YOLO is done as a regression problem and provides the class
probabilities of the detected images
Prediction in the entire image is done in a single algorithmic run.
YOLO algorithm consists of various variants including tiny YOLO and YOLOv1,
v2, v3, v4.
Popular because of its speed and accuracy.
27. Yolo evolution
Algorithm Description
The original YOLO - YOLO was the first object detection network to combine the problem of drawing
bounding boxes and identifying class labels in one end-to-end differentiable
network.
YOLOv2 - YOLOv2 made a number of iterative improvements on top of YOLO including
BatchNorm, higher resolution, and anchor boxes.
YOLOv3 - YOLOv3 built upon previous models by adding an objects score to bounding box
prediction, added connections to the backbone network layers and made predictions at
three separate levels of granularity to improve performance on smaller objects.
YOLOv4 - It is a one-stage detector with several components in it. It detects the object in real
time. The speed and accuracy is faster than other algorithm.
29. CSP DARKNET53
CSPDarknet53 is a convolutional neural network and
backbone for object detection.
It employs a strategy to partition the feature map of the
Image into two parts and then merges them through a
cross-stage hierarchy.
The use of a split and merge strategy allows for more
gradient flow through the network.
30. SPATIAL PYRAMID POOLING
A CNN consists of some Convolutional
(Conv) layers followed by some Fully-
Connected (FC) layers. Conv layers don’t
require fixed-size input .
The solution to this problem lies in the
Spatial Pyramid Pooling (SPP) layer. It is
placed between the last Conv layer and the
first FC layer and removes the fixed-size
constraint of the network.
The goal of the SPP layer is to pool the
variable-size features that come from the
Conv layer and generate fixed-length outputs
that will then be fed to the first FC layer of
the network.
31. BAG OF FREEBIES AND SPACIALS
‘Bag of Freebies’ (BoF) is a general framework of training strategies for improving
the overall accuracy of an object detection model.
The set of techniques or methods that change the training strategy or training cost
for improvement of model accuracy is termed as Bag of Freebies.
Bag of Specials (BoS) can be considered as an add-on for any object detectors
present right now to make them more accurate.
33. Raspberry pi 3B+
The Raspberry Pi 3 Model B+ is the latest product in the Raspberry Pi 3
range, boasting a 64-bit quad core processor running at 1.4GHz, dual-band
2.4GHz. and 5GHz wireless LAN, Bluetooth 4.2/BLE
34. Camera module v2
The Raspberry Pi Camera v2 is a high quality 4 mega pixel image
sensor custom designed add-on board for Raspberry Pi, featuring a fixed
focus lens.
35. Result
• In fig our device identifies the objects by classes
assigned to each object by its tag and has
dimensions on detected image.
• Detection Speeds: You can reduce the time it takes
to detect an clear image by setting the speed of
detection speed to “fast”, “faster” and “fastest”.
• As a result this device gives output in 2sec to 5 sec.
• Accuracy of the output is tested on 250 images and
there is different accuracy on different types of
image quality and position.
• The average accuracy of the device is 90%.
• According to types of image the accuracy showed
in table:-
Images
Clear
images
Blurred
images
Void
background
Conjusted
background
Near images 97% 96% 90% 88%
Far images 92% 89% 86% 82%
Fig: it shows the output that is detected
37. Advantage of yolo algorithm
YOLO algorithm is important because of the following reasons:
Speed : This algorithm improves the speed of detection because it can predict
objects in real-time.
High accuracy: YOLO is a predictive technique that provides accurate results. It
use Convolutional implementation that means that if you have 3*3 grid (i.e.,
divide image into 9 grid cells) then you don't need to run the algorithm 9 times to
validate presence of object in each grid cell rather this is one single convolutional
implementation.
Learning capabilities: The algorithm has excellent learning capabilities that
enable it to learn the representations of objects and apply them in object
detection.
38. Advantage
This work is implemented using GTTS.
Easy to set up.
Open source tools were used for this project.
Cheap and cost-efficient.
This project will work on device only no need to buy any extra things.
39. Conclusion
Simple Indian object recognition system
based on yolo algorithm has been
proposed.
The system has been written in OpenCV.
40. Future Work
Enhancing the accuracy by building a model of features for each object
class.
Working now on using local features instead of template matching
Enhancing the best frame to be processed for runtime application
Adding more objects to the database.