This paper evaluates the effects of input image degradation on performance of image object detectors. The purpose of the evaluation is to determine usability of the detectors trained on original images in adverse conditions. SSD and Faster R-CNN based pedestrian and cyclist detector performance with images degraded with motion blur, out-of-focus blur, and JPEG compression artefacts, most commonly occurring in mobile or static traffic systems. An experiment was designed to assess the effect of degradations on detection precision and cross class confusion. The paper describes the two datasets created for this evaluation, evaluation of a number of detectors on increasingly more degraded images, comparison of their performance, and assessment of their tolerance to different types of image degradation as well as a discussion of the results.
An Analysis of Various Deep Learning Algorithms for Image Processingvivatechijri
Various applications of image processing has given it a wider scope when it comes to data analysis.
Various Machine Learning Algorithms provide a powerful environment for training modules effectively to
identify various entities of images and segment the same accordingly. Rather one can observe that though the
image classifiers like the Support Vector Machines (SVM) or Random Forest Algorithms do justice to the task,
deep learning algorithms like the Artificial Neural Networks (ANN) and its subordinates, the very well-known
and extremely powerful Algorithm Convolution Neural Networks (CNN) can provide a new dimension to the
image processing domain. It has way higher accuracy and computational power for classifying images further
and segregating their various entities as individual components of the image working region. Major focus will
be on the Region Convolution Neural Networks (R-CNN) algorithm and how well it provides the pixel-level
segmentation further using its better successors like the Fast-Faster and Mask R-CNN versions.
VEHICLE CLASSIFICATION USING THE CONVOLUTION NEURAL NETWORK APPROACHJANAK TRIVEDI
We present vehicle detection classification using the Convolution
Neural Network (CNN) of the deep learning approach. The automatic vehicle
classification for traffic surveillance video systems is challenging for the Intelligent
Transportation System (ITS) to build a smart city. In this article, three different
vehicles: bike, car and truck classification are considered for around 3,000 bikes,
6,000 cars, and 2,000 images of trucks. CNN can automatically absorb and extract
different vehicle dataset’s different features without a manual selection of features.
The accuracy of CNN is measured in terms of the confidence values of the detected
object. The highest confidence value is about 0.99 in the case of the bike category
vehicle classification. The automatic vehicle classification supports building an
electronic toll collection system and identifying emergency vehicles in the traffic
Cracks on the concrete surface are one of the earliest symptoms of degradation of the structure which isfundamental for the upkeep as properly the non-stop publicity will lead to the severe injury to the environment.Manual inspection is the acclaimed approach for the crack inspection. In the guide inspection, the diagram of thecrack is organized manually, and the conditions of the irregularities are noted. Since the guide strategy absolutelyrelies upon on the specialist’s expertise and experience, it lacks objectivity in the quantitative analysis. So,automated image-based crack detection is proposed as a replacement. The proposed gadget comprises pictureprocessing and facts acquisition methodologies for crack detection and evaluation of surface degradation. Theacquired outcomes exhibit that the deployment of image processing in an nice way is a key step towards theinspection of giant infrastructures
Visual odometry & slam utilizing indoor structured environmentsNAVER Engineering
Visual odometry (VO) and simultaneous localization and mapping (SLAM) are fundamental building blocks for various applications from autonomous vehicles to virtual and augmented reality (VR/AR).
To improve the accuracy and robustness of the VO & SLAM approaches, we exploit multiple lines and orthogonal planar features, such as walls, floors, and ceilings, common in man-made indoor environments.
We demonstrate the effectiveness of the proposed VO & SLAM algorithms through an extensive evaluation on a variety of RGB-D datasets and compare with other state-of-the-art methods.
Classification and Detection of Vehicles using Deep Learningijtsrd
The vehicle classification and detecting its license plate are important tasks in intelligent security and transportation systems. The traditional methods of vehicle classification and detection are highly complex which provides coarse grained results due to suffering from limited viewpoints. Because of the latest achievements of Deep Learning, it was successfully applied to image classification and detection of objects. This paper presents a method based on a convolutional neural network, which consists of two steps vehicle classification and vehicle license plate recognition. Several typical neural network modules have been applied in training and testing the vehicle Classification and detection of license plate model, such as CNN convolutional neural networks , TensorFlow, Tesseract OCR. The proposed method can identify the vehicle type, number plate and other information accurately. This model provides security and log details regarding vehicles by using AI Surveillance. It guides the surveillance operators and assists human resources. With the help of the original training dataset and enriched testing dataset, the algorithm can obtain results with an average accuracy of about 97.32 in the classification and detection of vehicles. By increasing the amount of the data, the mean error and misclassification rate gradually decreases. So, this algorithm which is based on Deep Learning has good superiority and adaptability. When compared to the leading methods in the challenging Image datasets, our deep learning approach obtains highly competitive results. Finally, this paper proposes modern methods for the improvement of the algorithm and prospects the development direction of deep learning in the field of machine learning and artificial intelligence. Madde Pavan Kumar | Dr. K. Manivel | N. Jayanthi "Classification & Detection of Vehicles using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-3 , April 2020, URL: https://www.ijtsrd.com/papers/ijtsrd30353.pdf Paper Url :https://www.ijtsrd.com/engineering/software-engineering/30353/classification-and-detection-of-vehicles-using-deep-learning/madde-pavan-kumar
An Analysis of Various Deep Learning Algorithms for Image Processingvivatechijri
Various applications of image processing has given it a wider scope when it comes to data analysis.
Various Machine Learning Algorithms provide a powerful environment for training modules effectively to
identify various entities of images and segment the same accordingly. Rather one can observe that though the
image classifiers like the Support Vector Machines (SVM) or Random Forest Algorithms do justice to the task,
deep learning algorithms like the Artificial Neural Networks (ANN) and its subordinates, the very well-known
and extremely powerful Algorithm Convolution Neural Networks (CNN) can provide a new dimension to the
image processing domain. It has way higher accuracy and computational power for classifying images further
and segregating their various entities as individual components of the image working region. Major focus will
be on the Region Convolution Neural Networks (R-CNN) algorithm and how well it provides the pixel-level
segmentation further using its better successors like the Fast-Faster and Mask R-CNN versions.
VEHICLE CLASSIFICATION USING THE CONVOLUTION NEURAL NETWORK APPROACHJANAK TRIVEDI
We present vehicle detection classification using the Convolution
Neural Network (CNN) of the deep learning approach. The automatic vehicle
classification for traffic surveillance video systems is challenging for the Intelligent
Transportation System (ITS) to build a smart city. In this article, three different
vehicles: bike, car and truck classification are considered for around 3,000 bikes,
6,000 cars, and 2,000 images of trucks. CNN can automatically absorb and extract
different vehicle dataset’s different features without a manual selection of features.
The accuracy of CNN is measured in terms of the confidence values of the detected
object. The highest confidence value is about 0.99 in the case of the bike category
vehicle classification. The automatic vehicle classification supports building an
electronic toll collection system and identifying emergency vehicles in the traffic
Cracks on the concrete surface are one of the earliest symptoms of degradation of the structure which isfundamental for the upkeep as properly the non-stop publicity will lead to the severe injury to the environment.Manual inspection is the acclaimed approach for the crack inspection. In the guide inspection, the diagram of thecrack is organized manually, and the conditions of the irregularities are noted. Since the guide strategy absolutelyrelies upon on the specialist’s expertise and experience, it lacks objectivity in the quantitative analysis. So,automated image-based crack detection is proposed as a replacement. The proposed gadget comprises pictureprocessing and facts acquisition methodologies for crack detection and evaluation of surface degradation. Theacquired outcomes exhibit that the deployment of image processing in an nice way is a key step towards theinspection of giant infrastructures
Visual odometry & slam utilizing indoor structured environmentsNAVER Engineering
Visual odometry (VO) and simultaneous localization and mapping (SLAM) are fundamental building blocks for various applications from autonomous vehicles to virtual and augmented reality (VR/AR).
To improve the accuracy and robustness of the VO & SLAM approaches, we exploit multiple lines and orthogonal planar features, such as walls, floors, and ceilings, common in man-made indoor environments.
We demonstrate the effectiveness of the proposed VO & SLAM algorithms through an extensive evaluation on a variety of RGB-D datasets and compare with other state-of-the-art methods.
Classification and Detection of Vehicles using Deep Learningijtsrd
The vehicle classification and detecting its license plate are important tasks in intelligent security and transportation systems. The traditional methods of vehicle classification and detection are highly complex which provides coarse grained results due to suffering from limited viewpoints. Because of the latest achievements of Deep Learning, it was successfully applied to image classification and detection of objects. This paper presents a method based on a convolutional neural network, which consists of two steps vehicle classification and vehicle license plate recognition. Several typical neural network modules have been applied in training and testing the vehicle Classification and detection of license plate model, such as CNN convolutional neural networks , TensorFlow, Tesseract OCR. The proposed method can identify the vehicle type, number plate and other information accurately. This model provides security and log details regarding vehicles by using AI Surveillance. It guides the surveillance operators and assists human resources. With the help of the original training dataset and enriched testing dataset, the algorithm can obtain results with an average accuracy of about 97.32 in the classification and detection of vehicles. By increasing the amount of the data, the mean error and misclassification rate gradually decreases. So, this algorithm which is based on Deep Learning has good superiority and adaptability. When compared to the leading methods in the challenging Image datasets, our deep learning approach obtains highly competitive results. Finally, this paper proposes modern methods for the improvement of the algorithm and prospects the development direction of deep learning in the field of machine learning and artificial intelligence. Madde Pavan Kumar | Dr. K. Manivel | N. Jayanthi "Classification & Detection of Vehicles using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-3 , April 2020, URL: https://www.ijtsrd.com/papers/ijtsrd30353.pdf Paper Url :https://www.ijtsrd.com/engineering/software-engineering/30353/classification-and-detection-of-vehicles-using-deep-learning/madde-pavan-kumar
Advertisement billboard detection and geotagging system with inductive transf...TELKOMNIKA JOURNAL
In this paper, we propose an approach to detect and geotag advertisement billboard in real-time
condition. Our approach is using AlexNet’s Deep Convolutional Neural Network (DCNN) as a pre-trained
neural network with 1000 categories for image classification. To improve the performance of
the pre-trained neural network, we retrain the network by adding more advertisement billboard images
using inductive transfer learning approach. Then, we fine-tuned the output layer into advertisement
billboard related categories. Furthermore, the detected advertisement billboard images will be geotagged
by inserting Exif metadata into the image file. Experimental results show that the approach achieves
92.7% training accuracy for advertisement billboard detection, while for overall testing results it will give
71.86% testing accuracy.
Deconstructing SfM-Net architecture and beyond
"SfM-Net, a geometry-aware neural network for motion estimation in videos that decomposes frame-to-frame pixel motion in terms of scene and object depth, camera motion and 3D object rotations and translations. Given a sequence of frames, SfM-Net predicts depth, segmentation, camera and rigid object motions, converts those into a dense frame-to-frame motion field (optical flow), differentiably warps frames in time to match pixels and back-propagates."
Alternative download:
https://www.dropbox.com/s/aezl7ro8sy2xq7j/sfm_net_v2.pdf?dl=0
COMPARATIVE STUDY ON VEHICLE DETECTION TECHNIQUES IN AERIAL SURVEILLANCEIJCI JOURNAL
Aerial surveillance system becomes a great trendy on past decades. Aerial surveillance vehicle tracking techniques plays a vital role and give rising to optimistic techniques continuously. This system can be very handy in various applications such as police, traffic monitoring, natural disaster and military. It is often covers large area and providing better perspective of moving objects. The detection of moving vehicle can be both from the dynamic aerial imagery, wide area motion imagery or images under low resolution and also the static in nature. It has been very difficult issue whether identify the object in the air view, the camera angles, movement objects and motionless object. This paper deals with comparative study on various vehicle detection and tracking approach in aerial videos with its experimental results and measures working condition, hit rate and false alarm rate
Lecture 1 for Digital Image Processing (2nd Edition)Moe Moe Myint
-What is Digital Image Processing?
-The Origins of Digital Image Processing
-Examples of Fields that Use Digital Image Processing
-Fundamentals Steps in Digital Image Processing
-Components of an Image Processing System
A new Compton scattered tomography modality and its application to material n...irjes
Imaging modalities exploiting the use of Compton scattering are currently under active investigation. However, despite many innovative contributions, this topic still poses a formidable mathematical and technical challenge. Due to the very particular nature of the Compton effect, the main problem consists of obtaining the reconstruction of the object electron density. Investigations on Compton scatter imaging for biological tissues, organs and the like have been performed and studied widely over the years. However in material sciences, in particular in non-destructive evaluation and control, this type of imaging procedure is just at its beginning. In this paper, we present a new scanning process which collects scattered radiation to reconstruct the internal electronic distribution of industrial materials. As an illustration, we shall look at one of the most widely used construction material: concrete and its variants in civil engineering. The Compton scattered radiation approach is particularly efficient in imaging steel frame and voids imbedded in bulk concrete objects.
We present numerical simulation results to demonstrate the viability and performances of this imaging modality.
Keywords :- Compton scattering , Gamma-ray imaging , Non-destructive testing/evaluation (NDT/NDE), Concrete: structure and defects, Radon transform
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHMcsandit
Computer vision approaches are increasingly used in mobile robotic systems, since they allow
to obtain a very good representation of the environment by using low-power and cheap sensors.
In particular it has been shown that they can compete with standard solutions based on laser
range scanners when dealing with the problem of simultaneous localization and mapping
(SLAM), where the robot has to explore an unknown environment while building a map of it and
localizing in the same map. We present a package for simultaneous localization and mapping in
ROS (Robot Operating System) using a monocular camera sensor only. Experimental results in
real scenarios as well as on standard datasets show that the algorithm is able to track the
trajectory of the robot and build a consistent map of small environments, while running in near
real-time on a standard PC.
[Paper research] GOSELO: for Robot navigation using Reactive neural networksJehong Lee
GOSELO: Goal-Direction Obstacle and Self-Location Map for Robot Navigation Using Reactive Neural Networks 라는 논문을 중심으로, mobile platform의 Path planning을 CNN으로 End-to-End 방식으로 수행하는 방법에 관하여 소개합니다.
광주과학기술원 인공지능 스터디 A-GIST 모임에서 발표했습니다.
발표영상(유튜브, 한국어): https://youtu.be/l-gKjzWKuHA
Traffic Accident Detection
The Problem is to build a reliable and accurate traffic accident detection system using state-of-the-art object detection models. The system should be able to analyze live video feeds from CCTV cameras and determine whether an accident has occurred or not, based on the presence of specific visual cues such as collision, and vehicle damage. The system's performance will be evaluated based on accuracy, precision, recall, and computational efficiency.
Faster R-CNN (ResNet-50)
YOLO_V5_ Overview
SSD_mobilenet_v2_320*320
Advertisement billboard detection and geotagging system with inductive transf...TELKOMNIKA JOURNAL
In this paper, we propose an approach to detect and geotag advertisement billboard in real-time
condition. Our approach is using AlexNet’s Deep Convolutional Neural Network (DCNN) as a pre-trained
neural network with 1000 categories for image classification. To improve the performance of
the pre-trained neural network, we retrain the network by adding more advertisement billboard images
using inductive transfer learning approach. Then, we fine-tuned the output layer into advertisement
billboard related categories. Furthermore, the detected advertisement billboard images will be geotagged
by inserting Exif metadata into the image file. Experimental results show that the approach achieves
92.7% training accuracy for advertisement billboard detection, while for overall testing results it will give
71.86% testing accuracy.
Deconstructing SfM-Net architecture and beyond
"SfM-Net, a geometry-aware neural network for motion estimation in videos that decomposes frame-to-frame pixel motion in terms of scene and object depth, camera motion and 3D object rotations and translations. Given a sequence of frames, SfM-Net predicts depth, segmentation, camera and rigid object motions, converts those into a dense frame-to-frame motion field (optical flow), differentiably warps frames in time to match pixels and back-propagates."
Alternative download:
https://www.dropbox.com/s/aezl7ro8sy2xq7j/sfm_net_v2.pdf?dl=0
COMPARATIVE STUDY ON VEHICLE DETECTION TECHNIQUES IN AERIAL SURVEILLANCEIJCI JOURNAL
Aerial surveillance system becomes a great trendy on past decades. Aerial surveillance vehicle tracking techniques plays a vital role and give rising to optimistic techniques continuously. This system can be very handy in various applications such as police, traffic monitoring, natural disaster and military. It is often covers large area and providing better perspective of moving objects. The detection of moving vehicle can be both from the dynamic aerial imagery, wide area motion imagery or images under low resolution and also the static in nature. It has been very difficult issue whether identify the object in the air view, the camera angles, movement objects and motionless object. This paper deals with comparative study on various vehicle detection and tracking approach in aerial videos with its experimental results and measures working condition, hit rate and false alarm rate
Lecture 1 for Digital Image Processing (2nd Edition)Moe Moe Myint
-What is Digital Image Processing?
-The Origins of Digital Image Processing
-Examples of Fields that Use Digital Image Processing
-Fundamentals Steps in Digital Image Processing
-Components of an Image Processing System
A new Compton scattered tomography modality and its application to material n...irjes
Imaging modalities exploiting the use of Compton scattering are currently under active investigation. However, despite many innovative contributions, this topic still poses a formidable mathematical and technical challenge. Due to the very particular nature of the Compton effect, the main problem consists of obtaining the reconstruction of the object electron density. Investigations on Compton scatter imaging for biological tissues, organs and the like have been performed and studied widely over the years. However in material sciences, in particular in non-destructive evaluation and control, this type of imaging procedure is just at its beginning. In this paper, we present a new scanning process which collects scattered radiation to reconstruct the internal electronic distribution of industrial materials. As an illustration, we shall look at one of the most widely used construction material: concrete and its variants in civil engineering. The Compton scattered radiation approach is particularly efficient in imaging steel frame and voids imbedded in bulk concrete objects.
We present numerical simulation results to demonstrate the viability and performances of this imaging modality.
Keywords :- Compton scattering , Gamma-ray imaging , Non-destructive testing/evaluation (NDT/NDE), Concrete: structure and defects, Radon transform
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHMcsandit
Computer vision approaches are increasingly used in mobile robotic systems, since they allow
to obtain a very good representation of the environment by using low-power and cheap sensors.
In particular it has been shown that they can compete with standard solutions based on laser
range scanners when dealing with the problem of simultaneous localization and mapping
(SLAM), where the robot has to explore an unknown environment while building a map of it and
localizing in the same map. We present a package for simultaneous localization and mapping in
ROS (Robot Operating System) using a monocular camera sensor only. Experimental results in
real scenarios as well as on standard datasets show that the algorithm is able to track the
trajectory of the robot and build a consistent map of small environments, while running in near
real-time on a standard PC.
[Paper research] GOSELO: for Robot navigation using Reactive neural networksJehong Lee
GOSELO: Goal-Direction Obstacle and Self-Location Map for Robot Navigation Using Reactive Neural Networks 라는 논문을 중심으로, mobile platform의 Path planning을 CNN으로 End-to-End 방식으로 수행하는 방법에 관하여 소개합니다.
광주과학기술원 인공지능 스터디 A-GIST 모임에서 발표했습니다.
발표영상(유튜브, 한국어): https://youtu.be/l-gKjzWKuHA
Traffic Accident Detection
The Problem is to build a reliable and accurate traffic accident detection system using state-of-the-art object detection models. The system should be able to analyze live video feeds from CCTV cameras and determine whether an accident has occurred or not, based on the presence of specific visual cues such as collision, and vehicle damage. The system's performance will be evaluated based on accuracy, precision, recall, and computational efficiency.
Faster R-CNN (ResNet-50)
YOLO_V5_ Overview
SSD_mobilenet_v2_320*320
Real Time Object Identification for Intelligent Video Surveillance ApplicationsEditor IJCATR
Intelligent video surveillance system has emerged as a very important research topic in the computer vision field in the
recent years. It is well suited for a broad range of applications such as to monitor activities at traffic intersections for detecting
congestions and predict the traffic flow. Object classification in the field of video surveillance is a key component of smart
surveillance software. Two robust methodology and algorithms adopted for people and object classification for automated surveillance
systems is proposed in this paper. First method uses background subtraction model for detecting the object motion. The background
subtraction and image segmentation based on morphological transformation for tracking and object classification on highways is
proposed. This algorithm uses erosion followed by dilation on various frames. Proposed algorithm in first method, segments the image
by preserving important edges which improves the adaptive background mixture model and makes the system learn faster and more
accurately. The system used in second method adopts the object detection method without background subtraction because of the static
object detection. Segmentation is done by the bounding box registration technique. Then the classification is done with the multiclass
SVM using the edge histogram as features. The edge histograms are calculated for various bin values in different environment. The
result obtained demonstrates the effectiveness of the proposed approach.
Traffic Congestion Detection Using Deep Learningijtsrd
Despite the huge amount of traffic surveillance videos and images have been accumulated in the daily monitoring, deep learning approaches have been underutilized in the application of traffic intelligent management and control. Traffic images, including various illumination, weather conditions, and vast scenarios are considered and preprocessed to set up a proper training dataset. In order to detect traffic congestion, a network structure is proposed based on residual learning to be pre trained and fine tuned. The network is then transferred to the traffic application and retrained with self established training dataset to generate the Traffic Net. The accuracy of Traffic Net to classify congested and uncongested road states reaches 99 for the validation dataset and 95 for the testing dataset. The trained model is stored in cloud storage for easy access for application from anywhere. The proposed Traffic Net can be used by a regional detection of traffic congestion on a large scale surveillance system. The effectiveness and efficiencies are magnificently demonstrated with quick detection in the high accuracy in the case study. The experimental trial could extend its successful application to traffic surveillance system and has potential enhancement for intelligent transport system in future. Anusha C | Dr. J. Bhuvana "Traffic Congestion Detection Using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-2 , February 2022, URL: https://www.ijtsrd.com/papers/ijtsrd49401.pdf Paper URL: https://www.ijtsrd.com/computer-science/other/49401/traffic-congestion-detection-using-deep-learning/anusha-c
Automated Traffic sign board classification system is one of the key technologies of Intelligent
Transportation Systems (ITS). Traffic Surveillance System is being more and important with improving
urban scale and increasing number of vehicles. This Paper presents an intelligent sign board
classification method based on blob analysis in traffic surveillance. Processing is done by three main
steps: moving object segmentation, blob analysis, and classifying. A Sign board is modelled as a
rectangular patch and classified via blob analysis. By processing the blob of sign boards, the meaningful
features are extracted. Tracking moving targets is achieved by comparing the extracted features with
training data. After classifying the sign boards the system will intimate to user in the form of alarms,
sound waves. The experimental results show that the proposed system can provide real-time and useful
information for traffic surveillance.
Noise Removal in Traffic Sign Detection SystemsCSEIJJournal
The application of Traffic sign detection and recognition is growing in traffic assistant driving systems and
automatic driving systems. It helps drivers and automatic driving systems to detect and recognize the
traffic signs effectively. However, it is found that it may be difficult for these systems to work in challenging
environments like rain, haze, hue, etc. To help the detection systems to have better performance in
challenging conditions like rain and haze, we propose the use of a deep learning technique based on a
Convolutional Neural Network to process visual data. The processed data could be used in the detection.
We are using the NoiseNet model [11], a noise reduction network for our architecture. The model is
trained to enhance images in patches instead of as a whole. The training is done using the Challenging
Unreal and Real Environment - Traffic Sign Detection Dataset(CURE-TSD) which contains videos of
different roads in various challenging situations. The
Semantic Concept Detection in Video Using Hybrid Model of CNN and SVM Classif...CSCJournals
In today's era of digitization and fast internet, many video are uploaded on websites, a mechanism is required to access this video accurately and efficiently. Semantic concept detection achieve this task accurately and is used in many application like multimedia annotation, video summarization, annotation, indexing and retrieval. Video retrieval based on semantic concept is efficient and challenging research area. Semantic concept detection bridges the semantic gap between low level extraction of features from key-frame or shot of video and high level interpretation of the same as semantics. Semantic Concept detection automatically assigns labels to video from predefined vocabulary. This task is considered as supervised machine learning problem. Support vector machine (SVM) emerged as default classifier choice for this task. But recently Deep Convolutional Neural Network (CNN) has shown exceptional performance in this area. CNN requires large dataset for training. In this paper, we present framework for semantic concept detection using hybrid model of SVM and CNN. Global features like color moment, HSV histogram, wavelet transform, grey level co-occurrence matrix and edge orientation histogram are selected as low level features extracted from annotated groundtruth video dataset of TRECVID. In second pipeline, deep features are extracted using pretrained CNN. Dataset is partitioned in three segments to deal with data imbalance issue. Two classifiers are separately trained on all segments and fusion of scores is performed to detect the concepts in test dataset. The system performance is evaluated using Mean Average Precision for multi-label dataset. The performance of the proposed framework using hybrid model of SVM and CNN is comparable to existing approaches.
Inclined Image Recognition for Aerial Mapping using Deep Learning and Tree ba...TELKOMNIKA JOURNAL
One of the important capabilities of an unmanned aerial vehicle (UAV) is aerial mapping. Aerial mapping is an image registration problem, i.e., the problem of transforming different sets of images into one coordinate system. In image registration, the quality of the output is strongly influenced by the quality of input (i.e., images captured by the UAV). Therefore, selecting the quality of input images becomes important and one of the challenging task in aerial mapping because the ground truth in the mapping process is not given before the UAV flies. Typically, UAV takes images in sequence irrespective of its flight orientation and roll angle. These may result in the acquisition of bad quality images, possibly compromising the quality of mapping results, and increasing the computational cost of a registration process. To address these issues, we need a recognition system that is able to recognize images that are not suitable for the registration process. In this paper, we define these unsuitable images as “inclined images,” i.e., images captured by UAV that are not perpendicular to the ground. Although we can calculate the inclination angle using a gyroscope attached to the UAV, our interest here is to recognize these inclined images without the use of additional sensors in order to mimic how humans perform this task visually. To realize that, we utilize a deep learning method with the combination of tree-based models to build an inclined image recognition system. We have validated the proposed system with the images captured by the UAV. We collected 192 images and labelled them with two different levels of classes (i.e., coarse- and fine-classification). We compared this with several models and the results showed that our proposed system yielded an improvement of accuracy rate up to 3%.
NUMBER PLATE IMAGE DETECTION FOR FAST MOTION VEHICLES USING BLUR KERNEL ESTIM...paperpublications3
Abstract: Many recent advancements have been introduced in both hardware and software technologies. Out of these technologies, we gain a lot of interest in Image Processing. As the eccentric identification of a vehicle, number plate is a key clue to discover theft and over-speed vehicles. The captured images from the camera are always in low resolution and suffer severe loss of edge information, which cast great challenge to existing blind deblurring methods. The blur kernel can be showed as linear uniform convolution and with angle and length estimation. In this paper, sparse representation is used to identify the blur kernel. Then, the length of the motion kernel has been estimated with Radon transform in Fourier domain. We evaluate our approach on real-world images and compare with several popular blind image deblurring algorithms. Based on the results obtained the supremacy of our proposed approach in terms of efficacy.
Keywords: Image Segmentation, Angle Estimation, Length Estimation, Deconvolution Algorithm, Artificial Neural Network.
Title: NUMBER PLATE IMAGE DETECTION FOR FAST MOTION VEHICLES USING BLUR KERNEL ESTIMATION AND ANN
Author: S.KALAIVANI, K.PRAVEENA, T.PREETHI, N.PUNITHA
ISSN 2350-1022
International Journal of Recent Research in Mathematics Computer Science and Information Technology
Paper Publications
Accident vehicle types classification: a comparative study between different...nooriasukmaningtyas
Classifying and finding type of individual vehicles within an accident image are considered difficult problems. This research concentrates on accurately classifying and recognizing vehicle accidents in question. The aim to provide a comparative analysis of vehicle accidents. A number of network topologies are tested to arrive at convincing results and a variety of matrices are used in the evaluation process to identify the best networks. The best two networks are used with faster recurrent convolution neural network (Faster RCNN) and you only look once (YOLO) to determine which network will identifiably detect the location and type of the vehicle. In addition, two datasets are used in this research. In consequence, experiment results show that MobileNetV2 and ResNet50 have accomplished higher accuracy compared to the rest of the models, with 89.11% and 88.45% for the GAI dataset as well as 88.72% and 89.69% for KAI dataset, respectively. The findings reveal that the ResNet50 base network for YOLO achieved higher accuracy than MobileNetV2 for YOLO, ResNet50 for Faster RCNN with 83%, 81%, and 79% for GAI dataset and 79%, 78% and 74% for KAI dataset.
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofsAlex Pruden
This paper presents Reef, a system for generating publicly verifiable succinct non-interactive zero-knowledge proofs that a committed document matches or does not match a regular expression. We describe applications such as proving the strength of passwords, the provenance of email despite redactions, the validity of oblivious DNS queries, and the existence of mutations in DNA. Reef supports the Perl Compatible Regular Expression syntax, including wildcards, alternation, ranges, capture groups, Kleene star, negations, and lookarounds. Reef introduces a new type of automata, Skipping Alternating Finite Automata (SAFA), that skips irrelevant parts of a document when producing proofs without undermining soundness, and instantiates SAFA with a lookup argument. Our experimental evaluation confirms that Reef can generate proofs for documents with 32M characters; the proofs are small and cheap to verify (under a second).
Paper: https://eprint.iacr.org/2023/1886
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Enhancing Performance with Globus and the Science DMZGlobus
ESnet has led the way in helping national facilities—and many other institutions in the research community—configure Science DMZs and troubleshoot network issues to maximize data transfer performance. In this talk we will present a summary of approaches and tips for getting the most out of your network infrastructure using Globus Connect Server.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Welcome to the first live UiPath Community Day Dubai! Join us for this unique occasion to meet our local and global UiPath Community and leaders. You will get a full view of the MEA region's automation landscape and the AI Powered automation technology capabilities of UiPath. Also, hosted by our local partners Marc Ellis, you will enjoy a half-day packed with industry insights and automation peers networking.
📕 Curious on our agenda? Wait no more!
10:00 Welcome note - UiPath Community in Dubai
Lovely Sinha, UiPath Community Chapter Leader, UiPath MVPx3, Hyper-automation Consultant, First Abu Dhabi Bank
10:20 A UiPath cross-region MEA overview
Ashraf El Zarka, VP and Managing Director MEA, UiPath
10:35: Customer Success Journey
Deepthi Deepak, Head of Intelligent Automation CoE, First Abu Dhabi Bank
11:15 The UiPath approach to GenAI with our three principles: improve accuracy, supercharge productivity, and automate more
Boris Krumrey, Global VP, Automation Innovation, UiPath
12:15 To discover how Marc Ellis leverages tech-driven solutions in recruitment and managed services.
Brendan Lingam, Director of Sales and Business Development, Marc Ellis
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
State of ICS and IoT Cyber Threat Landscape Report 2024 preview
Performance Evaluation of CNN Based Pedestrian and Cyclist Detectors On Degraded Images
1. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 1
Performance Evaluation of CNN Based Pedestrian and Cyclist
Detectors On Degraded Images
Tomáš Zemčík zemcikt@feec.vutbr.cz
Faculty of Electrical Engineering and Communications
Dept. of Automation and Instrumentation
Brno University of Technology, Brno, Czech Republic
Lukáš Kratochvíla kratochvila@feec.vutbr.cz
Faculty of Electrical Engineering and Communications
Dept. of Automation and Instrumentation
Brno University of Technology, Brno, Czech Republic
Šimon Bilík bilik@feec.vutbr.cz
Faculty of Electrical Engineering and Communications
Dept. of Automation and Instrumentation
Brno University of Technology, Brno, Czech Republic
Ondřej Boštík bostik@feec.vutbr.cz
Faculty of Electrical Engineering and Communications
Dept. of Automation and Instrumentation
Brno University of Technology, Brno, Czech Republic
Pavel Zemčík zemcik@fit.vutbr.cz
Faculty of Information Technology
Dept. of Computer Graphics and Multimedia
Brno University of Technology, Brno, Czech Republic
Karel Horák horak@feec.vutbr.cz
Faculty of Electrical Engineering and Communications
Dept. of Automation and Instrumentation
Brno University of Technology, Brno, Czech Republic
Abstract
This paper evaluates the effects of input image degradation on performance of image object
detectors. The purpose of the evaluation is to determine usability of the detectors trained on
original images in adverse conditions. SSD and Faster R-CNN based pedestrian and cyclist
detector performance with images degraded with motion blur, out-of-focus blur, and JPEG
compression artefacts, most commonly occurring in mobile or static traffic systems. An
experiment was designed to assess the effect of degradations on detection precision and cross
class confusion. The paper describes the two datasets created for this evaluation, evaluation of a
number of detectors on increasingly more degraded images, comparison of their performance,
and assessment of their tolerance to different types of image degradation as well as a discussion
of the results.
Keywords: Object Detection, Image Degradation, Pedestrian Detection, Cyclist Detection, SSD,
Faster R-CNN.
1. INTRODUCTION
Image object detection is one of the keys to solving technical problems in machine vision-based
traffic application. CNN object detection architectures currently seem to be the most promising
ones towards robust and reliable image object detection in very difficult traffic environments. With
2. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 2
the latest advances in this field, the pressures of regulatory bodies to implement them into traffic
safety improving applications is higher than ever, as demonstrated by EU’s “Vision Zero” plan to
have zero road accident deaths by 2050 [1]. Pedestrians and cyclists represent the most
vulnerable road users making up for approximately 26% of all road accident-related deaths in
2018 according to WHO [2]. In the EU this number is even higher, at 32%, so it is no surprise that
detecting them is a central issue in Autonomous Vehicles (AV), Advanced Driver Assistance
Systems (ADAS), Intelligent Traffic Systems (ITS), and Intelligent Vehicle Systems (IVS). With
the shift towards more automated transportation a lot of research in this area is being done, such
as [3 - 5].
From the semantic point of view, both pedestrians and cyclists are very similar, as both classes
represent humans in different poses with or without a bicycle. Therefore, their appearance in
images is very similar. This similarity means that common features exist and can aid detector
training. On the other hand, while similarities exist across classes, also an enormous variety
occurs within individual instances of each class making distinguishing between pedestrians and
cyclists, especially in certain orientations, a challenging task [6].
Despite the aforementioned similarity in appearance, however, for a higher-level AV or ADAS
system, they represent dramatically different objects that have very different behavioural patterns
in traffic and as such need to be dealt with very differently in the trajectory planning algorithms,
thus it is needed to have a robust and accurate detector.
2. OBJECT DETECTION IN TRAFFIC APPLICATIONS
Since the publication of pedestrian targeted Viola Jones detector in 2003 [7], which is by many
regarded the first pedestrian-in-image detector [8], major advances in the field have been made.
In consequent years, the attention has shifted from hand-selected Haar-like features and boosted
classifiers through HOG features using SVM classifiers [9, 10] to deformable part models [11]. In
recent years, however, the most attention is concentrated on Deep Learning models based
around CNNs. The DL approach finally allows for the feature extraction to be trained along with
the classifiers [8].
Up until the rise of CNNs, detection methods of pedestrians and cyclists (and other classes) have
been largely kept separate requiring multiple runs of different algorithms on a single image to
detect various classes of road users [12]. With the capabilities of modern CNN meta-
architectures, no reason not to implement multiclass capability into today’s detectors exists
anymore and with deployed AV and ADAS, this is a de-facto standard [12].
In practical deployment, ideal conditions rarely apply and many applications need to perform well
even with poor input image quality. This paper evaluates performance of state-of-the-art object
detection architectures on varying quality datasets. We do not intend to create a detector to
compete with the state-of-the-art detectors.
Effects of image degradation on various CNN classification architectures were previously studied
[13, 14]. With results showing that while all tested architectures were affected, not all are affected
by the same way, with Roy at al. [13] concluding deeper networks are affected more. Pei et al.
[14] concluded that while algorithms exist for removing the tested degradations from image
(motion blur and haze), the improvement in CNN classification performance was insignificant. The
effect of image degradation on object detection networks has not been that extensively studied.
Single Shot multibox Detector (SSD) and Faster Region-based Convolutional Neural Network (R-
CNN) have been selected to be the candidates based on performance as measured by Tilgner
[15].
3. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 3
2.1 R-CNN Object Detector Family
Girshick et al. have presented their Region-based Convolutional Neural Network (R-CNN) in 2014
[16]. The R-CNN workflow consists of 3 steps, where in the first one, an input image is divided
into 2000 regions-of-interest (RoI). Consequently, all 2000 subimages are run through a CNN
feature extraction and finally classified via SVM.
FIGURE 1: R-CNN Workflow [16].
R-CNN outperformed most of its contemporaries in terms of accuracy but running the CNN on all
RoIs separately is obviously rather computationally expensive (tens of seconds per image), so
primarily to address this drawback, Girshick introduced Fast R-CNN [17] in 2015.
Fast R-CNN improves upon the original method by only running the CNN once on each input
image generating a convolutional feature map, by adding bounding box regression into the
network, and then by resizing the selected region of interest in the feature map to be classified
using a fully connected neural network. These improvements managed to improve computational
performance by an order of magnitude without any significant loss in detection accuracy [17].
In 2016 Ren et al. published another major improvement in the R-CNN object detector family - the
Faster R-CNN [18]. Similarly to Fast R-CNN, the convolutional feature map is only computed
once per image but unlike in the above method, the region proposals are handled by a custom
Region Proposal Network (RPN). These changes add the benefit of the method being
implementable in GPUs and such acceleration improves the computational performance by yet
another order of magnitude into the frame rate range of units of frames per second (FPS), making
it a viable option for real time detection.
FIGURE 2: Faster R-CNN Workflow.
2.2 SSD Object Detector
Single Shot Multibox Detector (SSD) was published by Liu et al. in 2016 as a competitive real-
time object detection method and at that time, it outclassed its contemporaries including Faster R-
CNN both in terms of mAP and framerate on VOC2007 [19].
4. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 4
SSD uses a pre-trained image classification CNN base as a feature extractor to which further
progressively smaller convolutional layers are attached. This progressive decrease in feature
map size means that objects of very different scales can be detected [19]. As opposed to
preceding object detection approaches, single shot detectors like YOLO [20] and SSD [19] do
away with a trained region proposal approach and generally apply a variant of a grid based RoI
generator, in SSD, a set of default bounding boxes is generated for each feature map. These
default boxes are of pre-selected aspect ratios in order to accommodate different object classes
[19].
FIGURE 3: SSD Workflow (NMS - Non Maximum Suppression).
From all the default boxes classification, scores for all classes are extracted and Non Maximum
Suppression is used to select only the strongest of the closely located detections.
SSD, when first published, outperformed Faster R-CNN on Pascal VOC and COCO datasets both
in mAP and framerate. SSD300 performed with mAP 74.3% compared to 73.2% of Faster R-
CNN, at 46 FPS compared to 7 FPS [19].
3. IMAGE DEGRADATIONS
In traffic images, different image degradations appear in many cases, such as JPEG compression
artifacts, motion blur, and out-of-focus blur, that have been selected as the representative ones.
JPEG compression is still one of the most widely used compression algorithms. Based on 8x8
block DCT, the compression leaves rather prominent artifacts in images compressed at higher
compression rates [21]. These artifacts are mostly ringing (caused by missing higher frequencies
as more of the DCT is truncated) and block-to-block discontinuities. Although more modern and
better compression standards have since been introduced, JPEG is still used across the industry
especially in static traffic monitoring systems where lowering data sizes is important. Methods
exist for JPEG artifact removal [22], but they require additional computational costs.
Motion blur is a blur caused by objects in the imaging moving too fast relatively to the exposure
time. Motion blur affects mobile and static traffic systems similarly, while in static installations it is
the objects to be detected that can move, in mobile systems it can be both motion of the camera
and motion of the scene at the same time causing the blur. Similarly to the JPEG artifact removal,
motion blur removal has been studied and removal methods are available [23].
Out-of-focus blur is the final type of degradation discussed. It is caused by objects in the scene
being outside of the camera’s depth-of-field. This type of degradation affects static traffic cameras
the most. This type of blur can be readily removed as well [24].
5. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 5
As stated above, all selected degradations are to an extent removable and their avoidance or
removal is advisable. Should they still occur, their removal might be computationally costly and
time consuming and that is why investigating their effect on object detection is important.
4. DATASETS & EXPERIMENTS
To assess performance of the selected architectures, an experiment was devised where
pedestrian and cyclist detectors trained on undegraded data applied on newly compiled datasets
with varying magnitude of degradation.
Detectors were evaluated in terms of their average precision for each class. Average precision
being used as defined by the Pascal VOC challenge [25, 26], and in terms of their cross-class
confusion (cases where a pedestrian is detected to be a cyclist and vice versa).
4.1 Object Detectors
Two object detection models were trained of the selected architectures. The models trained by
the authors were trained to detect both pedestrians and cyclists, while those selected for
comparison only detect pedestrians.
A Faster R-CNN based detector has been selected for evaluation. The particular variant selected
uses Resnet101 [27] as a feature extractor. The detector was trained on a combination of the left
colour image of the stereo pair in the KITTI 2012 stereo dataset [28] and authors’ own pedestrian
dataset. Data augmentation using horizontal flip and limited skew were used to further improve
the training. The training has been done through Tensorflow Object Detection API [29]. We will
further refer to this detector as FRCNN Zemcik. For comparison, two more faster R-CNN
detectors were used, FRCNN Tilgner [15] and FRCNN Kitti zoo [29], both using Resnet101 as a
feature extractor.
Representing SSD architecture an SSDLite model SSD Zemcik has been trained on the same
data. The model utilizes MobileNetV2 as the CNN backbone [30]. The same architecture trained
by Tilgner will be used for comparison as SSD Tilgner [15].
4.2 Datasets
Testing datasets were collected to assess quality of all measured detectors. The datasets were
designed to be representative of a multitude of traffic situation and individual images were not
selected to purposely help or hinder detection. Two datasets have been compiled for this
purpose:
Small Pedestrian and Cyclist Dataset - SPCD - consists of 200 images gathered using a state-of-
the-art mid-range dashcam. The image data has been gathered in urban and suburban conditions
in Brno, Czech Republic. This dataset only contains daytime images and all images contain at
least one instance of a pedestrian or a cyclist.
FIGURE 4: Selected Images from SPCD.
6. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 6
The Low Quality Traffic Dataset – LQTD - consists of over 400 images taken using a purposely
selected low resolution low image quality dashcam with further image degradations, such as item
reflections on the windscreen or more dirty windscreen achieved by placing the camera out of the
wiper arch. Similarly to the SPCD, it was gathered in daytime and all weather. Majority of the
images contain an instance of a pedestrian, a cyclist or both.
FIGURE 5: Selected Images from LQTD.
All three types of degradations were applied on all images of both datasets consecutively in
varying severity. JPEG compression was applied with quality graduated from 100% to 10% to
achieve images with increasing severity of JPEG artifact degradation.
FIGURE 6: JPEG compressed image with quality of 100%, 50%, and 10% respectively.
Motion blur was applied using a horizontal linear averaging filter from 2 to 20 pixels in width.
FIGURE 7: Motion blurred images with kernels of 0px, 10px and 20px.
7. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 7
The out-of-focus blur was approximated by application of a linear averaging filter with circular
aperture simulating a Point Spread Function of an out-of-focus lens. The circle radius was
selected from 1 to 10 pixels to correspond to the width of the motion blur kernel size.
FIGURE 8: Out-of-focus blurred images with kernel radius of 0px, 5px and 10px.
JPEG (quality) 100% 90% 80% 70% 60% 50% 40% 30% 20% 10%
SPCD 1.000 0.987 0.980 0.973 0.961 0.960 0.950 0.926 0.870 0.811
LQTD 1.000 0.975 0.962 0.952 0.941 0.933 0.923 0.906 0.878 0.811
Motion (kernel) 2px 4px 6px 8px 10px 12px 14px 16px 18px 20px
SPCD 0.973 0.957 0.934 0.904 0.875 0.851 0.830 0.812 0.797 0.784
LQTD 0.966 0.940 0.908 0.880 0.857 0.837 0.821 0.808 0.797 0.787
OoF (kernel rad) 1px 2px 3px 4px 5px 6px 7px 8px 9px 10px
SPCD 0.973 0.936 0.882 0.830 0.767 0.732 0.706 0.681 0.653 0.650
LQTD 0.957 0.902 0.834 0.787 0.737 0.708 0.686 0.664 0.648 0.635
TABLE 1: Mean Structural Similarity Index Measure for all degradations on both datasets (SPCD and
LQTD) with respect to JPEG compression quality, motion blur kernel width and out-of-focus blur kernel
radius.
The severity of the degradation was quantified using Structural Similarity Index Measure (SSIM)
[31]. Table 1 gives the mean SSIM for both SPCD and LQTD datasets with each degradation
severity.
5. RESULTS
The results of the experiments described above are presented in the following charts.
Figure 9 gives the class average precision as per the Pascal VOC metric [25, 26] for each
classifier using both datasets. Pedestrian detectors are shown in solid line, while cyclist detectors
are in dashed line. In all cases, the severity of the degradation increases from right to left.
Figure 10 gives the cross-class confusion for each of the classifiers. The confusion was quantified
using the same Pascal VOC average precision metric, only the pedestrian and cyclist ground
truth labels were switched. The solid line represents cyclists in the image that were erroneously
detected and classified as pedestrians, while dashed line represents pedestrians erroneously
labelled as cyclists.
8. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 8
FIGURE 9: Average precision of all classifiers on both datasets.
9. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 9
FIGURE 10: Cross-confusion of all classifiers on both datasets.
10. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 10
From figure 9 it is apparent that in all cases additional degradation reduces overall performance
of all object detectors. It is interesting to note that while pedestrian detection precision decreases
comparably for motion blur and out-of-focus blur in both datasets respectively, in the SPCD
dataset the cyclist detection precision is much more affected by out-of-focus blur than motion
blur.
Figure 10 shows that in the LQTD dataset there is no apparent trend for cross-class confusion to
increase with increasing severity of the image degradation. On the SPCD dataset however, it is
apparent that cross-class confusion in case of cyclists erroneously labelled as pedestrians
increases significantly (by as much as 15% increase in Pascal VOC defined average precision for
inverted labels) with increasing kernel sizes of both motion blur and the out-of-focus blur.
As mentioned above the cross-class confusion is not as critical as the detection itself, but it still
has a bearing on how a higher-level path planning system sees the detected object, it is therefore
desirable to keep the confusions to a minimum.
Knowing the extent of data degradation robustness the architectures have, it is now possible to
design the ADAS systems so that the more performance hindering degradations are eliminated in
the design (especially in the image acquisition and pre-processing stages). This also makes it
possible to save resources trying to eliminate certain degradations that after all are not critical to
the performance of the system (such as high quality JPEG compression).
6. CONCLUSIONS
In this paper, effects of image degradation were evaluated on selected object detection CNN
architectures trained for detecting pedestrians and cyclists in typical traffic imagery. Not
surprisingly, the performance of both SSD and Faster R-CNN based detectors dropped with
increasing degradation severity.
The measurements show that artifacts of JPEG compression do not degrade the performance
significantly with high quality - low compressions and only take effect with the rarely used high
compressions. The effect was also much more pronounced with the lower resolution dataset
because of the fixed block size used in JPEG compression.
The effects of motion blur were again more severe on the lower resolution LQTD dataset. On the
higher resolution SPCD dataset, a significant growth of pedestrian and cyclist confusions
occurred in both directions with the increase in the magnitude of the motion blur. The effects of
out-of-focus blur were very similar to the effect of the motion blur.
Better understanding of the effects of image degradation on CNN based object detectors will help
create more robust and efficient ADAS systems and will help push us one step closer towards
fully automated transportation systems.
In future work, we would like to further investigate the effects of image degradation on traffic
object detection, and we would also like to investigate how different approaches to the
degradation effect mitigation (data augmentation with degradations pre-training and image
degradation removal before detection) compare, and we would like to implement the selected
degradation effect mitigation mitigating solutions in a practical system.
7. REFERENCES
[1] European Commission (2019). EU Road Safety Policy Framework 2021-2030 - Next steps
towards "Vision Zero". Luxembourg: Office for Official Publications of the European
Communities.
11. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 11
[2] World Health Organization (2018). Global Status Report on Road Safety 2018 [Online].
Available:https://www.who.int/publications/i/item/9789241565684#:~:text=The%20Global%2
0status%20report%20on,people%20aged%205%2D29%20years. [Nov 15, 2020]
[3] V. Viswanathan & R. Hussein (2017, Apr.). Applications of Image Processing and Real-Time
embedded Systems in Autonomous Cars: A Short Review, International Journal of Image
Processing (IJIP), [Online] Volume (11) : Issue (2). Available:
https://www.cscjournals.org/manuscript/Journals/IJIP/Volume11/Issue2/IJIP-1114.pdf. [Jan
23, 2021]
[4] S. Kim, I. Wisanggeni, R. Ros & R. Hussein (2020, Jan.). Detecting Fatigue Driving Through
PERCLOS: A Review, International Journal of Image Processing (IJIP), [Online] Volume
(14) : Issue (1) : 2020. Available:
https://www.cscjournals.org/manuscript/Journals/IJIP/Volume14/Issue1/IJIP-1194.pdf. [Jan
23, 2021]
[5] Z. Cheng, L. Jeng & K. Li (2018, Nov.) Behavioral Classification of Drivers for Driving
Efficiency Related ADAS Using Artificial Neural Network, 2018 IEEE International
Conference on Advanced Manufacturing (ICAM), [Online] pp. 173-176. Available:
https://ieeexplore.ieee.org/document/8614836. [Nov 15, 2020]
[6] O. Fard & S. Hamidreza. Efficient multi-class object detection with a hierarchy of classes.
Aubière, 2015, pp. 36-40.
[7] Viola, Jones, & Snow (2003, Oct.). Detecting pedestrians using patterns of motion and
appearance. In Proceedings Ninth IEEE International Conference on Computer Vision
[Online]. (pp. 734-741 vol.2). Available: https://ieeexplore.ieee.org/document/1238422 [Nov
15, 2020]
[8] R. Benenson, M. Omran, J. Hosang, & B. Schiele. (2014, Nov.). Ten Years of Pedestrian
Detection, What Have We Learned?. [Online]. Available: https://arxiv.org/abs/1411.4304
[Nov 13, 2020]
[9] N. Dalal, & B. Triggs (2005). Histograms of oriented gradients for human detection. In 2005
IEEE Computer Society Conference on Computer Vision and Pattern Recognition
(CVPR'05) (pp. 886-893 vol. 1).
[10] Wojek C., Schiele B. (2008) A Performance Evaluation of Single and Multi-feature People
Detection. In: Rigoll G. (eds) Pattern Recognition. DAGM 2008. Lecture Notes in Computer
Science, vol 5096. Springer, Berlin, Heidelberg. [Online]. Available:
https://doi.org/10.1007/978-3-540-69321-5_9 [Nov 23, 2020]
[11] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, & D. Ramanan (2010). Object Detection
with Discriminatively Trained Part-Based Models IEEE Transactions on Pattern Analysis and
Machine Intelligence, 32(9), 1627-1645.
[12] W. Tian, & M. Lauer (2015, Sep.). Fast Cyclist Detection by Cascaded Detector and
Geometric Constraint. In 2015 IEEE 18th International Conference on Intelligent
Transportation Systems [Online]. (pp. 1286-1291). Available:
https://ieeexplore.ieee.org/document/7313303 [Dec 05, 2020]
[13] Roy, P., Ghosh, S., Bhattacharya, S., & Pal, U. (2018, Jul.). Effects of Degradations on
Deep Neural Network Architectures. [Online]. Available: https://arxiv.org/abs/1807.10108.
[Dec 01, 2020]
[14] Y. Pei, Y. Huang, Q. Zou, X. Zhang, & S. Wang (2019, Nov.). Effects of Image Degradation
and Degradation Removal to CNN-based Image Classification, IEEE Transactions on
12. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 12
Pattern Analysis and Machine Intelligence, [Online]. 1-1. Available:
https://ieeexplore.ieee.org/document/8889765. [Nov 18, 2020]
[15] M. Tilgner. “Pedestrians Detection in Traffic Environment by Machine Learning” Master’s
thesis, Brno University of Technology, 2019
[16] R. Girshick, J. Donahue, T. Darrell, & J. Malik (2014, Oct.). Rich Feature Hierarchies for
Accurate Object Detection and Semantic Segmentation. In 2014 IEEE Conference on
Computer Vision and Pattern Recognition [Online]. (pp. 580-587). Available:
https://arxiv.org/abs/1311.2524. [Dec 01, 2020]
[17] R. Girshick (2015). Fast R-CNN. In 2015 IEEE International Conference on Computer Vision
(ICCV) (pp. 1440-1448).
[18] S. Ren, K. He, R. Girshick, & J. Sun (2016, Jan). Faster R-CNN: Towards Real-Time Object
Detection with Region Proposal Networks IEEE Transactions on Pattern Analysis and
Machine Intelligence, [Online]. 39(6), 1137-1149. Available: https://arxiv.org/abs/1506.01497
[Dec 02, 2020]
[19] Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., & Berg, A. (2016, Dec).
SSD: Single Shot MultiBox Detector, Lecture Notes in Computer Science, [Online]. 21–37.
Available: https://arxiv.org/abs/1512.02325. [Dec 15, 2020]
[20] Redmon, J., Divvala, S., Girshick, R.B., & Farhadi, A. (2016). You Only Look Once: Unified,
Real-Time Object Detection. 2016 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), 779-788.
[21] International Telecommunication Union (1992), T.81 – DIGITAL COMPRESSION AND
CODING OF CONTINUOUS-TONE STILL IMAGES – REQUIREMENTS AND
GUIDELINES. [Online]. Available: https://www.w3.org/Graphics/JPEG/itu-t81.pdf [Nov 12,
2020]
[22] P. Svoboda, M. Hradiš, D. Bařina, & P. Zemčík (2016, May.). Compression Artifacts
Removal Using Convolutional Neural Networks, Journal of WSCG, [Online]. 24(2), 63–72.
Available: https://arxiv.org/abs/1605.00366. [Nov 21, 2020]
[23] P. Svoboda, M. Hradiš, L. Maršík, & P. Zemčík (2016, Feb.). CNN for license plate motion
deblurring. In IEEE International Conference on Image Processing (ICIP) [Online]. (pp.
3832–3836). IEEE Signal Processing Society. Available: https://arxiv.org/abs/1602.07873.
[Dec 01, 2020]
[24] M. Hradiš, J. Kotera, P. Zemčík, & F. Šroubek (2015). Convolutional Neural Networks for
Direct Text Deblurring. In Proceedings of BMVC 2015 (pp. 1–13). The British Machine
Vision Association and Society for Pattern Recognition.
[25] Everingham, M., Van Gool, L., Williams, C., Winn, J., & Zisserman, A. (2010). The Pascal
Visual Object Classes (VOC) challenge International Journal of Computer Vision, 88, 303-
338.
[26] R. Padilla, S. L. Netto, & E. A. B. da Silva (2020, Jul.). A Survey on Performance Metrics for
Object-Detection Algorithms. In 2020 International Conference on Systems, Signals and
Image Processing (IWSSIP) [Online] (pp. 237-242). Available:
https://www.researchgate.net/publication/343194514_A_Survey_on_Performance_Metrics_f
or_Object-Detection_Algorithms. [Nov 15, 2020]
[27] K. He, X. Zhang, S. Ren, & J. Sun (2016, Dec.). Deep Residual Learning for Image
Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition
(CVPR) [Online] (pp. 770-778). Available: https://arxiv.org/abs/1512.03385. [Dec 01, 2020]
13. Tomáš Zemčík, Lukáš Kratochvíla, Šimon Bilík, Ondřej Boštík, Pavel Zemčík & Karel Horák
International Journal of Image Processing (IJIP), Volume (15) : Issue (1) : 2021 13
[28] Andreas Geiger, Philip Lenz, & Raquel Urtasun (2012). Are we ready for Autonomous
Driving? The KITTI Vision Benchmark Suite. In Conference on Computer Vision and Pattern
Recognition (CVPR).
[29] Huang, J., Rathod, V., Sun, C., Zhu, M., Korattikara, A., Fathi, A., Fischer, I., Wojna, Z.,
Song, Y., Guadarrama, S., & Murphy, K. (2017, Apr.). Speed/Accuracy Trade-Offs for
Modern Convolutional Object Detectors. In CVPR 2017 [Online] (pp. 3296-3297). Available:
https://arxiv.org/abs/1611.10012. [Nov 15, 2020]
[30] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, & L. Chen (2018). MobileNetV2: Inverted
Residuals and Linear Bottlenecks. In 2018 IEEE/CVF Conference on Computer Vision and
Pattern Recognition (pp. 4510-4520).
[31] Zhou Wang, A. C. Bovik, H. R. Sheikh, & E. P. Simoncelli (2004, Apr.). Image quality
assessment: from error visibility to structural similarity, IEEE Transactions on Image
Processing, [Online] 13(4), 600-612. Available:
https://ieeexplore.ieee.org/document/1284395. [Dec 01, 2020]