SlideShare a Scribd company logo
1 of 75
1
Identification d’objets pour l’évaluation
de risque en conduite autonome
Séminaire IA & VA - 26 novembre 2021
• Introduction
– Présentation du laboratoire CIAD
– Véhicule autonome
• Attention visuelle
• Identification d’objets
OUTLINE
Laboratoire CIAD
Membres du laboratoire CIAD
5
Perception
de
l’env
Explicabilité
Raisonnement hybride Raisonnement distribué
Champs scientifiques principaux
Data/Graph
Mining
Ingénierie des
connaissances
Apprentissage
machine
Vision par ordinateur
Robotique
Systèmes multiagents
Program
mation
logique
Optimisation
meta-heuristique
ou bio-inspirée
UX
Réalité
virtuelle
TA
L
Génie
logiciel
Simulation
Visualisa
tion
données
6
Axes de recherche transversaux
01 03
02
Véracité - Valeur
Perception,qualification de
la véracité et de la valeur de
connaissances dans des
environnements intelligents
massifs
Recommendation et simulation
prescriptive pour des systèmes
distribués ou complexes
Comportements et raisonnement
distribués dans des systèmes
complexes
(e.g. cyber-physical systems)
Simulation
prescriptive
Raisonnement distribué
7
Principaux domaines d’applications
8
Plate-forme MOBILITECH : véhicules autonomes, drones, robots et UX
Type véhicule GEM Little Renault
Scenic
Renault
Zoe
Hyundai
Matériel Télémètre,
caméra,
central
inertielle, …
Télémètre,
caméra,
central
inertielle,
pile à
combustible,
…
Télémètre,
caméra, …
Télémètre, caméra, central inertielle,
…
Quantité 1 1 2 2 3 1
Thématique Acquisition
dedonnées
sur route
classique
Système
d’aideà la
conduite
(ADAS);
conduite
autonome
en milieu
contrôlé
Conduiteen
convoi
ADAS; conduiteautonomesur route
ferméeou ouverte
+ Drones, robots domestiques (Toyota) et industriels (Hunter)
Vidéo
Autonomous vehicles
Autonomous Driving system consists of three main parts
• Perception
• Planning
• Control
and each part includes different tasks that are expected to be fully understood by the
system.
AUTONOMOUS VEHICLES
Sensors and Perception Models
Relies heavily on an extensive
infrastructure of active and
passive sensors.
PERCEPTION
PERCEPTION
Environmental Sensing
Technology
LIDAR
CAMERA
S
RADAR
ULTRASONIC SENSORS
PLANNING
Future Trajectories based on the results coming from
perception part.
Choosing appropriately ego trajectory, and driving
behavior is created and planned.
CONTROL
• The control part is deeply coupled with the perception and planning parts.
• It guarantees that the vehicle follows the course set by the planning part and
controls the vehicle’s hardware (acceleration, braking, and steering using drivers
and actuators) for safe driving.
ENVIRONMENT PERCEPTION
Integrating advanced sensing and perception models
Environmental Sensing → Sensor Fusion
Technology
Environmental Perception Models → Computer Vision Technology
PERCEPTION
Environment Analysis &
Understanding
(Segmentation,Detection,
Recognition, Tracking…..)
SENSING
Camera
Inertial
Ultra
Sonic
RADAR
GPS
LIDAR
ENVIRONMENT PERCEPTION
The principal goal is:
• Designing computer systems that possess
the ability to capture, understand, and
interpret important visual information
contained with image, video and other
visual data.
• Then translate this data, using contextual
knowledge provided by human beings, into
insights used to drive decision making.
Computer Vision
Technology
ENVIRONMENT PERCEPTION
Training computer vision systems used to
involve learning process all the way to the
smallest granular units of visual data – PIXEL.
• The system records and evaluates digital
images on the basis of its raw data alone,
where minute differences in pixel density,
color saturation, and levels of lightness and
darkness determine the structure and
therefore identity of the larger object.
Computer Vision
Technology
ENVIRONMENT PERCEPTION
is one of the most remarkable things to come out of the deep learning and artificial
intelligence world. The advancements that deep learning has contributed to the computer
vision field have really set this field apart.
Computer Vision
Technology
ENVIRONMENT PERCEPTION
Combining sensor & vision technologies leads to
Environment Perception → Scene Understanding
Scene Understanding: Analysis of a scene, considering the semantic and geometric
context of its contents and the internal relations between them.
Humans can classify, locate, segment, and identify objects and features at one look.
ENVIRONMENT PERCEPTION
Humans can classify (object type and status moving/static), specify (spatial position),
identify (motion, position, direction, and velocity), and track these objects in the
driving scene.
In addition, humans focus their visual attention more on important or purposeful
elements and ignore unnecessary ones in their field of view.
Conferring these phenomenal abilities into machine-learning systems has been a long-
standing goal in the field of computer vision
ENVIRONMENT PERCEPTION
Numerous approaches and methods (classic algorithms and deep learning) have been
proposed to improve scene understanding and extract semantic information about the driving
environment from images and videos.
ENVIRONMENT PERCEPTION – Visual Attention
Ability of sensing and perceiving the driving environment is a key technology for ADAS and
autonomous driving.
Humans' visual attention
• Predicting or locating potential risk
• Understanding the driving environment
• Quickly locate objects of interest
• “Visual Attention” - selection of
relevant and filtering out of
irrelevant information from
cluttered visual scenes.
• Attention - operate on regions of
space, particular features of an
object, or on entire objects.
• Attention can also be directed either
overtly or covertly.
ENVIRONMENT PERCEPTION – Visual Attention
How should a machine learning system or an autonomous vehicle
acquire this ability to detect such attentions for safe driving?
• To tackle this
– Incorporate saliency mechanism (Salient object detection model) as visual
attention model
– Saliency refers to unique features (pixels, resolution etc.) of the image in the context
of visual processing. These unique features depict the visually alluring locations in an
image. Saliency map is a topographical representation of them.
– Salient object detection models mimic the behaviour of human beings and capture
the most salient region/object from the images or scenes.
ENVIRONMENT PERCEPTION – Visual Attention
The concept of saliency map was first proposed by Christof
Koch in 1985.
Feature Integration Theory - 1980 " defining human visual
search strategies
“The salient areas in the visual scene are identified by
combining or relating visual features information such as
color, orientation, spatial frequency, brightness, direction of
movement that direct human attention”
The visual attention methods using saliency are divided into
two categories:
– Bottom-up: Biologically inspired methods Image color
and intensity are common examples
– Top-down: True computational methods
Prior knowledge, memories, goals are common factors.
Bottom-up
ENVIRONMENT PERCEPTION – Visual Attention
Top-down
• Saliency Algorithms
Review and test some well known Saliency algorithms both classic and deep learning
Fig: Diffèrent Saliency Algorithms test on driving scene images
ENVIRONMENT PERCEPTION – Visual Attention
Existing works
Human eye tracking into the process Adding object detection
In -lab Simulation
Real Driving
Berkeley DeepDrive Attention Dr(eye)VE SAGE
ENVIRONMENT PERCEPTION – Visual Attention
The research shows that these models have contributed a lot to the deployment of
attention and have significant progress.
Some limitations and drawbacks
– Complexity of capturing the true driver attention
– The fixations: Subjected to various characteristics of driver
• Driving experience & habits, preference & intentions, capabilities, culture, age,
gender, etc
– Eye tracker record single location at each moment that driver is gazing at, while he
may be looking at multiple important objects in the scene.
– The computational cost for developing a dataset with
saliency maps has been relatively expensive.
ENVIRONMENT PERCEPTION – Visual Attention
Our Aim: Develop a visual attention framework capable of predicting the important
objects (road context) simultaneously in a driving scene.
ENVIRONMENT PERCEPTION – Visual Attention
We came up with a new idea by shifting the problem from PREDICTION
What/where the driver is looking, or most drivers would look at to
SELECTION What the driver should/must look at during driving
ENVIRONMENT PERCEPTION – Visual Attention
Saliency Heat-Map as Visual Attention for Autonomous Driving
using Generative Adversarial Network (GAN)
Our Approach
Generative Adversarial Network - GAN’ (by Ian Goodfellow in 2014)
– GAN is a type of neural network architecture for generative modeling. It involves using a
model to generate new examples that plausibly come from an existing distribution of
samples, such as generating new photographs that are similar but specifically different
from a dataset of existing photographs
Applications: to name few
Image-to-Image Translation
Face Frontal
View Generation
Photos to Emojis Photo Inpainting
ENVIRONMENT PERCEPTION – Visual Attention
Framework
• We borrow the pix2pix GAN architecture which is suitable for image-to-image translation
task and can be conditioned on the input image for generating corresponding output image.
ENVIRONMENT PERCEPTION – Visual Attention
Data Collection
– The data nowadays for developing saliency models are collected from human eye fixation or
gazing. These data as saliency map (gray-scale or heat map image) are obtained using Gaussian
probability function, that exhibits the probability of each image pixel that captures human
attention.
Examples of fixation selection – Prediction points
(MITdataset)
ENVIRONMENT PERCEPTION – Visual Attention
Data Collection
– We propose a different approach for data collection by taking advantage of
semantic label information from driving scenes datasets.
ENVIRONMENT PERCEPTION – Visual Attention
Our Framework Experiments:
• Visual results on VADD validation set
ENVIRONMENT PERCEPTION – Visual Attention
Our Framework Experiments:
• Visual results on different environments and weather conditions
ENVIRONMENT PERCEPTION – Visual Attention
Our Framework Experiments:
• Test on EU long-term and Synthia Datasets
ENVIRONMENT PERCEPTION – Visual Attention
Our Framework Experiments:
• Random Unseen Images
ENVIRONMENT PERCEPTION – Visual Attention
Our Framework Experiments:
• Against Eye Fixation Attention Models.
Visual Comparison with Dr(eye)VE Project results
ENVIRONMENT PERCEPTION – Visual Attention
Our Framework Experiments:
• Against Eye Fixation Attention Models.
Visual Comparison with BDD-A results
ENVIRONMENT PERCEPTION – Visual Attention
Ours BDD-A
Ours Dr(eye)Ve
BDD
Our Framework Experiments:
• Against Eye Fixation Attention Models
ENVIRONMENT PERCEPTION – Visual Attention
BDD
Our Framework Experiments:
• On VADD
ENVIRONMENT PERCEPTION – Visual Attention
Limitations and Future Work:
• Not all these objects are demanding all the time
• Detects false regions as salient
• Triggers false detection due to direct sunlight or reflections of light
 Incorporate depth, location, motion information, this could help for giving priorities within the
detected objects
ENVIRONMENT PERCEPTION – Visual Attention
ENVIRONMENT PERCEPTION
• Based on our research into visual attention, we have concluded that in addition to visual
attention, we require the incorporation of depth, location, and motion information, which
could aid in the assignment of priorities among the detected objects.
• It is critical to have a thorough understanding of each of the surrounding elements
(including the motion and geometry information).
• Road-users are critical for perception, planning and decision-making for both self-driving
cars and driver assistance systems.
• Some road-users, however, are more important for decision making than others because
of their respective intentions, ego-vehicle’s intention and their effects on each other.
 Road Geometry
 Road Boundaries
 Road Users
 Road Semantics:
ENVIRONMENT PERCEPTION
Aim: To propose a framework that can extract motion and geometry related information (i.e.,
object class, status, position, movement, speed and distance information) to identify object
(road-user) characteristics in the urban driving scenario.
Using these and other semantic cues, we may be able to determine the most important (prior)
objects in a given scene while driving.
ENVIRONMENT PERCEPTION – Object Identification
Motion and Geometry-related Information Fusion through
a Framework for Object Identification from a moving
camera in Urban Driving Scenarios
ENVIRONMENT PERCEPTION – Object Identification
We have reviewed the contributions of works most related to ours,
– i.e., scene understanding for driving by combining motion and geometry-related information.
SMS-Net
ENVIRONMENT PERCEPTION – Object Identification
FRAMEWORK FOR OBJECT IDENTIFICATION (FOI) - Workflow
ENVIRONMENT PERCEPTION – Object Identification
• Framework's components includes
– Disparity Estimation through the Semi-Global Matching
– Motion Estimation by Image Registration and Optical Flow
– Moving Object Detection (MOD)
– Information Extraction and Fusion
DISPARITY ESTIMATION
– We adopt a well known Semi-Global Matching (SGM) algorithm.
ENVIRONMENT PERCEPTION – Object Identification
ENVIRONMENT PERCEPTION – Object Identification
MOTION ESTIMATION
– One of the most widely used methods is optical flow (OF) estimation.
– OF provides satisfactory results when the camera is fixed or carefully displaced.
– However, the optical flow from image sequences acquired by a moving camera encodes two
pieces of information.
• The motion of the surrounding objects
• The ego vehicle's motion
results in significant motion vectors associated with the static objects, leading to an
incorrect perception of static objects as moving objects.
MOTION ESTIMATION
Approach to Motion Compensation:
– By recent trends in aerial and medical imaging, we suggest a procedure to be used called image
registration together with the optical flow method to overcome ego-motion and obtain true
estimation of the motion information.
IMAGE REGISTRATION
With Image Registration
Without Image Registration
ENVIRONMENT PERCEPTION – Object Identification
MOTION ESTIMATION
Computing Optical Flow
ENVIRONMENT PERCEPTION – Object Identification
MOVING OBJECT DETECTION (MOD)
We followed a straightforward approach for moving object detection, that is
• To detect the objects of interest first
• Then, identify the moving ones from detected
objects from two consecutive frames
Current
frame
Previous
frame
Moving
Objects
OBJECTDETECTION
• Classification
• Localization
• Segmentation of all the
objects in the scene
IDENTIFY MOVING OBJECTS
• Recognitionof Moving
objectsfromdetectedobjects
inthe scene
ENVIRONMENT PERCEPTION – Object Identification
MOVING OBJECT DETECTION (MOD)
E D
MOD
Moving Objects Mask
Moving Object
Detection
Encoder-Decoder
Network
Superimpose image frame and
corresponding moving object mask, and
integrate Bbox and Class information
Seg Mask
Bbox and Class
Segmentation
Network
The proposed architecture incorporates the object segmentation, and binary pixel classification based on
temporal information
The object segmentation part is a pre-trained CenterMask Lite segmentation network inference, which gives the bounding
boxes, category probabilities, and segmentation mask for each object of interest
The temporal processing part is an encoder-decoder network (EDNet) that identifies the moving objects using segmented
masks of consecutive frames
ENVIRONMENT PERCEPTION – Object Identification
MOVING OBJECT DETECTION (MOD)
The proposed model labels the moving objects (as
white) and the static/background (as black) from
sequence pair images.
ENVIRONMENT PERCEPTION – Object Identification
PROPOSED MOD DATASET
We developed a large dataset for moving object detection (from KITTI and EU-life long
datasets) covering all the dynamic objects like all types of vehicles, pedestrians, cyclists,
motorcyclists, bus, train, and truck.
ENVIRONMENT PERCEPTION – Object Identification
MOVING OBJECT DETECTION (MOD) - RESULTS
ENVIRONMENT PERCEPTION – Object Identification
MOVING OBJECT DETECTION (MOD) - RESULTS
EU-LT SEQ_02
ENVIRONMENT PERCEPTION – Object Identification
FULLY COMPENSATED OPTICAL FLOW (FCOF)
ENVIRONMENT PERCEPTION – Object Identification
FUSION OF MOD, FCOF AND DISPARITY
The results of each stage of the proposed framework, such as disparity, moving object
detection, and motion estimation, are fused to extract information such as object ID, static
or moving, distance, direction, position, and velocity.
– Labelling and Scaling
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
KITTI SEQ_05
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
KITTI SEQ_84
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
KITTI SEQ_71
ENVIRONMENT PERCEPTION – Object Identification
FOI - Results
KITTI SEQ_59
ENVIRONMENT PERCEPTION – Object Identification
CONCLUSION
• Proposal of a new framework for object identification (FOI) from a moving camera in complex urban
driving environment.
• The framework relies only upon the images captured from the stereo camera.
• It extracts the information related to the object, including class, status (moving/static), direction, velocity,
position, and distance from the ego vehicle.
• Other contributions related to Moving Object Detection (MOD) as it is considered as the critical task.
– A new dataset for Moving Object Détection is built from existing driving datasets KITTI and EU
long-term, covering dynamic objects like all types of vehicles, pedestrians, bicyclists, and
motorcyclists.
– Propose to use image registration as a tool for ego-motion compensation for urban driving scenarios.
– A new model for moving object detection is developed by integrating an encoder-decoder network
with a segmentation model.
ENVIRONMENT PERCEPTION – Object Identification
LIMITATIONS: Through the experiments, it is found that
Objects overlap or very close to each other Object reflection Object motion speed is same as ego-vehicle speed
ENVIRONMENT PERCEPTION – Object Identification
FUTUREWORK
– More attention should be paid to improving the overall speed of the proposed framework.
– Build a new dataset – Fully Compensated Optical Flow color maps.
– Future studies will be devoted to develop a framework that can prioritize objects in a driving
scene according to the situation.
– We will use the perception data (object identification) to plan secure and smooth trajectories for
the objects of interest using their dynamics limits, navigation comfort and safety, and the traffic
rules.
– Lane detection, traffic sign detection, and live traffic light detection could be integrated into the
framework, which would help the system in various tasks such as lane-change, obstacle
avoidance, and combined in critical driving situations.
ENVIRONMENT PERCEPTION – Object Identification
Road-user importance estimation during a left turn maneuver
• In real-world driving, there can be a variety of items in the immediate neighborhood of the
ego-vehicle at any given time.
• Some items have a direct impact on the behavior of the ego-vehicle (e.g., brake, steer),
while others have the potential to be a danger, and still others do not pose a danger
currently or soon.
• The ability to determine how important or relevant a given object is to an ego-decision
vehicle's is critical for both driver assistance systems and self-driving vehicles.
Establishing trust with human drivers or passengers
Demonstrating transparency to law enforcement
Promoting a human-centric thought process etc
Road-user importance estimation during a left turn maneuver
THANKS
Contact:
Yassine Ruichek
Laboratoire CIAD
Université de Bourgogne Franche Comté (UBFC)
Université de Technologie Belfort Montbéliard (UTBM)
yassine.ruichek@utbm.fr– 06 73 22 10 19

More Related Content

Similar to Séminaire IA & VA- Yassine Ruichek, UTBM

Image Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learningImage Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learningPRATHAMESH REGE
 
Vision-Based Motorcycle Crash Detection and Reporting Using Deep Learning
Vision-Based Motorcycle Crash Detection and Reporting Using Deep LearningVision-Based Motorcycle Crash Detection and Reporting Using Deep Learning
Vision-Based Motorcycle Crash Detection and Reporting Using Deep LearningIRJET Journal
 
IRJET- Application of MCNN in Object Detection
IRJET-  	  Application of MCNN in Object DetectionIRJET-  	  Application of MCNN in Object Detection
IRJET- Application of MCNN in Object DetectionIRJET Journal
 
Fcv core sawhney
Fcv core sawhneyFcv core sawhney
Fcv core sawhneyzukun
 
Introduction-to-Computer-Vision PPPP.pptx
Introduction-to-Computer-Vision PPPP.pptxIntroduction-to-Computer-Vision PPPP.pptx
Introduction-to-Computer-Vision PPPP.pptxShivaniPaswan6
 
Artificial Intelligence_Strategy.pptx
Artificial Intelligence_Strategy.pptxArtificial Intelligence_Strategy.pptx
Artificial Intelligence_Strategy.pptxSureshMaddi1
 
A Literature Survey on Image Linguistic Visual Question Answering
A Literature Survey on Image Linguistic Visual Question AnsweringA Literature Survey on Image Linguistic Visual Question Answering
A Literature Survey on Image Linguistic Visual Question AnsweringIRJET Journal
 
Computer vision and robotics
Computer vision and roboticsComputer vision and robotics
Computer vision and roboticsBiniam Asnake
 
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...IRJET Journal
 
Sanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance SystemSanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance SystemIRJET Journal
 
A Traffic Sign Classifier Model using Sage Maker
A Traffic Sign Classifier Model using Sage MakerA Traffic Sign Classifier Model using Sage Maker
A Traffic Sign Classifier Model using Sage Makerijtsrd
 
Real-Time Pertinent Maneuver Recognition for Surveillance
Real-Time Pertinent Maneuver Recognition for SurveillanceReal-Time Pertinent Maneuver Recognition for Surveillance
Real-Time Pertinent Maneuver Recognition for SurveillanceIRJET Journal
 
Person Acquisition and Identification Tool
Person Acquisition and Identification ToolPerson Acquisition and Identification Tool
Person Acquisition and Identification ToolIRJET Journal
 
Visual Saliency Model Using Sift and Comparison of Learning Approaches
Visual Saliency Model Using Sift and Comparison of Learning ApproachesVisual Saliency Model Using Sift and Comparison of Learning Approaches
Visual Saliency Model Using Sift and Comparison of Learning Approachescsandit
 
Saksham presentation
Saksham presentationSaksham presentation
Saksham presentationSakshamTurki
 
Human Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueHuman Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueIRJET Journal
 
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...Editor IJMTER
 

Similar to Séminaire IA & VA- Yassine Ruichek, UTBM (20)

Image Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learningImage Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learning
 
F0932733
F0932733F0932733
F0932733
 
Vision-Based Motorcycle Crash Detection and Reporting Using Deep Learning
Vision-Based Motorcycle Crash Detection and Reporting Using Deep LearningVision-Based Motorcycle Crash Detection and Reporting Using Deep Learning
Vision-Based Motorcycle Crash Detection and Reporting Using Deep Learning
 
Computer vesion
Computer vesionComputer vesion
Computer vesion
 
IRJET- Application of MCNN in Object Detection
IRJET-  	  Application of MCNN in Object DetectionIRJET-  	  Application of MCNN in Object Detection
IRJET- Application of MCNN in Object Detection
 
Fcv core sawhney
Fcv core sawhneyFcv core sawhney
Fcv core sawhney
 
Introduction-to-Computer-Vision PPPP.pptx
Introduction-to-Computer-Vision PPPP.pptxIntroduction-to-Computer-Vision PPPP.pptx
Introduction-to-Computer-Vision PPPP.pptx
 
Artificial Intelligence_Strategy.pptx
Artificial Intelligence_Strategy.pptxArtificial Intelligence_Strategy.pptx
Artificial Intelligence_Strategy.pptx
 
Image recognition
Image recognitionImage recognition
Image recognition
 
A Literature Survey on Image Linguistic Visual Question Answering
A Literature Survey on Image Linguistic Visual Question AnsweringA Literature Survey on Image Linguistic Visual Question Answering
A Literature Survey on Image Linguistic Visual Question Answering
 
Computer vision and robotics
Computer vision and roboticsComputer vision and robotics
Computer vision and robotics
 
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
 
Sanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance SystemSanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance System
 
A Traffic Sign Classifier Model using Sage Maker
A Traffic Sign Classifier Model using Sage MakerA Traffic Sign Classifier Model using Sage Maker
A Traffic Sign Classifier Model using Sage Maker
 
Real-Time Pertinent Maneuver Recognition for Surveillance
Real-Time Pertinent Maneuver Recognition for SurveillanceReal-Time Pertinent Maneuver Recognition for Surveillance
Real-Time Pertinent Maneuver Recognition for Surveillance
 
Person Acquisition and Identification Tool
Person Acquisition and Identification ToolPerson Acquisition and Identification Tool
Person Acquisition and Identification Tool
 
Visual Saliency Model Using Sift and Comparison of Learning Approaches
Visual Saliency Model Using Sift and Comparison of Learning ApproachesVisual Saliency Model Using Sift and Comparison of Learning Approaches
Visual Saliency Model Using Sift and Comparison of Learning Approaches
 
Saksham presentation
Saksham presentationSaksham presentation
Saksham presentation
 
Human Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueHuman Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision Technique
 
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
 

More from Mahdi Zarg Ayouna

Séminaire IA & VA- Bertrand Leroy, VEDECOM
Séminaire IA & VA- Bertrand Leroy, VEDECOMSéminaire IA & VA- Bertrand Leroy, VEDECOM
Séminaire IA & VA- Bertrand Leroy, VEDECOMMahdi Zarg Ayouna
 
Séminaire IA & VA- Pierre Duthon, CEREMA
Séminaire IA & VA- Pierre Duthon, CEREMASéminaire IA & VA- Pierre Duthon, CEREMA
Séminaire IA & VA- Pierre Duthon, CEREMAMahdi Zarg Ayouna
 
Séminaire IA & VA- Arnaud de La Fortelle, Mines ParisTech
Séminaire IA & VA- Arnaud de La Fortelle, Mines ParisTechSéminaire IA & VA- Arnaud de La Fortelle, Mines ParisTech
Séminaire IA & VA- Arnaud de La Fortelle, Mines ParisTechMahdi Zarg Ayouna
 
Séminaire IA & VA- Dominique Gruyer, Univ Gustave Eiffel
Séminaire IA & VA- Dominique Gruyer, Univ Gustave EiffelSéminaire IA & VA- Dominique Gruyer, Univ Gustave Eiffel
Séminaire IA & VA- Dominique Gruyer, Univ Gustave EiffelMahdi Zarg Ayouna
 
Séminaire IA & VA- Aurélien Cord, Stanley Robotics
Séminaire IA & VA- Aurélien Cord, Stanley RoboticsSéminaire IA & VA- Aurélien Cord, Stanley Robotics
Séminaire IA & VA- Aurélien Cord, Stanley RoboticsMahdi Zarg Ayouna
 
Etienne Come : Apprentissage non supervisé, application à des données smart-c...
Etienne Come : Apprentissage non supervisé, application à des données smart-c...Etienne Come : Apprentissage non supervisé, application à des données smart-c...
Etienne Come : Apprentissage non supervisé, application à des données smart-c...Mahdi Zarg Ayouna
 
Simulation and optimization of dynamic ridesharing services
Simulation and optimization of dynamic ridesharing servicesSimulation and optimization of dynamic ridesharing services
Simulation and optimization of dynamic ridesharing servicesMahdi Zarg Ayouna
 
Compréhension de la mobilité : données ousimulation ?
Compréhension de la mobilité : données ousimulation ?Compréhension de la mobilité : données ousimulation ?
Compréhension de la mobilité : données ousimulation ?Mahdi Zarg Ayouna
 
Seminaire ia mobilite_furno_smoreda_small
Seminaire ia mobilite_furno_smoreda_smallSeminaire ia mobilite_furno_smoreda_small
Seminaire ia mobilite_furno_smoreda_smallMahdi Zarg Ayouna
 
20210318 seminaire ia_mobilite_ratp_finale
20210318 seminaire ia_mobilite_ratp_finale20210318 seminaire ia_mobilite_ratp_finale
20210318 seminaire ia_mobilite_ratp_finaleMahdi Zarg Ayouna
 

More from Mahdi Zarg Ayouna (10)

Séminaire IA & VA- Bertrand Leroy, VEDECOM
Séminaire IA & VA- Bertrand Leroy, VEDECOMSéminaire IA & VA- Bertrand Leroy, VEDECOM
Séminaire IA & VA- Bertrand Leroy, VEDECOM
 
Séminaire IA & VA- Pierre Duthon, CEREMA
Séminaire IA & VA- Pierre Duthon, CEREMASéminaire IA & VA- Pierre Duthon, CEREMA
Séminaire IA & VA- Pierre Duthon, CEREMA
 
Séminaire IA & VA- Arnaud de La Fortelle, Mines ParisTech
Séminaire IA & VA- Arnaud de La Fortelle, Mines ParisTechSéminaire IA & VA- Arnaud de La Fortelle, Mines ParisTech
Séminaire IA & VA- Arnaud de La Fortelle, Mines ParisTech
 
Séminaire IA & VA- Dominique Gruyer, Univ Gustave Eiffel
Séminaire IA & VA- Dominique Gruyer, Univ Gustave EiffelSéminaire IA & VA- Dominique Gruyer, Univ Gustave Eiffel
Séminaire IA & VA- Dominique Gruyer, Univ Gustave Eiffel
 
Séminaire IA & VA- Aurélien Cord, Stanley Robotics
Séminaire IA & VA- Aurélien Cord, Stanley RoboticsSéminaire IA & VA- Aurélien Cord, Stanley Robotics
Séminaire IA & VA- Aurélien Cord, Stanley Robotics
 
Etienne Come : Apprentissage non supervisé, application à des données smart-c...
Etienne Come : Apprentissage non supervisé, application à des données smart-c...Etienne Come : Apprentissage non supervisé, application à des données smart-c...
Etienne Come : Apprentissage non supervisé, application à des données smart-c...
 
Simulation and optimization of dynamic ridesharing services
Simulation and optimization of dynamic ridesharing servicesSimulation and optimization of dynamic ridesharing services
Simulation and optimization of dynamic ridesharing services
 
Compréhension de la mobilité : données ousimulation ?
Compréhension de la mobilité : données ousimulation ?Compréhension de la mobilité : données ousimulation ?
Compréhension de la mobilité : données ousimulation ?
 
Seminaire ia mobilite_furno_smoreda_small
Seminaire ia mobilite_furno_smoreda_smallSeminaire ia mobilite_furno_smoreda_small
Seminaire ia mobilite_furno_smoreda_small
 
20210318 seminaire ia_mobilite_ratp_finale
20210318 seminaire ia_mobilite_ratp_finale20210318 seminaire ia_mobilite_ratp_finale
20210318 seminaire ia_mobilite_ratp_finale
 

Recently uploaded

Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksSoftradix Technologies
 
Unlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsUnlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsPrecisely
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersThousandEyes
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024BookNet Canada
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Alan Dix
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDGMarianaLemus7
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentationphoebematthew05
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Wonjun Hwang
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsHyundai Motor Group
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):comworks
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfngoud9212
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Neo4j
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 

Recently uploaded (20)

Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other Frameworks
 
Unlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsUnlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power Systems
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping Elbows
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDG
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentation
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdf
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort ServiceHot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
 
Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 

Séminaire IA & VA- Yassine Ruichek, UTBM

  • 1. 1 Identification d’objets pour l’évaluation de risque en conduite autonome Séminaire IA & VA - 26 novembre 2021
  • 2. • Introduction – Présentation du laboratoire CIAD – Véhicule autonome • Attention visuelle • Identification d’objets OUTLINE
  • 5. 5 Perception de l’env Explicabilité Raisonnement hybride Raisonnement distribué Champs scientifiques principaux Data/Graph Mining Ingénierie des connaissances Apprentissage machine Vision par ordinateur Robotique Systèmes multiagents Program mation logique Optimisation meta-heuristique ou bio-inspirée UX Réalité virtuelle TA L Génie logiciel Simulation Visualisa tion données
  • 6. 6 Axes de recherche transversaux 01 03 02 Véracité - Valeur Perception,qualification de la véracité et de la valeur de connaissances dans des environnements intelligents massifs Recommendation et simulation prescriptive pour des systèmes distribués ou complexes Comportements et raisonnement distribués dans des systèmes complexes (e.g. cyber-physical systems) Simulation prescriptive Raisonnement distribué
  • 8. 8 Plate-forme MOBILITECH : véhicules autonomes, drones, robots et UX Type véhicule GEM Little Renault Scenic Renault Zoe Hyundai Matériel Télémètre, caméra, central inertielle, … Télémètre, caméra, central inertielle, pile à combustible, … Télémètre, caméra, … Télémètre, caméra, central inertielle, … Quantité 1 1 2 2 3 1 Thématique Acquisition dedonnées sur route classique Système d’aideà la conduite (ADAS); conduite autonome en milieu contrôlé Conduiteen convoi ADAS; conduiteautonomesur route ferméeou ouverte + Drones, robots domestiques (Toyota) et industriels (Hunter) Vidéo
  • 10. Autonomous Driving system consists of three main parts • Perception • Planning • Control and each part includes different tasks that are expected to be fully understood by the system. AUTONOMOUS VEHICLES
  • 11. Sensors and Perception Models Relies heavily on an extensive infrastructure of active and passive sensors. PERCEPTION
  • 13. PLANNING Future Trajectories based on the results coming from perception part. Choosing appropriately ego trajectory, and driving behavior is created and planned.
  • 14. CONTROL • The control part is deeply coupled with the perception and planning parts. • It guarantees that the vehicle follows the course set by the planning part and controls the vehicle’s hardware (acceleration, braking, and steering using drivers and actuators) for safe driving.
  • 15. ENVIRONMENT PERCEPTION Integrating advanced sensing and perception models Environmental Sensing → Sensor Fusion Technology Environmental Perception Models → Computer Vision Technology PERCEPTION Environment Analysis & Understanding (Segmentation,Detection, Recognition, Tracking…..) SENSING Camera Inertial Ultra Sonic RADAR GPS LIDAR
  • 16. ENVIRONMENT PERCEPTION The principal goal is: • Designing computer systems that possess the ability to capture, understand, and interpret important visual information contained with image, video and other visual data. • Then translate this data, using contextual knowledge provided by human beings, into insights used to drive decision making. Computer Vision Technology
  • 17. ENVIRONMENT PERCEPTION Training computer vision systems used to involve learning process all the way to the smallest granular units of visual data – PIXEL. • The system records and evaluates digital images on the basis of its raw data alone, where minute differences in pixel density, color saturation, and levels of lightness and darkness determine the structure and therefore identity of the larger object. Computer Vision Technology
  • 18. ENVIRONMENT PERCEPTION is one of the most remarkable things to come out of the deep learning and artificial intelligence world. The advancements that deep learning has contributed to the computer vision field have really set this field apart. Computer Vision Technology
  • 19. ENVIRONMENT PERCEPTION Combining sensor & vision technologies leads to Environment Perception → Scene Understanding Scene Understanding: Analysis of a scene, considering the semantic and geometric context of its contents and the internal relations between them. Humans can classify, locate, segment, and identify objects and features at one look.
  • 20. ENVIRONMENT PERCEPTION Humans can classify (object type and status moving/static), specify (spatial position), identify (motion, position, direction, and velocity), and track these objects in the driving scene. In addition, humans focus their visual attention more on important or purposeful elements and ignore unnecessary ones in their field of view. Conferring these phenomenal abilities into machine-learning systems has been a long- standing goal in the field of computer vision
  • 21. ENVIRONMENT PERCEPTION Numerous approaches and methods (classic algorithms and deep learning) have been proposed to improve scene understanding and extract semantic information about the driving environment from images and videos.
  • 22. ENVIRONMENT PERCEPTION – Visual Attention Ability of sensing and perceiving the driving environment is a key technology for ADAS and autonomous driving. Humans' visual attention • Predicting or locating potential risk • Understanding the driving environment • Quickly locate objects of interest
  • 23. • “Visual Attention” - selection of relevant and filtering out of irrelevant information from cluttered visual scenes. • Attention - operate on regions of space, particular features of an object, or on entire objects. • Attention can also be directed either overtly or covertly. ENVIRONMENT PERCEPTION – Visual Attention
  • 24. How should a machine learning system or an autonomous vehicle acquire this ability to detect such attentions for safe driving? • To tackle this – Incorporate saliency mechanism (Salient object detection model) as visual attention model – Saliency refers to unique features (pixels, resolution etc.) of the image in the context of visual processing. These unique features depict the visually alluring locations in an image. Saliency map is a topographical representation of them. – Salient object detection models mimic the behaviour of human beings and capture the most salient region/object from the images or scenes. ENVIRONMENT PERCEPTION – Visual Attention
  • 25. The concept of saliency map was first proposed by Christof Koch in 1985. Feature Integration Theory - 1980 " defining human visual search strategies “The salient areas in the visual scene are identified by combining or relating visual features information such as color, orientation, spatial frequency, brightness, direction of movement that direct human attention” The visual attention methods using saliency are divided into two categories: – Bottom-up: Biologically inspired methods Image color and intensity are common examples – Top-down: True computational methods Prior knowledge, memories, goals are common factors. Bottom-up ENVIRONMENT PERCEPTION – Visual Attention Top-down
  • 26. • Saliency Algorithms Review and test some well known Saliency algorithms both classic and deep learning Fig: Diffèrent Saliency Algorithms test on driving scene images ENVIRONMENT PERCEPTION – Visual Attention
  • 27. Existing works Human eye tracking into the process Adding object detection In -lab Simulation Real Driving Berkeley DeepDrive Attention Dr(eye)VE SAGE ENVIRONMENT PERCEPTION – Visual Attention
  • 28. The research shows that these models have contributed a lot to the deployment of attention and have significant progress. Some limitations and drawbacks – Complexity of capturing the true driver attention – The fixations: Subjected to various characteristics of driver • Driving experience & habits, preference & intentions, capabilities, culture, age, gender, etc – Eye tracker record single location at each moment that driver is gazing at, while he may be looking at multiple important objects in the scene. – The computational cost for developing a dataset with saliency maps has been relatively expensive. ENVIRONMENT PERCEPTION – Visual Attention
  • 29. Our Aim: Develop a visual attention framework capable of predicting the important objects (road context) simultaneously in a driving scene. ENVIRONMENT PERCEPTION – Visual Attention We came up with a new idea by shifting the problem from PREDICTION What/where the driver is looking, or most drivers would look at to SELECTION What the driver should/must look at during driving
  • 30. ENVIRONMENT PERCEPTION – Visual Attention Saliency Heat-Map as Visual Attention for Autonomous Driving using Generative Adversarial Network (GAN)
  • 31. Our Approach Generative Adversarial Network - GAN’ (by Ian Goodfellow in 2014) – GAN is a type of neural network architecture for generative modeling. It involves using a model to generate new examples that plausibly come from an existing distribution of samples, such as generating new photographs that are similar but specifically different from a dataset of existing photographs Applications: to name few Image-to-Image Translation Face Frontal View Generation Photos to Emojis Photo Inpainting ENVIRONMENT PERCEPTION – Visual Attention
  • 32. Framework • We borrow the pix2pix GAN architecture which is suitable for image-to-image translation task and can be conditioned on the input image for generating corresponding output image. ENVIRONMENT PERCEPTION – Visual Attention
  • 33. Data Collection – The data nowadays for developing saliency models are collected from human eye fixation or gazing. These data as saliency map (gray-scale or heat map image) are obtained using Gaussian probability function, that exhibits the probability of each image pixel that captures human attention. Examples of fixation selection – Prediction points (MITdataset) ENVIRONMENT PERCEPTION – Visual Attention
  • 34. Data Collection – We propose a different approach for data collection by taking advantage of semantic label information from driving scenes datasets. ENVIRONMENT PERCEPTION – Visual Attention
  • 35. Our Framework Experiments: • Visual results on VADD validation set ENVIRONMENT PERCEPTION – Visual Attention
  • 36. Our Framework Experiments: • Visual results on different environments and weather conditions ENVIRONMENT PERCEPTION – Visual Attention
  • 37. Our Framework Experiments: • Test on EU long-term and Synthia Datasets ENVIRONMENT PERCEPTION – Visual Attention
  • 38. Our Framework Experiments: • Random Unseen Images ENVIRONMENT PERCEPTION – Visual Attention
  • 39. Our Framework Experiments: • Against Eye Fixation Attention Models. Visual Comparison with Dr(eye)VE Project results ENVIRONMENT PERCEPTION – Visual Attention
  • 40. Our Framework Experiments: • Against Eye Fixation Attention Models. Visual Comparison with BDD-A results ENVIRONMENT PERCEPTION – Visual Attention
  • 41. Ours BDD-A Ours Dr(eye)Ve BDD Our Framework Experiments: • Against Eye Fixation Attention Models ENVIRONMENT PERCEPTION – Visual Attention
  • 42. BDD Our Framework Experiments: • On VADD ENVIRONMENT PERCEPTION – Visual Attention
  • 43. Limitations and Future Work: • Not all these objects are demanding all the time • Detects false regions as salient • Triggers false detection due to direct sunlight or reflections of light  Incorporate depth, location, motion information, this could help for giving priorities within the detected objects ENVIRONMENT PERCEPTION – Visual Attention
  • 44. ENVIRONMENT PERCEPTION • Based on our research into visual attention, we have concluded that in addition to visual attention, we require the incorporation of depth, location, and motion information, which could aid in the assignment of priorities among the detected objects. • It is critical to have a thorough understanding of each of the surrounding elements (including the motion and geometry information). • Road-users are critical for perception, planning and decision-making for both self-driving cars and driver assistance systems. • Some road-users, however, are more important for decision making than others because of their respective intentions, ego-vehicle’s intention and their effects on each other.
  • 45.  Road Geometry  Road Boundaries  Road Users  Road Semantics: ENVIRONMENT PERCEPTION
  • 46. Aim: To propose a framework that can extract motion and geometry related information (i.e., object class, status, position, movement, speed and distance information) to identify object (road-user) characteristics in the urban driving scenario. Using these and other semantic cues, we may be able to determine the most important (prior) objects in a given scene while driving. ENVIRONMENT PERCEPTION – Object Identification
  • 47. Motion and Geometry-related Information Fusion through a Framework for Object Identification from a moving camera in Urban Driving Scenarios ENVIRONMENT PERCEPTION – Object Identification
  • 48. We have reviewed the contributions of works most related to ours, – i.e., scene understanding for driving by combining motion and geometry-related information. SMS-Net ENVIRONMENT PERCEPTION – Object Identification
  • 49. FRAMEWORK FOR OBJECT IDENTIFICATION (FOI) - Workflow ENVIRONMENT PERCEPTION – Object Identification
  • 50. • Framework's components includes – Disparity Estimation through the Semi-Global Matching – Motion Estimation by Image Registration and Optical Flow – Moving Object Detection (MOD) – Information Extraction and Fusion DISPARITY ESTIMATION – We adopt a well known Semi-Global Matching (SGM) algorithm. ENVIRONMENT PERCEPTION – Object Identification
  • 51. ENVIRONMENT PERCEPTION – Object Identification MOTION ESTIMATION – One of the most widely used methods is optical flow (OF) estimation. – OF provides satisfactory results when the camera is fixed or carefully displaced. – However, the optical flow from image sequences acquired by a moving camera encodes two pieces of information. • The motion of the surrounding objects • The ego vehicle's motion results in significant motion vectors associated with the static objects, leading to an incorrect perception of static objects as moving objects.
  • 52. MOTION ESTIMATION Approach to Motion Compensation: – By recent trends in aerial and medical imaging, we suggest a procedure to be used called image registration together with the optical flow method to overcome ego-motion and obtain true estimation of the motion information. IMAGE REGISTRATION With Image Registration Without Image Registration ENVIRONMENT PERCEPTION – Object Identification
  • 53. MOTION ESTIMATION Computing Optical Flow ENVIRONMENT PERCEPTION – Object Identification
  • 54. MOVING OBJECT DETECTION (MOD) We followed a straightforward approach for moving object detection, that is • To detect the objects of interest first • Then, identify the moving ones from detected objects from two consecutive frames Current frame Previous frame Moving Objects OBJECTDETECTION • Classification • Localization • Segmentation of all the objects in the scene IDENTIFY MOVING OBJECTS • Recognitionof Moving objectsfromdetectedobjects inthe scene ENVIRONMENT PERCEPTION – Object Identification
  • 55. MOVING OBJECT DETECTION (MOD) E D MOD Moving Objects Mask Moving Object Detection Encoder-Decoder Network Superimpose image frame and corresponding moving object mask, and integrate Bbox and Class information Seg Mask Bbox and Class Segmentation Network The proposed architecture incorporates the object segmentation, and binary pixel classification based on temporal information The object segmentation part is a pre-trained CenterMask Lite segmentation network inference, which gives the bounding boxes, category probabilities, and segmentation mask for each object of interest The temporal processing part is an encoder-decoder network (EDNet) that identifies the moving objects using segmented masks of consecutive frames ENVIRONMENT PERCEPTION – Object Identification
  • 56. MOVING OBJECT DETECTION (MOD) The proposed model labels the moving objects (as white) and the static/background (as black) from sequence pair images. ENVIRONMENT PERCEPTION – Object Identification
  • 57. PROPOSED MOD DATASET We developed a large dataset for moving object detection (from KITTI and EU-life long datasets) covering all the dynamic objects like all types of vehicles, pedestrians, cyclists, motorcyclists, bus, train, and truck. ENVIRONMENT PERCEPTION – Object Identification
  • 58. MOVING OBJECT DETECTION (MOD) - RESULTS ENVIRONMENT PERCEPTION – Object Identification
  • 59. MOVING OBJECT DETECTION (MOD) - RESULTS EU-LT SEQ_02 ENVIRONMENT PERCEPTION – Object Identification
  • 60. FULLY COMPENSATED OPTICAL FLOW (FCOF) ENVIRONMENT PERCEPTION – Object Identification
  • 61. FUSION OF MOD, FCOF AND DISPARITY The results of each stage of the proposed framework, such as disparity, moving object detection, and motion estimation, are fused to extract information such as object ID, static or moving, distance, direction, position, and velocity. – Labelling and Scaling ENVIRONMENT PERCEPTION – Object Identification
  • 62. FOI - Results ENVIRONMENT PERCEPTION – Object Identification
  • 63. FOI - Results ENVIRONMENT PERCEPTION – Object Identification
  • 64. FOI - Results ENVIRONMENT PERCEPTION – Object Identification
  • 65. FOI - Results ENVIRONMENT PERCEPTION – Object Identification
  • 66. FOI - Results KITTI SEQ_05 ENVIRONMENT PERCEPTION – Object Identification
  • 67. FOI - Results KITTI SEQ_84 ENVIRONMENT PERCEPTION – Object Identification
  • 68. FOI - Results KITTI SEQ_71 ENVIRONMENT PERCEPTION – Object Identification
  • 69. FOI - Results KITTI SEQ_59 ENVIRONMENT PERCEPTION – Object Identification
  • 70. CONCLUSION • Proposal of a new framework for object identification (FOI) from a moving camera in complex urban driving environment. • The framework relies only upon the images captured from the stereo camera. • It extracts the information related to the object, including class, status (moving/static), direction, velocity, position, and distance from the ego vehicle. • Other contributions related to Moving Object Detection (MOD) as it is considered as the critical task. – A new dataset for Moving Object Détection is built from existing driving datasets KITTI and EU long-term, covering dynamic objects like all types of vehicles, pedestrians, bicyclists, and motorcyclists. – Propose to use image registration as a tool for ego-motion compensation for urban driving scenarios. – A new model for moving object detection is developed by integrating an encoder-decoder network with a segmentation model. ENVIRONMENT PERCEPTION – Object Identification
  • 71. LIMITATIONS: Through the experiments, it is found that Objects overlap or very close to each other Object reflection Object motion speed is same as ego-vehicle speed ENVIRONMENT PERCEPTION – Object Identification
  • 72. FUTUREWORK – More attention should be paid to improving the overall speed of the proposed framework. – Build a new dataset – Fully Compensated Optical Flow color maps. – Future studies will be devoted to develop a framework that can prioritize objects in a driving scene according to the situation. – We will use the perception data (object identification) to plan secure and smooth trajectories for the objects of interest using their dynamics limits, navigation comfort and safety, and the traffic rules. – Lane detection, traffic sign detection, and live traffic light detection could be integrated into the framework, which would help the system in various tasks such as lane-change, obstacle avoidance, and combined in critical driving situations. ENVIRONMENT PERCEPTION – Object Identification
  • 73. Road-user importance estimation during a left turn maneuver • In real-world driving, there can be a variety of items in the immediate neighborhood of the ego-vehicle at any given time. • Some items have a direct impact on the behavior of the ego-vehicle (e.g., brake, steer), while others have the potential to be a danger, and still others do not pose a danger currently or soon.
  • 74. • The ability to determine how important or relevant a given object is to an ego-decision vehicle's is critical for both driver assistance systems and self-driving vehicles. Establishing trust with human drivers or passengers Demonstrating transparency to law enforcement Promoting a human-centric thought process etc Road-user importance estimation during a left turn maneuver
  • 75. THANKS Contact: Yassine Ruichek Laboratoire CIAD Université de Bourgogne Franche Comté (UBFC) Université de Technologie Belfort Montbéliard (UTBM) yassine.ruichek@utbm.fr– 06 73 22 10 19