Serene 2015
Davide Scaramuzza
Abstract: With drones becoming more and more popular, safety is a big concern. A critical situation occurs when a drone temporarily loses its GPS position information, which might lead it to crash. This can happen, for instance, when flying close to buildings where GPS signal is lost. In such situations, it is desirable that the drone can rely on fall-back systems and regain stable flight as soon as possible. In this talk, I will present novel methods to automatically recover and stabilize a quadrotor from any initial condition or execute emergency landing. On the one hand, this new technology will allow quadrotors to be launched by simply tossing them in the air, like a “baseball ball”. On the other hand, it will allow them to recover back into stable flight or land on a safe area after a system failure. Since this technology does not rely on any external infrastructure, such as GPS, it enables the safe use of drones in both indoor and outdoor environments. Thus, it can become relevant for commercial use of drones, such as parcel delivery.
Recent videos:
Automatic failure recovery without GPS: https://youtu.be/pGU1s6Y55JI
Autonomous Landing-site detection and landing: https://youtu.be/phaBKFwfcJ4
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision GroupLihang Li
This is the slides about DTAM for my group meeting report, hope it does help to anyone who will want to implement DTAM and need to understand it deeply.
This document summarizes a tutorial on replicable evaluation of recommender systems presented at ACM RecSys 2015. The tutorial covered background on recommender systems and motivation for proper evaluation. It discussed evaluating recommender systems as a "black box" process involving data splitting, recommendation generation, candidate item selection, and metric computation. The presenters emphasized the importance of replicating and reproducing evaluation results to validate findings and advance the field. They provided guidelines for reproducible experimental design and highlighted the need to distinguish between replicability and reproducibility. The tutorial included a demonstration of replicating results and concluded by discussing next steps like agreeing on standard implementations and incentivizing reproducibility.
This document provides an overview of recommendation systems including: how they work using content-based, collaborative filtering, and hybrid approaches; challenges like performance, modelization, and biases; common evaluation methods including offline metrics, A/B testing, and the online-offline gap; interactive learning using reinforcement learning; and concludes with best practices like prioritizing criteria, blending approaches, and continuous evaluation and monitoring. A hands-on tutorial is also proposed to implement a neural network-based recommendation system.
Hot Stand-By Disaster Recovery Solutions for Ensuring the Resilience of Railw...SERENEWorkshop
Specifications of modern railway control systems often include resilience requirements in order to quickly and safely recovery from disasters (e.g. system-level failures). To that aim, spatial redundancy is required, with main and backup systems installed in fully isolated buildings, together with very short switchover times from main to backup systems in case of disasters. In order to fulfil those requirements, Ansaldo STS has developed a system-level hot stand-by solution allowing to quickly and smoothly switch from the main system to the back-up one, ensuring the necessary continuity of service and transparency to train supervisors and other operators. The functional architecture of such a solution is able to keep aligned the safety-critical nucleuses, typically based on N-modular redundancy (i.e. ‘KooM’ voting), of the main and the back-up systems. Such a coherent alignment must be kept in terms of both interfaced field devices (e.g. interlocking signals, track circuits, switch points, etc.) – on the ‘bottom’ level – and control room Human Machine Interfaces (HMI) – on the ‘top’ level. The solution is based on heterogeneous and redundant network links (copper/fiber Ethernet/HyperRing) at different levels of system architecture. In this speech, the reference architecture and the fault-tolerance functionalities for disaster recovery are provided, considering the requirements of real railway and mass-transit installations.
This document describes a project to develop a quadrotor drone capable of autonomous navigation in GPS-denied environments using visual sensors. The drone is equipped with an optical flow sensor for velocity and altitude control, an IMU, and cameras for visual SLAM. Software runs ROS and uses a monocular visual odometry algorithm. Initial flight tests show the drone can hold position and velocity with wind. Future work will integrate the visual SLAM for full position control and add safety features.
This document discusses the history and fundamentals of visual odometry (VO) and simultaneous localization and mapping (SLAM). It provides an overview of key developments in VO from the 1980s to present day, including the first real-time VO implementation on a robot in 1980 and use of VO on the Mars rovers in 2004. The document also summarizes the differences between VO, SFM, and V-SLAM, and describes common approaches to feature extraction, motion estimation, and optimization in VO pipelines.
IRJET- Behavior Analysis from Videos using Motion based Feature ExtractionIRJET Journal
This document proposes a technique for analyzing human behavior in videos using motion-based feature extraction. It discusses how previous approaches have used spatial and temporal features to detect abnormal behaviors. The proposed approach extracts motion features from videos to represent each video with a single feature vector, rather than extracting features from each individual frame. This reduces the feature space and unnecessary information. The technique involves preprocessing videos into frames, extracting motion features, using KNN classification on the features to classify behaviors as normal or abnormal, and evaluating the method's performance on various metrics like accuracy, recall, and precision. Testing on fight and riot datasets showed the motion-based approach achieved higher accuracy, recall, precision and F-measure than a non-motion based approach.
Deconstructing SfM-Net architecture and beyond
"SfM-Net, a geometry-aware neural network for motion estimation in videos that decomposes frame-to-frame pixel motion in terms of scene and object depth, camera motion and 3D object rotations and translations. Given a sequence of frames, SfM-Net predicts depth, segmentation, camera and rigid object motions, converts those into a dense frame-to-frame motion field (optical flow), differentiably warps frames in time to match pixels and back-propagates."
Alternative download:
https://www.dropbox.com/s/aezl7ro8sy2xq7j/sfm_net_v2.pdf?dl=0
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision GroupLihang Li
This is the slides about DTAM for my group meeting report, hope it does help to anyone who will want to implement DTAM and need to understand it deeply.
This document summarizes a tutorial on replicable evaluation of recommender systems presented at ACM RecSys 2015. The tutorial covered background on recommender systems and motivation for proper evaluation. It discussed evaluating recommender systems as a "black box" process involving data splitting, recommendation generation, candidate item selection, and metric computation. The presenters emphasized the importance of replicating and reproducing evaluation results to validate findings and advance the field. They provided guidelines for reproducible experimental design and highlighted the need to distinguish between replicability and reproducibility. The tutorial included a demonstration of replicating results and concluded by discussing next steps like agreeing on standard implementations and incentivizing reproducibility.
This document provides an overview of recommendation systems including: how they work using content-based, collaborative filtering, and hybrid approaches; challenges like performance, modelization, and biases; common evaluation methods including offline metrics, A/B testing, and the online-offline gap; interactive learning using reinforcement learning; and concludes with best practices like prioritizing criteria, blending approaches, and continuous evaluation and monitoring. A hands-on tutorial is also proposed to implement a neural network-based recommendation system.
Hot Stand-By Disaster Recovery Solutions for Ensuring the Resilience of Railw...SERENEWorkshop
Specifications of modern railway control systems often include resilience requirements in order to quickly and safely recovery from disasters (e.g. system-level failures). To that aim, spatial redundancy is required, with main and backup systems installed in fully isolated buildings, together with very short switchover times from main to backup systems in case of disasters. In order to fulfil those requirements, Ansaldo STS has developed a system-level hot stand-by solution allowing to quickly and smoothly switch from the main system to the back-up one, ensuring the necessary continuity of service and transparency to train supervisors and other operators. The functional architecture of such a solution is able to keep aligned the safety-critical nucleuses, typically based on N-modular redundancy (i.e. ‘KooM’ voting), of the main and the back-up systems. Such a coherent alignment must be kept in terms of both interfaced field devices (e.g. interlocking signals, track circuits, switch points, etc.) – on the ‘bottom’ level – and control room Human Machine Interfaces (HMI) – on the ‘top’ level. The solution is based on heterogeneous and redundant network links (copper/fiber Ethernet/HyperRing) at different levels of system architecture. In this speech, the reference architecture and the fault-tolerance functionalities for disaster recovery are provided, considering the requirements of real railway and mass-transit installations.
This document describes a project to develop a quadrotor drone capable of autonomous navigation in GPS-denied environments using visual sensors. The drone is equipped with an optical flow sensor for velocity and altitude control, an IMU, and cameras for visual SLAM. Software runs ROS and uses a monocular visual odometry algorithm. Initial flight tests show the drone can hold position and velocity with wind. Future work will integrate the visual SLAM for full position control and add safety features.
This document discusses the history and fundamentals of visual odometry (VO) and simultaneous localization and mapping (SLAM). It provides an overview of key developments in VO from the 1980s to present day, including the first real-time VO implementation on a robot in 1980 and use of VO on the Mars rovers in 2004. The document also summarizes the differences between VO, SFM, and V-SLAM, and describes common approaches to feature extraction, motion estimation, and optimization in VO pipelines.
IRJET- Behavior Analysis from Videos using Motion based Feature ExtractionIRJET Journal
This document proposes a technique for analyzing human behavior in videos using motion-based feature extraction. It discusses how previous approaches have used spatial and temporal features to detect abnormal behaviors. The proposed approach extracts motion features from videos to represent each video with a single feature vector, rather than extracting features from each individual frame. This reduces the feature space and unnecessary information. The technique involves preprocessing videos into frames, extracting motion features, using KNN classification on the features to classify behaviors as normal or abnormal, and evaluating the method's performance on various metrics like accuracy, recall, and precision. Testing on fight and riot datasets showed the motion-based approach achieved higher accuracy, recall, precision and F-measure than a non-motion based approach.
Deconstructing SfM-Net architecture and beyond
"SfM-Net, a geometry-aware neural network for motion estimation in videos that decomposes frame-to-frame pixel motion in terms of scene and object depth, camera motion and 3D object rotations and translations. Given a sequence of frames, SfM-Net predicts depth, segmentation, camera and rigid object motions, converts those into a dense frame-to-frame motion field (optical flow), differentiably warps frames in time to match pixels and back-propagates."
Alternative download:
https://www.dropbox.com/s/aezl7ro8sy2xq7j/sfm_net_v2.pdf?dl=0
Person Detection in Maritime Search And Rescue OperationsIRJET Journal
This document discusses recent research on using computer vision and machine learning techniques for person detection in maritime search and rescue operations from images and video captured by drones. Specifically, it summarizes 12 research papers on this topic, covering approaches such as training convolutional neural networks on bird's eye view datasets to detect people from aerial images, using multiple detection methods like sliding windows and precise localization, combining data from multiple drones and sensors to optimize search efforts, and evaluating models on both RGB and thermal image datasets. The goal of this research is to automate part of the search process to make maritime rescue operations more efficient and effective.
Person Detection in Maritime Search And Rescue OperationsIRJET Journal
1) The document discusses using machine learning and computer vision techniques for person detection in maritime search and rescue operations using drones/UAVs. It aims to automatically detect people in images/videos captured by drones to help with search efforts.
2) A key challenge is that people appear small in drone footage and are often obscured by vegetation or terrain. The models need to be trained on similar bird's eye view data to achieve high accuracy. The document reviews different person detection models and their use in search and rescue.
3) It discusses recent work involving using efficient neural networks like MobileNet for object detection from drones. Other work involves using depth sensors and pose estimation for person tracking, as well as using distributed deep learning
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...IRJET Journal
The document proposes a system to detect suspicious human activity in crowdsourced video captured by surveillance cameras. The system uses Advanced Motion Detection (AMD) to detect moving objects and generate a reliable background model for analysis. A camera connected to a monitoring room would produce alert messages for any detected suspicious activity based on height, time, and body movement constraints. The system aims to automate real-time video processing for security applications like detecting unauthorized access. It extracts human objects from frames and identifies suspicious behavior using the AMD algorithm before sending alerts.
The document describes a blind assistance system called Sanjaya that uses object detection and depth estimation to help visually impaired individuals navigate environments. The system uses a SSD MobileNet model trained on the COCO dataset via TensorFlow's object detection API to identify objects in camera images in real-time. It then uses depth estimation to calculate distances and provides voice feedback alerts to users about detected objects and their proximity. The system aims to allow visually impaired people to have improved comprehension of their surroundings and navigation abilities.
Inertial Navigation for Quadrotor Using Kalman Filter with Drift Compensation IJECEIAES
The main disadvantage of an Inertial Navigation System is a low accuracy due to noise, bias, and drift error in the inertial sensor. This research aims to develop the accelerometer and gyroscope sensor for quadrotor navigation system, bias compensation, and Zero Velocity Compensation (ZVC). Kalman Filter is designed to reduce the noise on the sensor while bias compensation and ZVC are designed to eliminate the bias and drift error in the sensor data. Test results showed the Kalman Filter design is acceptable to reduce the noise in the sensor data. Moreover, the bias compensation and ZVC can reduce the drift error due to integration process as well as improve the position estimation accuracy of the quadrotor. At the time of testing, the system provided the accuracy above 90 % when it tested indoor.
Real Time Detection of Moving Object Based on Fpgaiosrjce
IOSR Journal of Electronics and Communication Engineering(IOSR-JECE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of electronics and communication engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electronics and communication engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
1. The document describes a method for real-time detection of moving objects based on background subtraction and its implementation on an FPGA. A static camera is used to capture video frames. The first frame is used as the reference background frame. Pixels in subsequent frames are compared to the background frame and objects are detected where pixel differences exceed a threshold.
2. The method was tested on sample surveillance videos. Background subtraction accurately detected moving objects in test videos in real-time. Future work may include identifying objects using face or palm recognition and activity recognition for visual surveillance applications.
IRJET- Criminal Recognization in CCTV Surveillance VideoIRJET Journal
This document presents research on criminal recognition in CCTV surveillance videos using deep learning. It proposes a method where a user can upload faces of known criminals. When CCTV footage is recorded, the application will monitor for these faces. If a face is recognized, the CCTV camera will track the identified person through multiple cameras by alerting other cameras. The system segments video into images, acquires images, recognizes human faces, constructs motion flows between cameras to track individuals. Experimental results on a dataset show the system's ability to extract patterns from faces and cluster images of different angled faces. The system aims to identify criminals across surveillance camera networks.
Human Motion Detection in Video Surveillance using Computer Vision TechniqueIRJET Journal
The document discusses a technique for detecting human motion in video surveillance using computer vision. It proposes a method called DECOLOR (Detecting Contiguous Outliers in the LOw-rank Representation) that formulates object detection as outlier detection in a low-rank representation of video frames. This allows it to detect moving objects flexibly without assumptions about foreground or background behavior. DECOLOR simultaneously performs object detection and background estimation using only the test video sequence, without requiring training data. The method models the outlier support explicitly and favors spatially contiguous outliers, making it suitable for detecting clustered foreground objects like people. It achieves more accurate detection and background estimation than state-of-the-art robust principal component analysis methods.
IRJET- Robust and Fast Detection of Moving Vechiles in Aerial Videos usin...IRJET Journal
This document discusses a method for detecting moving vehicles in aerial videos using sliding windows in MATLAB. It first performs foreground motion segmentation and trajectory accumulation to build a dynamic scene motion heat map. It then introduces a novel saliency-based background model that enhances moving objects to segment vehicles in high-heat areas. Key steps include compensating for camera motion, detecting independent motion via tracking image corners, and applying a sliding window classifier with optimized parameters like window size and orientation. Evaluation on urban videos shows the approach achieves 88% detection rates with only 2% false positives, outperforming alternative segmentation-based methods in processing time and accuracy. The goal is robust, real-time vehicle detection from aerial videos for applications like traffic monitoring.
Visual Mapping and Collision Avoidance Dynamic Environments in Dynamic Enviro...Darius Burschka
How conventional vision is more appropriate for control since it provides also error analysis. There is a lot of information in the images that is lost when converting to 3D
The document summarizes a study that compares the performance of camera-only, IMU-only, and combined camera-IMU systems for visual inertial navigation (VisNav) and simultaneous localization and mapping (SLAM). Data was collected from a camera and IMU mounted on a rolling platform traversing a test course. The camera performed poorly around corners while the IMU had offset errors, but combining the sensors was superior to either individually. Statistical analysis showed the mean error was significantly lower for the camera-IMU system compared to the camera-only and IMU-only systems. The study was limited but integrating the sensors helped compensate their individual weaknesses.
This document presents a method for tracking moving objects in video sequences using affine flow parameters combined with illumination insensitive template matching. The method extracts affine flow parameters from frames to model local object motion using affine transformations. It then applies template matching with illumination compensation to track objects across frames while being robust to illumination changes. The method is evaluated on various indoor and outdoor database videos and is shown to effectively track objects without false detections, handling issues like illumination variations, camera motion and dynamic backgrounds better than other methods.
This paper proposes a way of recognition foreground of moving object by Foreground Extraction
algorithm by Pan-Tilt-Zoom camera. It presents the combined process of Foreground Extraction and local
histogram process. Background images are often modeled as multiple frames and their corresponding
camera pan and tilt angles are determined. Initially have got to work out the foremost matchedbackground
from sequence of input frames based on camera pose information. The method is more continued by
compensating the matched background image with this image. Then Background Subtraction is completed
between modeled background and current image. Finally before local histogram process is completed,
noises are often removed by morphological operators. As a result correct foreground moving objects are
successfully extracted by implementing these four steps.
DETECTION OF MOVING OBJECT USING FOREGROUND EXTRACTION ALGORITHM BY PTZ CAMERAijistjournal
The document summarizes a proposed method for detecting moving objects in video sequences captured by a Pan-Tilt-Zoom (PTZ) camera. The method uses foreground extraction and local histogram processing. It first finds the most matched background frame using Euclidean distance. SURF keypoints are extracted and used to compute a homography matrix for background compensation. Foreground is extracted by subtracting the compensated background from the current frame. Morphological operations and local histogram processing are applied to remove noise and obtain the final extracted foreground containing the moving objects. The method is evaluated on a public dataset and is shown to achieve accurate segmentation when combining foreground extraction with local histogram processing.
An Innovative Moving Object Detection and Tracking System by Using Modified R...sipij
The ultimate goal of this study is to afford enhanced video object detection and tracking by eliminating the
limitations which are existing nowadays. Although high performance ratio for video object detection and
tracking is achieved in the earlier work it takes more time for computation. Consequently we are in need to
propose a novel video object detection and tracking technique so as to minimize the computational
complexity. Our proposed technique covers five stages they are preprocessing, segmentation, feature
extraction, background subtraction and hole filling. Originally the video clip in the database is split into
frames. Then preprocessing is performed so as to get rid of noise, an adaptive median filter is used in this
stage to eliminate the noise. The preprocessed image then undergoes segmentation by means of modified
region growing algorithm. The segmented image is subjected to feature extraction phase so as to extract
the multi features from the segmented image and the background image, the feature value thus obtained
are compared so as to attain optimal value, consequently a foreground image is attained in this stage. The
foreground image is then subjected to morphological operations of erosion and dilation so as to fill the
holes and to get the object accurately as these foreground image contains holes and discontinuities. Thus
the moving object is tracked in this stage. This method will be employed in MATLAB platform and the
outcomes will be studied and compared with the existing techniques so as to reveal the performance of the
novel video object detection and tracking technique.
IRJET- A Review Analysis to Detect an Object in Video Surveillance SystemIRJET Journal
This document reviews techniques for detecting objects in video surveillance systems. It discusses common object detection methods like frame differencing, optical flow, and background subtraction. Frame differencing detects motion by calculating pixel differences between frames but cannot detect still objects. Optical flow estimates pixel motion between frames to detect objects. Background subtraction models the static background and detects objects by subtracting current frames from the background model. The document analyzes these techniques and their use in video surveillance applications like traffic monitoring and security. It concludes more research is needed to improve object classification accuracy and handle challenges like camera motion.
Robust and Efficient Coupling of Perception to Actuation with Metric and Non-...Darius Burschka
The talk motivates a re-thinking of the way, how perception passes the information to the control modules. Metric information is not a native space of the camera and apparently also not used in biology for navigation. Early abstraction of information from images loses a lot of important information that can be directly used for following (Visual-Servoing), motion estimation(Motion Blurr), and collision relations(Optical Flow Clustering). I present in this talk ways, how we use the image information in "classical way" that does not require any learning and runs on low-power CPUs.
Considering Execution Environment Resilience: A White-Box ApproachSERENEWorkshop
The document discusses an approach called semi-purification for automatically generating unit test cases from source code. Semi-purification replaces dependencies like global variables and database calls in the source code with function parameters. This allows existing automated test case generation tools to be used by treating the semi-purified code as if it were pure. Challenges discussed include handling shared subroutines, loops, and concurrency. The goal is to increase test coverage for complex, distributed systems with frequent changes like those used at CERN.
Engineering Cross-Layer Fault Tolerance in Many-Core SystemsSERENEWorkshop
1) The document discusses engineering cross-layer fault tolerance in many-core systems. It proposes a cross-layer approach where fault tolerance is distributed across system layers rather than handled solely within a single layer.
2) A motivating example of cross-layer fault tolerance is discussed using TCP/IP, where errors can be detected and recovered across multiple layers for improved efficiency.
3) The challenges of ensuring cross-layer fault tolerance for many-core systems containing tens to thousands of cores are discussed to improve reliability, performance and energy efficiency.
4) The plan is to implement a case study of a car number plate recognition application to gain experience with cross-layer fault tolerance, and apply order graphs to model performance
More Related Content
Similar to Towards Robust and Safe Autonomous Drones
Person Detection in Maritime Search And Rescue OperationsIRJET Journal
This document discusses recent research on using computer vision and machine learning techniques for person detection in maritime search and rescue operations from images and video captured by drones. Specifically, it summarizes 12 research papers on this topic, covering approaches such as training convolutional neural networks on bird's eye view datasets to detect people from aerial images, using multiple detection methods like sliding windows and precise localization, combining data from multiple drones and sensors to optimize search efforts, and evaluating models on both RGB and thermal image datasets. The goal of this research is to automate part of the search process to make maritime rescue operations more efficient and effective.
Person Detection in Maritime Search And Rescue OperationsIRJET Journal
1) The document discusses using machine learning and computer vision techniques for person detection in maritime search and rescue operations using drones/UAVs. It aims to automatically detect people in images/videos captured by drones to help with search efforts.
2) A key challenge is that people appear small in drone footage and are often obscured by vegetation or terrain. The models need to be trained on similar bird's eye view data to achieve high accuracy. The document reviews different person detection models and their use in search and rescue.
3) It discusses recent work involving using efficient neural networks like MobileNet for object detection from drones. Other work involves using depth sensors and pose estimation for person tracking, as well as using distributed deep learning
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...IRJET Journal
The document proposes a system to detect suspicious human activity in crowdsourced video captured by surveillance cameras. The system uses Advanced Motion Detection (AMD) to detect moving objects and generate a reliable background model for analysis. A camera connected to a monitoring room would produce alert messages for any detected suspicious activity based on height, time, and body movement constraints. The system aims to automate real-time video processing for security applications like detecting unauthorized access. It extracts human objects from frames and identifies suspicious behavior using the AMD algorithm before sending alerts.
The document describes a blind assistance system called Sanjaya that uses object detection and depth estimation to help visually impaired individuals navigate environments. The system uses a SSD MobileNet model trained on the COCO dataset via TensorFlow's object detection API to identify objects in camera images in real-time. It then uses depth estimation to calculate distances and provides voice feedback alerts to users about detected objects and their proximity. The system aims to allow visually impaired people to have improved comprehension of their surroundings and navigation abilities.
Inertial Navigation for Quadrotor Using Kalman Filter with Drift Compensation IJECEIAES
The main disadvantage of an Inertial Navigation System is a low accuracy due to noise, bias, and drift error in the inertial sensor. This research aims to develop the accelerometer and gyroscope sensor for quadrotor navigation system, bias compensation, and Zero Velocity Compensation (ZVC). Kalman Filter is designed to reduce the noise on the sensor while bias compensation and ZVC are designed to eliminate the bias and drift error in the sensor data. Test results showed the Kalman Filter design is acceptable to reduce the noise in the sensor data. Moreover, the bias compensation and ZVC can reduce the drift error due to integration process as well as improve the position estimation accuracy of the quadrotor. At the time of testing, the system provided the accuracy above 90 % when it tested indoor.
Real Time Detection of Moving Object Based on Fpgaiosrjce
IOSR Journal of Electronics and Communication Engineering(IOSR-JECE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of electronics and communication engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electronics and communication engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
1. The document describes a method for real-time detection of moving objects based on background subtraction and its implementation on an FPGA. A static camera is used to capture video frames. The first frame is used as the reference background frame. Pixels in subsequent frames are compared to the background frame and objects are detected where pixel differences exceed a threshold.
2. The method was tested on sample surveillance videos. Background subtraction accurately detected moving objects in test videos in real-time. Future work may include identifying objects using face or palm recognition and activity recognition for visual surveillance applications.
IRJET- Criminal Recognization in CCTV Surveillance VideoIRJET Journal
This document presents research on criminal recognition in CCTV surveillance videos using deep learning. It proposes a method where a user can upload faces of known criminals. When CCTV footage is recorded, the application will monitor for these faces. If a face is recognized, the CCTV camera will track the identified person through multiple cameras by alerting other cameras. The system segments video into images, acquires images, recognizes human faces, constructs motion flows between cameras to track individuals. Experimental results on a dataset show the system's ability to extract patterns from faces and cluster images of different angled faces. The system aims to identify criminals across surveillance camera networks.
Human Motion Detection in Video Surveillance using Computer Vision TechniqueIRJET Journal
The document discusses a technique for detecting human motion in video surveillance using computer vision. It proposes a method called DECOLOR (Detecting Contiguous Outliers in the LOw-rank Representation) that formulates object detection as outlier detection in a low-rank representation of video frames. This allows it to detect moving objects flexibly without assumptions about foreground or background behavior. DECOLOR simultaneously performs object detection and background estimation using only the test video sequence, without requiring training data. The method models the outlier support explicitly and favors spatially contiguous outliers, making it suitable for detecting clustered foreground objects like people. It achieves more accurate detection and background estimation than state-of-the-art robust principal component analysis methods.
IRJET- Robust and Fast Detection of Moving Vechiles in Aerial Videos usin...IRJET Journal
This document discusses a method for detecting moving vehicles in aerial videos using sliding windows in MATLAB. It first performs foreground motion segmentation and trajectory accumulation to build a dynamic scene motion heat map. It then introduces a novel saliency-based background model that enhances moving objects to segment vehicles in high-heat areas. Key steps include compensating for camera motion, detecting independent motion via tracking image corners, and applying a sliding window classifier with optimized parameters like window size and orientation. Evaluation on urban videos shows the approach achieves 88% detection rates with only 2% false positives, outperforming alternative segmentation-based methods in processing time and accuracy. The goal is robust, real-time vehicle detection from aerial videos for applications like traffic monitoring.
Visual Mapping and Collision Avoidance Dynamic Environments in Dynamic Enviro...Darius Burschka
How conventional vision is more appropriate for control since it provides also error analysis. There is a lot of information in the images that is lost when converting to 3D
The document summarizes a study that compares the performance of camera-only, IMU-only, and combined camera-IMU systems for visual inertial navigation (VisNav) and simultaneous localization and mapping (SLAM). Data was collected from a camera and IMU mounted on a rolling platform traversing a test course. The camera performed poorly around corners while the IMU had offset errors, but combining the sensors was superior to either individually. Statistical analysis showed the mean error was significantly lower for the camera-IMU system compared to the camera-only and IMU-only systems. The study was limited but integrating the sensors helped compensate their individual weaknesses.
This document presents a method for tracking moving objects in video sequences using affine flow parameters combined with illumination insensitive template matching. The method extracts affine flow parameters from frames to model local object motion using affine transformations. It then applies template matching with illumination compensation to track objects across frames while being robust to illumination changes. The method is evaluated on various indoor and outdoor database videos and is shown to effectively track objects without false detections, handling issues like illumination variations, camera motion and dynamic backgrounds better than other methods.
This paper proposes a way of recognition foreground of moving object by Foreground Extraction
algorithm by Pan-Tilt-Zoom camera. It presents the combined process of Foreground Extraction and local
histogram process. Background images are often modeled as multiple frames and their corresponding
camera pan and tilt angles are determined. Initially have got to work out the foremost matchedbackground
from sequence of input frames based on camera pose information. The method is more continued by
compensating the matched background image with this image. Then Background Subtraction is completed
between modeled background and current image. Finally before local histogram process is completed,
noises are often removed by morphological operators. As a result correct foreground moving objects are
successfully extracted by implementing these four steps.
DETECTION OF MOVING OBJECT USING FOREGROUND EXTRACTION ALGORITHM BY PTZ CAMERAijistjournal
The document summarizes a proposed method for detecting moving objects in video sequences captured by a Pan-Tilt-Zoom (PTZ) camera. The method uses foreground extraction and local histogram processing. It first finds the most matched background frame using Euclidean distance. SURF keypoints are extracted and used to compute a homography matrix for background compensation. Foreground is extracted by subtracting the compensated background from the current frame. Morphological operations and local histogram processing are applied to remove noise and obtain the final extracted foreground containing the moving objects. The method is evaluated on a public dataset and is shown to achieve accurate segmentation when combining foreground extraction with local histogram processing.
An Innovative Moving Object Detection and Tracking System by Using Modified R...sipij
The ultimate goal of this study is to afford enhanced video object detection and tracking by eliminating the
limitations which are existing nowadays. Although high performance ratio for video object detection and
tracking is achieved in the earlier work it takes more time for computation. Consequently we are in need to
propose a novel video object detection and tracking technique so as to minimize the computational
complexity. Our proposed technique covers five stages they are preprocessing, segmentation, feature
extraction, background subtraction and hole filling. Originally the video clip in the database is split into
frames. Then preprocessing is performed so as to get rid of noise, an adaptive median filter is used in this
stage to eliminate the noise. The preprocessed image then undergoes segmentation by means of modified
region growing algorithm. The segmented image is subjected to feature extraction phase so as to extract
the multi features from the segmented image and the background image, the feature value thus obtained
are compared so as to attain optimal value, consequently a foreground image is attained in this stage. The
foreground image is then subjected to morphological operations of erosion and dilation so as to fill the
holes and to get the object accurately as these foreground image contains holes and discontinuities. Thus
the moving object is tracked in this stage. This method will be employed in MATLAB platform and the
outcomes will be studied and compared with the existing techniques so as to reveal the performance of the
novel video object detection and tracking technique.
IRJET- A Review Analysis to Detect an Object in Video Surveillance SystemIRJET Journal
This document reviews techniques for detecting objects in video surveillance systems. It discusses common object detection methods like frame differencing, optical flow, and background subtraction. Frame differencing detects motion by calculating pixel differences between frames but cannot detect still objects. Optical flow estimates pixel motion between frames to detect objects. Background subtraction models the static background and detects objects by subtracting current frames from the background model. The document analyzes these techniques and their use in video surveillance applications like traffic monitoring and security. It concludes more research is needed to improve object classification accuracy and handle challenges like camera motion.
Robust and Efficient Coupling of Perception to Actuation with Metric and Non-...Darius Burschka
The talk motivates a re-thinking of the way, how perception passes the information to the control modules. Metric information is not a native space of the camera and apparently also not used in biology for navigation. Early abstraction of information from images loses a lot of important information that can be directly used for following (Visual-Servoing), motion estimation(Motion Blurr), and collision relations(Optical Flow Clustering). I present in this talk ways, how we use the image information in "classical way" that does not require any learning and runs on low-power CPUs.
Similar to Towards Robust and Safe Autonomous Drones (20)
Considering Execution Environment Resilience: A White-Box ApproachSERENEWorkshop
The document discusses an approach called semi-purification for automatically generating unit test cases from source code. Semi-purification replaces dependencies like global variables and database calls in the source code with function parameters. This allows existing automated test case generation tools to be used by treating the semi-purified code as if it were pure. Challenges discussed include handling shared subroutines, loops, and concurrency. The goal is to increase test coverage for complex, distributed systems with frequent changes like those used at CERN.
Engineering Cross-Layer Fault Tolerance in Many-Core SystemsSERENEWorkshop
1) The document discusses engineering cross-layer fault tolerance in many-core systems. It proposes a cross-layer approach where fault tolerance is distributed across system layers rather than handled solely within a single layer.
2) A motivating example of cross-layer fault tolerance is discussed using TCP/IP, where errors can be detected and recovered across multiple layers for improved efficiency.
3) The challenges of ensuring cross-layer fault tolerance for many-core systems containing tens to thousands of cores are discussed to improve reliability, performance and energy efficiency.
4) The plan is to implement a case study of a car number plate recognition application to gain experience with cross-layer fault tolerance, and apply order graphs to model performance
This document presents research on assessing risk to determine the optimal level of redundancy needed when moving critical applications to the cloud. It develops fault tree models based on the physical structure of clouds to calculate failure frequency. It factors in varying resource quality and the costs of downtime and VMs. The results show that deploying between 4-6 redundant VMs provides significant availability gains and reduces total costs by lowering risk compared to basic redundancy approaches. The aim was met of leveraging cloud features in modeling to support high-value, mission critical applications on public clouds.
Biological Immunity and Software Resilience: Two Faces of the Same Coin?SERENEWorkshop
The document discusses the similarities between biological immunity and software resilience. It proposes that biological systems are resilient, with the immune system being a prime example due to its ability to adapt, make decisions through distributed agents, and defend the body through learning. An actor-based model is presented as a way to engineer resilience into software by drawing inspiration from immune system principles like replication, containment, and delegation. A bio-inspired architecture is described that uses supervisor actors to detect changes and spawn helper/killer actors to address issues while maintaining system function. Future work areas are identified like automatic failure recognition, dynamic learning, and multi-layer management of failures.
This document summarizes a presentation on system-level concurrent error detection. It discusses specifying reliability constraints in system specifications, design methodologies that provide error detection capabilities through redundancy, and a two-level hardware/software partitioning approach that first considers traditional costs and then analyzes reliability constraints. The goal is to adopt design for reliability approaches earlier in the system design process to significantly impact costs like timing, energy and area.
SERENE 2014 School: Measurement-Driven Resilience Design of Cloud-Based Cyber...SERENEWorkshop
SERENE 2014 School on Engineering Resilient Cyber Physical Systems
Talk: Measurement-Driven Resilience Design of Cloud-Based Cyber-Physical Systems, by Imre Kocsis
SERENE 2014 School: Resilience in Cyber-Physical Systems: Challenges and Oppo...SERENEWorkshop
SERENE 2014 School on Engineering Resilient Cyber Physical Systems
Talk: Resilience in Cyber-Physical Systems: Challenges and Opportunities, by Gabor Karsai
SERENE 2014 Workshop: Panel on "Views on Runtime Resilience Assessment of Dyn...SERENEWorkshop
The document summarizes a panel discussion on views of runtime resilience assessment of dynamic software systems held at SERENE 2014 in Budapest, Hungary. The panelists represented different domains related to resilience assessment, software engineering, dynamic systems design, and dependable computing. They discussed key challenges around metrics for characterizing resilience, defining dynamic workloads and changeloads, monitoring unbounded and dynamic systems, maintaining accurate runtime models, and standardizing resilience assessment techniques. The panelists emphasized the need for predictive monitoring and adaptation, rather than just detection, to ensure resilience in increasingly complex and evolving software systems.
SERENE 2014 Workshop: Paper "Combined Error Propagation Analysis and Runtime ...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 4: Monitoring
Paper 3: Combined Error Propagation Analysis and Runtime Event Detection in Process-driven Systems
SERENE 2014 Workshop: Paper "Simulation Testing and Model Checking: A Case St...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 3: Verification and Validation
Paper 2: Simulation Testing and Model Checking: A Case Study Comparing these Approaches
SERENE 2014 Workshop: Paper "Adaptive Domain-Specific Service Monitoring"SERENEWorkshop
- An adaptive service monitoring approach that considers domain-specific errors, such as codec errors for streamed media, in addition to generic errors.
- The approach adapts the monitoring frequency for a particular service and error type based on the historical error rate to reduce monitoring costs.
- An evaluation using real-world data from Smart TV services found that the adaptive approach reduced monitoring costs by 30% with negligible impact on error detection quality.
SERENE 2014 Workshop: Paper "Verification and Validation of a Pressure Contro...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 3: Verification and Validation
Paper 1: Verification and Validation of a Pressure Control Unit for Hydraulic Systems
SERENE 2014 Workshop: Paper "Using Instrumentation for Quality Assessment of ...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 4: Monitoring
Paper 1: Using Instrumentation for Quality Assessment of Resilient Software in Embedded Systems
SERENE 2014 Workshop: Paper "Advanced Modelling, Simulation and Verification ...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 3: Verification and Validation
Paper 3: Advanced Modelling, Simulation and Verification for Future Traffic Regulation Optimisation
SERENE 2014 Workshop: Paper "Formal Fault Tolerance Analysis of Algorithms fo...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 2: Analysis of Resilience
Paper : Formal Fault Tolerance Analysis of Algorithms for Redundant Systems in Early Design Stages
SERENE 2014 Workshop: Paper "Modelling Resilience of Data Processing Capabili...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 2: Analysis of Resilience
Paper 1: Modelling Resilience of Data Processing Capabilities of CPS
SERENE 2014 Workshop: Paper "Automatic Generation of Description Files for Hi...SERENEWorkshop
SERENE 2014 - 6th International Workshop on Software Engineering for Resilient Systems
http://serene.disim.univaq.it/
Session 1: Design of Resilient Systems
Paper 3: "Automatic Generation of Description Files for Highly Available Systems"
Levelised Cost of Hydrogen (LCOH) Calculator ManualMassimo Talia
The aim of this manual is to explain the
methodology behind the Levelized Cost of
Hydrogen (LCOH) calculator. Moreover, this
manual also demonstrates how the calculator
can be used for estimating the expenses associated with hydrogen production in Europe
using low-temperature electrolysis considering different sources of electricity
A high-Speed Communication System is based on the Design of a Bi-NoC Router, ...DharmaBanothu
The Network on Chip (NoC) has emerged as an effective
solution for intercommunication infrastructure within System on
Chip (SoC) designs, overcoming the limitations of traditional
methods that face significant bottlenecks. However, the complexity
of NoC design presents numerous challenges related to
performance metrics such as scalability, latency, power
consumption, and signal integrity. This project addresses the
issues within the router's memory unit and proposes an enhanced
memory structure. To achieve efficient data transfer, FIFO buffers
are implemented in distributed RAM and virtual channels for
FPGA-based NoC. The project introduces advanced FIFO-based
memory units within the NoC router, assessing their performance
in a Bi-directional NoC (Bi-NoC) configuration. The primary
objective is to reduce the router's workload while enhancing the
FIFO internal structure. To further improve data transfer speed,
a Bi-NoC with a self-configurable intercommunication channel is
suggested. Simulation and synthesis results demonstrate
guaranteed throughput, predictable latency, and equitable
network access, showing significant improvement over previous
designs
Open Channel Flow: fluid flow with a free surfaceIndrajeet sahu
Open Channel Flow: This topic focuses on fluid flow with a free surface, such as in rivers, canals, and drainage ditches. Key concepts include the classification of flow types (steady vs. unsteady, uniform vs. non-uniform), hydraulic radius, flow resistance, Manning's equation, critical flow conditions, and energy and momentum principles. It also covers flow measurement techniques, gradually varied flow analysis, and the design of open channels. Understanding these principles is vital for effective water resource management and engineering applications.
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...PriyankaKilaniya
Energy efficiency has been important since the latter part of the last century. The main object of this survey is to determine the energy efficiency knowledge among consumers. Two separate districts in Bangladesh are selected to conduct the survey on households and showrooms about the energy and seller also. The survey uses the data to find some regression equations from which it is easy to predict energy efficiency knowledge. The data is analyzed and calculated based on five important criteria. The initial target was to find some factors that help predict a person's energy efficiency knowledge. From the survey, it is found that the energy efficiency awareness among the people of our country is very low. Relationships between household energy use behaviors are estimated using a unique dataset of about 40 households and 20 showrooms in Bangladesh's Chapainawabganj and Bagerhat districts. Knowledge of energy consumption and energy efficiency technology options is found to be associated with household use of energy conservation practices. Household characteristics also influence household energy use behavior. Younger household cohorts are more likely to adopt energy-efficient technologies and energy conservation practices and place primary importance on energy saving for environmental reasons. Education also influences attitudes toward energy conservation in Bangladesh. Low-education households indicate they primarily save electricity for the environment while high-education households indicate they are motivated by environmental concerns.
This study Examines the Effectiveness of Talent Procurement through the Imple...DharmaBanothu
In the world with high technology and fast
forward mindset recruiters are walking/showing interest
towards E-Recruitment. Present most of the HRs of
many companies are choosing E-Recruitment as the best
choice for recruitment. E-Recruitment is being done
through many online platforms like Linkedin, Naukri,
Instagram , Facebook etc. Now with high technology E-
Recruitment has gone through next level by using
Artificial Intelligence too.
Key Words : Talent Management, Talent Acquisition , E-
Recruitment , Artificial Intelligence Introduction
Effectiveness of Talent Acquisition through E-
Recruitment in this topic we will discuss about 4important
and interlinked topics which are
Blood finder application project report (1).pdfKamal Acharya
Blood Finder is an emergency time app where a user can search for the blood banks as
well as the registered blood donors around Mumbai. This application also provide an
opportunity for the user of this application to become a registered donor for this user have
to enroll for the donor request from the application itself. If the admin wish to make user
a registered donor, with some of the formalities with the organization it can be done.
Specialization of this application is that the user will not have to register on sign-in for
searching the blood banks and blood donors it can be just done by installing the
application to the mobile.
The purpose of making this application is to save the user’s time for searching blood of
needed blood group during the time of the emergency.
This is an android application developed in Java and XML with the connectivity of
SQLite database. This application will provide most of basic functionality required for an
emergency time application. All the details of Blood banks and Blood donors are stored
in the database i.e. SQLite.
This application allowed the user to get all the information regarding blood banks and
blood donors such as Name, Number, Address, Blood Group, rather than searching it on
the different websites and wasting the precious time. This application is effective and
user friendly.
9. Transportation Search and rescue Aerial photography
Law enforcementInspection Agriculture
Today’s Applications of MAVs
10. How to fly a drone
Remote control
Requires line of sight or communication link
Requires skilled pilots
Drone crash during soccer
match, Brasilia, 2013Interior of an earthquake-damaged building in Japan
GPS-based navigation
Doesn’t work indoors
Can be unreliable outdoors
11. Problems of GPS
Does not work indoors
Even outdoors it is not a reliable service
Satellite coverage
Multipath problem
15. This robot is «blind»
How do we Localize without GPS ?
16. This robot is «blind»
How do we Localize without GPS ?
Motion capture system
Markers
17. This robot can «see»This robot is «blind»
How do we Localize without GPS ?
Motion capture system
Markers
18. AutonomousVision-based Navigation in GPS-denied Environments
[Scaramuzza, Achtelik, Weiss, Fraundorfer, et al., Vision-Controlled Micro Flying Robots: from System
Design to Autonomous Navigation and Mapping in GPS-denied Environments, IEEE RAM, 2014]
19. Problems with Vision-controlled MAVs
Quadrotors have the potential to navigate quickly through unstructured
environments, enter and rapidly explore collapsed buildings, but…
Autonomous operation is currently restricted to controlled environments
Autonomous maneuvers with onboard cameras are still slow and inaccurate
compared to those attainable with motion-capture systems
Why?
Perception algorithms are mature but not robust
Unlike lidars and Vicon, localization accuracy depends on depth & texture!
Sparse models instead of dense environment models
Control & perception have been mostly considered separately
Not capable of adapting to changes
20. Outline
Vision-based, GPS-denied Navigation
From Sparse to Dense Models
Active Vision and Control
Low-latency State Estimation for agile motion
24. Feature-based vs. Direct Methods
Feature-based (e.g., PTAM, Klein’08)
1. Feature extraction
2. Feature matching
3. RANSAC + P3P
4. Reprojection error
minimization
𝑇𝑘,𝑘−1 = arg min
𝑇
𝒖′𝑖 − 𝜋 𝒑𝑖
2
𝑖
Direct approaches (e.g., Meilland’13)
1. Minimize photometric error
𝑇𝑘,𝑘−1
𝐼 𝑘
𝒖′𝑖
𝒑𝑖
𝒖𝑖
𝐼 𝑘−1
𝑇𝑘,𝑘−1 = ?
𝒑𝑖
𝒖′𝑖𝒖𝑖
𝑇𝑘,𝑘−1 = arg min
𝑇
𝐼 𝑘 𝒖′𝑖 − 𝐼 𝑘−1(𝒖𝑖) 2
𝑖
[Soatto’95, Meilland and Comport, IROS 2013], DVO [Kerl et al., ICRA
2013], DTAM [Newcombe et al., ICCV ‘11], ...
25. Feature-based vs. Direct Methods
1. Feature extraction
2. Feature matching
3. RANSAC + P3P
4. Reprojection error
minimization
𝑇𝑘,𝑘−1 = arg min
𝑇
𝒖′𝑖 − 𝜋 𝒑𝑖
2
𝑖
Direct approaches
1. Minimize photometric error
𝑇𝑘,𝑘−1 = arg min
𝑇
𝐼 𝑘 𝒖′𝑖 − 𝐼 𝑘−1(𝒖𝑖) 2
𝑖
Large frame-to-frame motions
Slow (20-30 Hz) due to costly feature
extraction and matching
Not robust to high-frequency and
repetive texture
Every pixel in the image can be
exploited (precision, robustness)
Increasing camera frame-rate reduces
computational cost per frame
Limited to small frame-to-frame
motion
Feature-based (e.g., PTAM, Klein’08)
26. Feature-based vs. Direct Methods
1. Feature extraction
2. Feature matching
3. RANSAC + P3P
4. Reprojection error
minimization
𝑇𝑘,𝑘−1 = arg min
𝑇
𝒖′𝑖 − 𝜋 𝒑𝑖
2
𝑖
Direct approaches
1. Minimize photometric error
𝑇𝑘,𝑘−1 = arg min
𝑇
𝐼 𝑘 𝒖′𝑖 − 𝐼 𝑘−1(𝒖𝑖) 2
𝑖
Large frame-to-frame motions
Slow (20-30 Hz) due to costly feature
extraction and matching
Not robust to high-frequency and
repetive texture
Every pixel in the image can be
exploited (precision, robustness)
Increasing camera frame-rate reduces
computational cost per frame
Limited to small frame-to-frame
motion
Feature-based (e.g., PTAM, Klein’08)
Our solution:
SVO: Semi-direct Visual Odometry [ICRA’14]
Combines feature-based and direct methods
27. SVO: Experiments in real-world environments
[Forster, Pizzoli, Scaramuzza, «SVO: Semi Direct Visual Odometry», ICRA’14]
Robust to fast and abrupt motions
Video:
https://www.youtube.com/watch?v=2YnI
Mfw6bJY
28. Processing Times of SVO
Laptop (Intel i7, 2.8 GHz)
Embedded ARM Cortex-A9, 1.7 GHz
400 frames per second
Up to 70 frames per second
Open Source available at: github.com/uzh-rpg/rpg_svo
Works with and without ROS
Closed-Source professional edition available for companies
Source Code
30. Scale Ambiguity
With a single camera, we only know the relative scale
No information about the metric scale
31. Absolute Scale Estimation
The absolute pose 𝑥 is known up to a scale 𝑠, thus
𝑥 = 𝑠𝑥
IMU provides accelerations, thus
𝑣 = 𝑣0 + 𝑎 𝑡 𝑑𝑡
By derivating the first one and equating them
𝑠𝑥 = 𝑣0 + 𝑎 𝑡 𝑑𝑡
As shown in [Martinelli, TRO’12], for 6DOF, both 𝑠 and 𝑣0 can be determined in closed form
from a single feature observation and 3 views
The scale and velocity can then be tracked using
Filter-based approaches
Losely-coupled approaches [Lynen et al., IROS’13]
Tightly-coupled approached (e.g., Google TANGO) [Mourikis & Roumeliotis, TRO’12]
Optimization-based approaches [Leutenegger, RSS’13], [Forster, Scaramuzza, RSS’14]
32. Fusion is solved as a non-linear optimization problem (no Kalman filter):
Increased accuracy over filtering methods
IMU residuals Reprojection residuals
[Forster, Carlone, Dellaert, Scaramuzza, IMU Preintegration on Manifold for efficient Visual-Inertial
Maximum-a-Posteriori Estimation, RSS’15]
Visual-Inertial Fusion [RSS’15]
33. Comparison with Previous Works
Google TangoProposed ASLAM
Forster, Carlone, Dellaert, Scaramuzza, IMU Preintegration on Manifold for efficient Visual-Inertial Maximum-a-
Posteriori Estimation, Robotics Science and Systens’15
Accuracy: 0.1% of the travel distance
Video: https://www.youtube.com/watch?v=CsJkci5lfco
35. Quadrotor System
PX4 (IMU)
Global-Shutter Camera
• 752x480 pixels
• High dynamic range
• 90 fps
450 grams
Odroid U3 Computer
• Quad Core Odroid (ARM Cortex A-9) used in Samsung Galaxy S4 phones
• Runs Linux Ubuntu and ROS
36. Flight Results: Hovering
RMS error: 5 mm, height: 1.5 m – Down-looking camera
Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D
Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.
37. Flight Results: Indoor, aggressive flight
Speed: 4 m/s, height: 1.5 m – Down-looking camera
Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D
Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.
Video: https://www.youtube.com/watch?v=l3TCiCe_T3g
38. Autonomous Vision-based Flight over Mockup Disaster Zone
Firefighters’ training area, Zurich
Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D
Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.
Video:
https://www.youtube.com/watch?v
=3mNY9-DSUDk
39. Probabilistic Depth Estimation
Depth-Filter:
• Depth Filter for every feature
• Recursive Bayesian depth estimation
Mixture of Gaussian + Uniform distribution
[Forster, Pizzoli, Scaramuzza, SVO: Semi Direct Visual Odometry, IEEE ICRA’14]
40. Robustness to Dynamic Objects and Occlusions
• Depth uncertainty is crucial for safety and robustness
• Outliers are caused by wrong data association (e.g., moving objects, distortions)
• Probabilistic depth estimation models outliers
Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D
Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.
Video:
https://www.youtube.com/watch?v
=LssgKdDz5z0
42. Automatic Failure Recovery from Aggressive Flight [ICRA’15]
Faessler, Fontana, Forster, Scaramuzza, Automatic Re-Initialization and Failure Recovery for Aggressive Flight
with a Monocular Vision-Based Quadrotor, ICRA’15. Featured in IEEE Spectrum.
Video:
https://www.youtube.com/watch?v
=pGU1s6Y55JI
47. Automatic Failure Recovery from Aggressive Flight [ICRA’15]
Faessler, Fontana, Forster, Scaramuzza, Automatic Re-Initialization and Failure Recovery for Aggressive Flight
with a Monocular Vision-Based Quadrotor, ICRA’15. Featured in IEEE Spectrum.
48. From Sparse to Dense 3D
Models
[M. Pizzoli, C. Forster, D. Scaramuzza, REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA’14]
49. Goal: estimate depth of every pixel in real time
Pros:
- Advantageous for environment interaction (e.g., collision avoidance,
landing, grasping, industrial inspection, etc)
- Higher position accuracy
Cons: computationally expensive (requires GPU)
Dense Reconstruction in Real-Time
[ICRA’15] [IROS’13, SSRR’14]
50. Dense Reconstruction Pipeline
Local methods
Estimate depth for every pixel independently using photometric cost aggregation
Global methods
Refine the depth surface as a whole by enforcing smoothness constraint
(“Regularization”)
𝐸 𝑑 = 𝐸 𝑑 𝑑 + λ𝐸𝑠(𝑑)
Data term Regularization term:
penalizes non-smooth
surfaces
[Newcombe et al. 2011]
51. [M. Pizzoli, C. Forster, D. Scaramuzza, REMODE: Probabilistic, Monocular Dense Reconstruction in Real Time, ICRA’14]
REMODE: Probabilistic Monocular Dense Reconstruction [ICRA’14]
Running at 50 Hz on GPU on a Lenovo W530, i7
Video:
https://www.youtube.com/watch?v
=QTKd5UWCG0Q
53. Autonomus, Flying 3D Scanning [ JFR’15]
• Sensing, control, state estimation run onboard at 50 Hz (Odroid U3, ARM Cortex A9)
• Dense reconstruction runs live on video streamed to laptop (Lenovo W530, i7)
2x
Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D
Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.
Video:
https://www.youtube.com/watch?v
=7-kPiWaFYAc
54. • Sensing, control, state estimation run onboard at 50 Hz (Odroid U3, ARM Cortex A9)
• Dense reconstruction runs live on video streamed to laptop (Lenovo W530, i7)
Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D
Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.
Autonomus, Flying 3D Scanning [ JFR’15]
55. Applications: Industrial Inspection
Industrial collaboration with Parrot-SenseFly targets:
Real-time dense reconstruction with 5 cameras
Vision-based navigation
Dense 3D mapping in real time
Faessler, Fontana, Forster, Mueggler, Pizzoli, Scaramuzza, Autonomous, Vision-based Flight and Live Dense 3D
Mapping with a Quadrotor Micro Aerial Vehicle, Journal of Field Robotics, 2015.
Video: https://www.youtube.com/watch?v=gr00Bf0AP1k
56. Inspection of CERN tunnels
Problem: inspection of CERN tunnels currently done by technicians,
who expend much of their annual quota of safe radiation dose
Goal: inspection of LHC tunnel with autonomous drone
Challenge: low illumination, cluttered environment
57. Autonomous Landing-Spot Detection and Landing [ICRA’15]
Forster, Faessler, Fontana, Werlberger, Scaramuzza, Continuous On-Board Monocular-Vision-based Elevation Mapping
Applied to Autonomous Landing of Micro Aerial Vehicles, ICRA’15.
Video: https://www.youtube.com/watch?v=phaBKFwfcJ4
58. Having an autonomous landing-spot detection
can really help!
The Philae lander while approaching the comet on November 12, 2014