The document summarizes lecture 4 on reconstruction from two views. It discusses techniques for shape reconstruction including shape from X methods using multiple camera views or additional information. It then covers the triangulation principle for reconstructing 3D points from 2D point correspondences in multiple views. Finally, it introduces epipolar geometry which models the geometric relationship between two views and can be used to reconstruct the fundamental matrix and epipolar lines.
This document describes visual servo control of a mobile robot using homography-based image feedback. A camera is mounted on the robot to track features in the environment. Homography is estimated from matched features to relate the current and target camera views. The robot's motion model and a control law are derived to drive the robot from its initial position to the target position defined by the target image. Experimental results are presented to validate the visual servo control approach. The high-level goal is to navigate the robot to the target position using only image feedback from the mounted camera.
Build Your Own 3D Scanner: Surface ReconstructionDouglas Lanman
Build Your Own 3D Scanner:
Surface Reconstruction
http://mesh.brown.edu/byo3d/
SIGGRAPH 2009 Courses
Douglas Lanman and Gabriel Taubin
This course provides a beginner with the necessary mathematics, software, and practical details to leverage projector-camera systems in their own 3D scanning projects. An example-driven approach is used throughout; each new concept is illustrated using a practical scanner implemented with off-the-shelf parts. The course concludes by detailing how these new approaches are used in rapid prototyping, entertainment, cultural heritage, and web-based applications.
2015-06-17 FEKO Based ISAR Analysis for 3D Object ReconstructionDr. Ali Nassib
This document discusses using inverse synthetic aperture radar (ISAR) imaging techniques to reconstruct 3D images of objects from electromagnetic scattering data. It presents the mathematical models and simulation setup used. Simulations were conducted of two thin cylinders separated by either half a wavelength or a full wavelength. When separated by half a wavelength, the cylinders were not clearly resolved due to coupling effects. But when separated by a wavelength, the two cylinders were successfully reconstructed from the simulated scattering data using inverse scattering algorithms. Future work involves reconstructing the full dyadic contrast function and performing additional experiments.
The document contains 35 multiple choice physics questions related to topics including:
- Characteristics of virtual images formed by mirrors
- Focal lengths of spherical mirrors and lenses
- Conditions for real and virtual images formed by lenses and mirrors
- Diffraction, interference patterns from double slits, and properties of diffraction gratings
- Kinematics and relativity, including time dilation, length contraction, and momentum-energy relationships for objects moving at relativistic speeds
Here are the key steps in Harris corner detection:
1. Compute the autocorrelation matrix M for a window around each pixel using:
M = ∑w(x,y)[IxIx IxIy]
[IxIy IyIy]
Where Ix and Iy are the gradients of the image I in x and y directions.
2. Compute the corner response function R = det(M) - k(trace(M))^2
3. A large R value indicates a corner. Threshold R to find candidate corner points.
4. Refine the candidate locations using interpolation.
So in summary, Harris detection looks for pixels with
Templateless Marked Element Recognition Using Computer Visionshivam chaurasia
The document describes an algorithm for templateless marked element recognition in documents using computer vision. It discusses preprocessing steps like converting images to grayscale, blurring, and edge detection. It then describes detecting shapes like checkboxes and radio buttons using contour analysis and evaluating pixel thresholds to determine if elements are selected. Pseudocode provides details of the complete algorithm to detect and mark checked checkboxes and radio buttons on input images without predefined templates.
Vehicle tracking and distance estimation based on multiple image featuresYixin Chen
This document presents a vehicle tracking algorithm that uses multiple image features to detect and track vehicles in images captured from a moving vehicle. The algorithm aims to identify vehicles, track their movement, and estimate the distance between the tracked vehicle and the host vehicle. It uses features like corners, edges, gradients, vehicle symmetry, and image matching. Corners and edges are extracted from the bottom portion of vehicles which is less occluded. Vehicle width is estimated and then height and distance are estimated using width and optical perspective principles. The performance of the algorithm is evaluated on real-world video images.
This document describes visual servo control of a mobile robot using homography-based image feedback. A camera is mounted on the robot to track features in the environment. Homography is estimated from matched features to relate the current and target camera views. The robot's motion model and a control law are derived to drive the robot from its initial position to the target position defined by the target image. Experimental results are presented to validate the visual servo control approach. The high-level goal is to navigate the robot to the target position using only image feedback from the mounted camera.
Build Your Own 3D Scanner: Surface ReconstructionDouglas Lanman
Build Your Own 3D Scanner:
Surface Reconstruction
http://mesh.brown.edu/byo3d/
SIGGRAPH 2009 Courses
Douglas Lanman and Gabriel Taubin
This course provides a beginner with the necessary mathematics, software, and practical details to leverage projector-camera systems in their own 3D scanning projects. An example-driven approach is used throughout; each new concept is illustrated using a practical scanner implemented with off-the-shelf parts. The course concludes by detailing how these new approaches are used in rapid prototyping, entertainment, cultural heritage, and web-based applications.
2015-06-17 FEKO Based ISAR Analysis for 3D Object ReconstructionDr. Ali Nassib
This document discusses using inverse synthetic aperture radar (ISAR) imaging techniques to reconstruct 3D images of objects from electromagnetic scattering data. It presents the mathematical models and simulation setup used. Simulations were conducted of two thin cylinders separated by either half a wavelength or a full wavelength. When separated by half a wavelength, the cylinders were not clearly resolved due to coupling effects. But when separated by a wavelength, the two cylinders were successfully reconstructed from the simulated scattering data using inverse scattering algorithms. Future work involves reconstructing the full dyadic contrast function and performing additional experiments.
The document contains 35 multiple choice physics questions related to topics including:
- Characteristics of virtual images formed by mirrors
- Focal lengths of spherical mirrors and lenses
- Conditions for real and virtual images formed by lenses and mirrors
- Diffraction, interference patterns from double slits, and properties of diffraction gratings
- Kinematics and relativity, including time dilation, length contraction, and momentum-energy relationships for objects moving at relativistic speeds
Here are the key steps in Harris corner detection:
1. Compute the autocorrelation matrix M for a window around each pixel using:
M = ∑w(x,y)[IxIx IxIy]
[IxIy IyIy]
Where Ix and Iy are the gradients of the image I in x and y directions.
2. Compute the corner response function R = det(M) - k(trace(M))^2
3. A large R value indicates a corner. Threshold R to find candidate corner points.
4. Refine the candidate locations using interpolation.
So in summary, Harris detection looks for pixels with
Templateless Marked Element Recognition Using Computer Visionshivam chaurasia
The document describes an algorithm for templateless marked element recognition in documents using computer vision. It discusses preprocessing steps like converting images to grayscale, blurring, and edge detection. It then describes detecting shapes like checkboxes and radio buttons using contour analysis and evaluating pixel thresholds to determine if elements are selected. Pseudocode provides details of the complete algorithm to detect and mark checked checkboxes and radio buttons on input images without predefined templates.
Vehicle tracking and distance estimation based on multiple image featuresYixin Chen
This document presents a vehicle tracking algorithm that uses multiple image features to detect and track vehicles in images captured from a moving vehicle. The algorithm aims to identify vehicles, track their movement, and estimate the distance between the tracked vehicle and the host vehicle. It uses features like corners, edges, gradients, vehicle symmetry, and image matching. Corners and edges are extracted from the bottom portion of vehicles which is less occluded. Vehicle width is estimated and then height and distance are estimated using width and optical perspective principles. The performance of the algorithm is evaluated on real-world video images.
The document summarizes different techniques for pattern projection used in 3D shape acquisition, including passive stereo, active stereo using coded structured light, and classifications of pattern projection methods. It discusses techniques such as time multiplexing using binary patterns, spatial codification using De Bruijn sequences, and direct codification using color. Diagrams illustrate concepts like correspondence problems in passive stereo and how encoded patterns can reduce these issues in active stereo.
Linköping University has several student kitchens all over its campuses where students are given a possibility to warm their food. Critics claim that there are too few student kitchens and that the existing ones are usually overcrowded. That all kitchens are overcrowded at the same time has not been confirmed by sample inspections. One standing hypothesis is that students do not know where all the kitchens are, nor do they want to risk going to a kitchen in another building in case it is full as well.
The aim of this project is to develop a system that will provide the students with information regarding student kitchen usage. The system uses an computer vision approach, estimating the number of people currently using the kitchens. The system was developed using C++, the OpenCV library and the Qt5 library.
https://github.com/GroupDenseKitchen/KitchenOccupation
This document discusses object detection and tracking algorithms. It covers the following key points:
- Background modeling uses a mixture of Gaussian model to identify foreground objects from backgrounds. Tracking includes correlating current objects to previous ones and handling occlusion.
- Challenges include noisy images with false positives, shadows, and removing noise while detecting moving regions. Distance filtering and morphological operations help address these.
- Objects are identified using contours and bounding boxes. Kalman filtering can be used for prediction but has difficulties with smoothing.
- Performance is evaluated using multiple object tracking accuracy (MOTA) and multiple object tracking precision (MOTP) metrics. Areas for improvement include handling size changes and removing lost objects.
This document discusses several techniques for image-based rendering including light field rendering, the lumigraph, view-dependent texture mapping, the unstructured lumigraph, blending fields, and unstructured light fields. It provides intuitive explanations of each technique and how they represent scenes and allow novel views to be rendered from different positions and angles.
This document discusses cryo-electron microscopy (cryo-EM) 3D reconstruction techniques. It describes the cryo-EM imaging process and challenges in reconstructing 3D structures from 2D projection images, including large noise and data size. The document proposes a memory-saving algorithm using tight wavelet frames for cryo-EM 3D reconstruction that formulates the reconstruction as a sparse representation problem solved with soft-thresholding and gradient descent. Simulation results on an E. coli ribosome and experimental results on an adenovirus demonstrate the proposed algorithm can reconstruct 3D structures from noisy projection data.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/qualcomm/embedded-vision-training/videos/pages/may-2016-embedded-vision-summit-mangen
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Michael Mangen, Product Manager for Camera and Computer Vision at Qualcomm, presents the "High-resolution 3D Reconstruction on a Mobile Processor" tutorial at the May 2016 Embedded Vision Summit.
Computer vision has come a long way. Use cases that were previously not possible in mass-market devices are now more accessible thanks to advances in depth sensors and mobile processors. In this presentation, Mangen provides an overview of how we are able to implement high-resolution 3D reconstruction – a capability typically requiring cloud/server processing – on a mobile processor. This is an exciting example of how new sensor technology and advanced mobile processors are bringing computer vision capabilities to broader markets.
A complete illustrated ppt on 3D printing technology. All the additive processes,Future and effects are well described with relevant diagram and images.Must download for attractive seminar presentation.3D Printing technology could revolutionize and re-shape the world. Advances in 3D printing technology can significantly change and improve the way we manufacture products and produce goods worldwide. If the last industrial revolution brought us mass production and the advent of economies of scale - the digital 3D printing revolution could bring mass manufacturing back a full circle - to an era of mass personalization, and a return to individual craftsmanship.
Structure from motion is a computer vision technique used to recover the three-dimensional structure of a scene and the camera motion from a set of images. It involves detecting feature points in multiple images, matching corresponding points across images, estimating camera poses and orientations, and reconstructing the 3D geometry of scene points. Large-scale structure from motion can reconstruct scenes from thousands of images but requires solving very large optimization problems. Applications include 3D modeling, surveying, robot navigation, virtual reality, augmented reality, and simultaneous localization and mapping.
Structure from motion is a computer vision technique used to recover the three-dimensional structure of a scene and the camera motion from a set of images. It can be used to build 3D models of scenes without any prior knowledge of the camera parameters or 3D locations of the scene points. Structure from motion involves detecting feature points in multiple images, matching the features between images, estimating the fundamental matrices between image pairs, and then optimizing a bundle adjustment problem to simultaneously compute the 3D structure and camera motion parameters. Some applications of structure from motion include 3D modeling, surveying, robot navigation, virtual and augmented reality, and visual effects.
New geometric interpretation and analytic solution for quadrilateral reconstr...Joo-Haeng Lee
Accepted as poster presentation for ICPR 2014, Stockholm, Sweden on August 24~28, 2014.
[Revised Version]
Title: New geometric interpretation and analytic solution for quadrilateral reconstruction
Author: Joo-Haeng Lee
Affiliation: Human-Robot Interaction Research Team, ETRI, KOREA
Abstract:
A new geometric framework, called generalized coupled line camera (GCLC), is proposed to derive an analytic solution to reconstruct an unknown scene quadrilateral and the relevant projective structure from a single or multiple image quadrilaterals. We extend the previous approach developed for rectangle to handle arbitrary scene quadrilaterals. First, we generalize a single line camera by removing the centering constraint that the principal axis should bisect a scene line. Then, we couple a pair of generalized line cameras to model a frustum with a quadrilateral base. Finally, we show that the scene quadrilateral and the center of projection can be analytically reconstructed from a single view when prior knowledge on the quadrilateral is available. A completely unknown quadrilateral can be reconstructed from four views through non-linear optimization. We also describe a improved method to handle an off-centered case by geometrically inferring a centered proxy quadrilateral, which accelerates a reconstruction process without relying on homography. The proposed method is easy to implement since each step is expressed as a simple analytic equation. We present the experimental results on real and synthetic examples.
[Submitted Version]
Title: Generalized Coupled Line Cameras and Application in Quadrilateral Reconstruction
Abstract:
Coupled line camera (CLC) provides a geometric framework to derive an analytic solution to reconstruct an unknown scene rectangle and the relevant projective structure from a single image quadrilateral. We extend this approach as generalized coupled line camera (GCLC) to handle a scene quadrilateral. First, we generalize a single line camera by removing the centering constraint that the principal axis should bisect a scene line. Then, we couple a pair of generalized line cameras to model a frustum with a quadrilateral base. Finally, we show that the scene quadrilateral and the center of projection can be analytically reconstructed from a single view when prior knowledge on the quadrilateral is available. ...
This project finds real objects dimensions from their images using OpenCV Java using two methods - Reference Method and Stereo Method.
Code for this project is available on my GitHub:
https://github.com/shyamabhuvanendran/VirtualRuler_CV
The document discusses camera models and imaging in computer vision. It describes the pinhole camera model as the simplest model and the perspective projection model as most commonly used. The key parameters discussed are intrinsic parameters like focal length and principal point, and extrinsic parameters like camera rotation and translation. Lens distortion is also covered. Homography is introduced as a transformation between projective planes that maps lines to lines.
SIGGRAPH 2014 Course on Computational Cameras and Displays (part 4)Matthew O'Toole
Recent advances in both computational photography and displays have given rise to a new generation of computational devices. Computational cameras and displays provide a visual experience that goes beyond the capabilities of traditional systems by adding computational power to optics, lights, and sensors. These devices are breaking new ground in the consumer market, including lightfield cameras that redefine our understanding of pictures (Lytro), displays for visualizing 3D/4D content without special eyewear (Nintendo 3DS), motion-sensing devices that use light coded in space or time to detect motion and position (Kinect, Leap Motion), and a movement toward ubiquitous computing with wearable cameras and displays (Google Glass).
This short (1.5 hour) course serves as an introduction to the key ideas and an overview of the latest work in computational cameras, displays, and light transport.
This document proposes using dual back-to-back Kinect sensors mounted on a robot to capture a 3D model of a large indoor scene. Traditionally, one Kinect is slid across an area, but this requires prominent features and careful handling. The dual Kinect setup requires calibrating the relative pose between the sensors. Since they do not share a view, traditional calibration is not possible. The authors place a dual-face checkerboard on top with a mirror to enable each Kinect to view the same calibration object. This allows estimating the pose between the sensors using a mirror-based algorithm. After capturing local models, the two Kinect views can be merged into a combined 3D model with a larger field of view.
Focus set based reconstruction of micro-objectsJan Wedekind
The document discusses focus set based reconstruction of micro-objects for micro-robotics applications. It describes acquiring a focus set of images at different depths, computing a local sharpness measure for each pixel, and determining a depth map by finding the depth with maximum sharpness for each pixel. It also addresses issues like systematic errors and proposes a multiscale approach to reduce errors while maintaining real-time performance.
Fisheye Omnidirectional View in Autonomous DrivingYu Huang
This document discusses several papers related to using omnidirectional/fisheye camera views for autonomous driving applications. The papers propose methods for tasks like image classification, object detection, scene understanding from 360 degree camera data. Specific approaches discussed include graph-based classification of omnidirectional images, learning spherical convolutions for 360 degree imagery, spherical CNNs, and networks for scene understanding and 3D object detection using around view monitoring camera systems.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles demonstrate the new model outperforms conventional models in correction accuracy and recovering locations of high objects.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles show the self-adjustive model outperforms conventional models in correction accuracy and recovering locations of high objects.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles demonstrate the new model outperforms conventional models in correction accuracy and recovering locations of high objects.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles demonstrate the new model outperforms conventional models in correction accuracy and recovering locations of high objects.
The document summarizes different techniques for pattern projection used in 3D shape acquisition, including passive stereo, active stereo using coded structured light, and classifications of pattern projection methods. It discusses techniques such as time multiplexing using binary patterns, spatial codification using De Bruijn sequences, and direct codification using color. Diagrams illustrate concepts like correspondence problems in passive stereo and how encoded patterns can reduce these issues in active stereo.
Linköping University has several student kitchens all over its campuses where students are given a possibility to warm their food. Critics claim that there are too few student kitchens and that the existing ones are usually overcrowded. That all kitchens are overcrowded at the same time has not been confirmed by sample inspections. One standing hypothesis is that students do not know where all the kitchens are, nor do they want to risk going to a kitchen in another building in case it is full as well.
The aim of this project is to develop a system that will provide the students with information regarding student kitchen usage. The system uses an computer vision approach, estimating the number of people currently using the kitchens. The system was developed using C++, the OpenCV library and the Qt5 library.
https://github.com/GroupDenseKitchen/KitchenOccupation
This document discusses object detection and tracking algorithms. It covers the following key points:
- Background modeling uses a mixture of Gaussian model to identify foreground objects from backgrounds. Tracking includes correlating current objects to previous ones and handling occlusion.
- Challenges include noisy images with false positives, shadows, and removing noise while detecting moving regions. Distance filtering and morphological operations help address these.
- Objects are identified using contours and bounding boxes. Kalman filtering can be used for prediction but has difficulties with smoothing.
- Performance is evaluated using multiple object tracking accuracy (MOTA) and multiple object tracking precision (MOTP) metrics. Areas for improvement include handling size changes and removing lost objects.
This document discusses several techniques for image-based rendering including light field rendering, the lumigraph, view-dependent texture mapping, the unstructured lumigraph, blending fields, and unstructured light fields. It provides intuitive explanations of each technique and how they represent scenes and allow novel views to be rendered from different positions and angles.
This document discusses cryo-electron microscopy (cryo-EM) 3D reconstruction techniques. It describes the cryo-EM imaging process and challenges in reconstructing 3D structures from 2D projection images, including large noise and data size. The document proposes a memory-saving algorithm using tight wavelet frames for cryo-EM 3D reconstruction that formulates the reconstruction as a sparse representation problem solved with soft-thresholding and gradient descent. Simulation results on an E. coli ribosome and experimental results on an adenovirus demonstrate the proposed algorithm can reconstruct 3D structures from noisy projection data.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/qualcomm/embedded-vision-training/videos/pages/may-2016-embedded-vision-summit-mangen
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Michael Mangen, Product Manager for Camera and Computer Vision at Qualcomm, presents the "High-resolution 3D Reconstruction on a Mobile Processor" tutorial at the May 2016 Embedded Vision Summit.
Computer vision has come a long way. Use cases that were previously not possible in mass-market devices are now more accessible thanks to advances in depth sensors and mobile processors. In this presentation, Mangen provides an overview of how we are able to implement high-resolution 3D reconstruction – a capability typically requiring cloud/server processing – on a mobile processor. This is an exciting example of how new sensor technology and advanced mobile processors are bringing computer vision capabilities to broader markets.
A complete illustrated ppt on 3D printing technology. All the additive processes,Future and effects are well described with relevant diagram and images.Must download for attractive seminar presentation.3D Printing technology could revolutionize and re-shape the world. Advances in 3D printing technology can significantly change and improve the way we manufacture products and produce goods worldwide. If the last industrial revolution brought us mass production and the advent of economies of scale - the digital 3D printing revolution could bring mass manufacturing back a full circle - to an era of mass personalization, and a return to individual craftsmanship.
Structure from motion is a computer vision technique used to recover the three-dimensional structure of a scene and the camera motion from a set of images. It involves detecting feature points in multiple images, matching corresponding points across images, estimating camera poses and orientations, and reconstructing the 3D geometry of scene points. Large-scale structure from motion can reconstruct scenes from thousands of images but requires solving very large optimization problems. Applications include 3D modeling, surveying, robot navigation, virtual reality, augmented reality, and simultaneous localization and mapping.
Structure from motion is a computer vision technique used to recover the three-dimensional structure of a scene and the camera motion from a set of images. It can be used to build 3D models of scenes without any prior knowledge of the camera parameters or 3D locations of the scene points. Structure from motion involves detecting feature points in multiple images, matching the features between images, estimating the fundamental matrices between image pairs, and then optimizing a bundle adjustment problem to simultaneously compute the 3D structure and camera motion parameters. Some applications of structure from motion include 3D modeling, surveying, robot navigation, virtual and augmented reality, and visual effects.
New geometric interpretation and analytic solution for quadrilateral reconstr...Joo-Haeng Lee
Accepted as poster presentation for ICPR 2014, Stockholm, Sweden on August 24~28, 2014.
[Revised Version]
Title: New geometric interpretation and analytic solution for quadrilateral reconstruction
Author: Joo-Haeng Lee
Affiliation: Human-Robot Interaction Research Team, ETRI, KOREA
Abstract:
A new geometric framework, called generalized coupled line camera (GCLC), is proposed to derive an analytic solution to reconstruct an unknown scene quadrilateral and the relevant projective structure from a single or multiple image quadrilaterals. We extend the previous approach developed for rectangle to handle arbitrary scene quadrilaterals. First, we generalize a single line camera by removing the centering constraint that the principal axis should bisect a scene line. Then, we couple a pair of generalized line cameras to model a frustum with a quadrilateral base. Finally, we show that the scene quadrilateral and the center of projection can be analytically reconstructed from a single view when prior knowledge on the quadrilateral is available. A completely unknown quadrilateral can be reconstructed from four views through non-linear optimization. We also describe a improved method to handle an off-centered case by geometrically inferring a centered proxy quadrilateral, which accelerates a reconstruction process without relying on homography. The proposed method is easy to implement since each step is expressed as a simple analytic equation. We present the experimental results on real and synthetic examples.
[Submitted Version]
Title: Generalized Coupled Line Cameras and Application in Quadrilateral Reconstruction
Abstract:
Coupled line camera (CLC) provides a geometric framework to derive an analytic solution to reconstruct an unknown scene rectangle and the relevant projective structure from a single image quadrilateral. We extend this approach as generalized coupled line camera (GCLC) to handle a scene quadrilateral. First, we generalize a single line camera by removing the centering constraint that the principal axis should bisect a scene line. Then, we couple a pair of generalized line cameras to model a frustum with a quadrilateral base. Finally, we show that the scene quadrilateral and the center of projection can be analytically reconstructed from a single view when prior knowledge on the quadrilateral is available. ...
This project finds real objects dimensions from their images using OpenCV Java using two methods - Reference Method and Stereo Method.
Code for this project is available on my GitHub:
https://github.com/shyamabhuvanendran/VirtualRuler_CV
The document discusses camera models and imaging in computer vision. It describes the pinhole camera model as the simplest model and the perspective projection model as most commonly used. The key parameters discussed are intrinsic parameters like focal length and principal point, and extrinsic parameters like camera rotation and translation. Lens distortion is also covered. Homography is introduced as a transformation between projective planes that maps lines to lines.
SIGGRAPH 2014 Course on Computational Cameras and Displays (part 4)Matthew O'Toole
Recent advances in both computational photography and displays have given rise to a new generation of computational devices. Computational cameras and displays provide a visual experience that goes beyond the capabilities of traditional systems by adding computational power to optics, lights, and sensors. These devices are breaking new ground in the consumer market, including lightfield cameras that redefine our understanding of pictures (Lytro), displays for visualizing 3D/4D content without special eyewear (Nintendo 3DS), motion-sensing devices that use light coded in space or time to detect motion and position (Kinect, Leap Motion), and a movement toward ubiquitous computing with wearable cameras and displays (Google Glass).
This short (1.5 hour) course serves as an introduction to the key ideas and an overview of the latest work in computational cameras, displays, and light transport.
This document proposes using dual back-to-back Kinect sensors mounted on a robot to capture a 3D model of a large indoor scene. Traditionally, one Kinect is slid across an area, but this requires prominent features and careful handling. The dual Kinect setup requires calibrating the relative pose between the sensors. Since they do not share a view, traditional calibration is not possible. The authors place a dual-face checkerboard on top with a mirror to enable each Kinect to view the same calibration object. This allows estimating the pose between the sensors using a mirror-based algorithm. After capturing local models, the two Kinect views can be merged into a combined 3D model with a larger field of view.
Focus set based reconstruction of micro-objectsJan Wedekind
The document discusses focus set based reconstruction of micro-objects for micro-robotics applications. It describes acquiring a focus set of images at different depths, computing a local sharpness measure for each pixel, and determining a depth map by finding the depth with maximum sharpness for each pixel. It also addresses issues like systematic errors and proposes a multiscale approach to reduce errors while maintaining real-time performance.
Fisheye Omnidirectional View in Autonomous DrivingYu Huang
This document discusses several papers related to using omnidirectional/fisheye camera views for autonomous driving applications. The papers propose methods for tasks like image classification, object detection, scene understanding from 360 degree camera data. Specific approaches discussed include graph-based classification of omnidirectional images, learning spherical convolutions for 360 degree imagery, spherical CNNs, and networks for scene understanding and 3D object detection using around view monitoring camera systems.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles demonstrate the new model outperforms conventional models in correction accuracy and recovering locations of high objects.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles show the self-adjustive model outperforms conventional models in correction accuracy and recovering locations of high objects.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles demonstrate the new model outperforms conventional models in correction accuracy and recovering locations of high objects.
A SELF-ADJUSTIVE GEOMETRIC CORRECTION METHOD FOR SERIOUSLY OBLIQUE AERO IMAGE...grssieee
This document presents a self-adjustive geometric correction method for seriously oblique aerial images. It analyzes projection errors caused by the curvature of the Earth and terrain relief. A ternary quadratic polynomial model is used, with adjustments to better correct for relief-induced projection errors. Experiments on images taken at large viewing angles demonstrate the new model outperforms conventional models in correction accuracy and recovering locations of high objects.
1) Pixels (u,v) in an image are projections of 3D points (X,Y,Z) in the world onto the image plane according to the camera's intrinsic and extrinsic parameters.
2) The process of projecting 3D points into 2D pixels is called forward projection and can be represented by a perspective projection matrix that models the camera's intrinsic properties and its position and orientation relative to the world frame.
3) Intrinsic parameters include focal length, principal point, and pixel scale factors that define the camera's imaging geometry, while extrinsic parameters define the rigid transformation between the world and camera coordinate frames.
COSC 426 Lecture 5 on Mathematical Principles Behind AR Registration. Given by Adrian Clark from the HIT Lab NZ at the University of Canterbury, August 8, 2012
The flow of baseline estimation using a single omnidirectional cameraTELKOMNIKA JOURNAL
1. The document describes a method for estimating the baseline of a single omnidirectional camera using optical flow tracking of points on an object.
2. As the camera is moved horizontally, tracking points on an object in panoramic images produces coordinate shifts that are saved and represented as graphs.
3. Analyzing the graphs allows determining the equation that estimates the baseline flow and coefficients of the equation.
Solving the Pose Ambiguity via a Simple Concentric Circle ConstraintDr. Amarjeet Singh
Estimating the pose of objects with circle feature from images is a basic and important question in computer vision
community. This paper is focused on the ambiguity problem in pose estimation of circle feature, and a new method is proposed based
on the concentric circle constraint. The pose of a single circle feature, in general, can be determined from its projection in the image
plane with a pre-calibrated camera. However, there are generally two possible sets of pose parameters. By introducing the concentric
circle constraint, interference from the false solution can be excluded. On the basis of element at infinity in projective geometry and
the Euclidean distance invariant, cases that concentric circles are coplanar and non-coplanar are discussed respectively. Experiments
on these two cases are performed to validate the proposed method.
The document describes a method for creating panoramic images from video frames. Key steps include camera calibration to determine intrinsic parameters, feature detection and matching between frames using SIFT or Shi-Tomasi features, selecting key frames when sufficient camera movement is detected, and stitching the key frames onto a cylindrical projection to create the panorama. Experimental results show Shi-Tomasi with optical flow is faster than SIFT with FLANN for feature matching.
Similar to Lecture 4 Reconstruction from Two Views (19)
This document summarizes camera calibration methods. It begins by introducing the pinhole camera model and describing its four step process: from world to camera coordinates, projection to the image plane, modeling lens distortion, and conversion to image coordinates. It then overviews several calibration methods, including the method of Hall which uses a linear transformation matrix, and the method of Faugeras-Toscani which obtains camera parameters through an iterative process accounting for radial distortion. The document focuses on explaining the method of Hall in detail, showing how its modeling leads to equations that can be solved using a pseudoinverse to obtain the camera calibration parameters.
This document outlines the contents of Lecture 1 on rigid body transformations. The key topics covered include Cartesian coordinates, points and vectors, inner products and cross products, translations using translation matrices, rotations using rotation matrices, and homogeneous coordinates for representing transformations. The lecture will define and provide examples of how to represent and compute rigid body transformations such as translations and rotations between coordinate systems.
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
LAND USE LAND COVER AND NDVI OF MIRZAPUR DISTRICT, UPRAHUL
This Dissertation explores the particular circumstances of Mirzapur, a region located in the
core of India. Mirzapur, with its varied terrains and abundant biodiversity, offers an optimal
environment for investigating the changes in vegetation cover dynamics. Our study utilizes
advanced technologies such as GIS (Geographic Information Systems) and Remote sensing to
analyze the transformations that have taken place over the course of a decade.
The complex relationship between human activities and the environment has been the focus
of extensive research and worry. As the global community grapples with swift urbanization,
population expansion, and economic progress, the effects on natural ecosystems are becoming
more evident. A crucial element of this impact is the alteration of vegetation cover, which plays a
significant role in maintaining the ecological equilibrium of our planet.Land serves as the foundation for all human activities and provides the necessary materials for
these activities. As the most crucial natural resource, its utilization by humans results in different
'Land uses,' which are determined by both human activities and the physical characteristics of the
land.
The utilization of land is impacted by human needs and environmental factors. In countries
like India, rapid population growth and the emphasis on extensive resource exploitation can lead
to significant land degradation, adversely affecting the region's land cover.
Therefore, human intervention has significantly influenced land use patterns over many
centuries, evolving its structure over time and space. In the present era, these changes have
accelerated due to factors such as agriculture and urbanization. Information regarding land use and
cover is essential for various planning and management tasks related to the Earth's surface,
providing crucial environmental data for scientific, resource management, policy purposes, and
diverse human activities.
Accurate understanding of land use and cover is imperative for the development planning
of any area. Consequently, a wide range of professionals, including earth system scientists, land
and water managers, and urban planners, are interested in obtaining data on land use and cover
changes, conversion trends, and other related patterns. The spatial dimensions of land use and
cover support policymakers and scientists in making well-informed decisions, as alterations in
these patterns indicate shifts in economic and social conditions. Monitoring such changes with the
help of Advanced technologies like Remote Sensing and Geographic Information Systems is
crucial for coordinated efforts across different administrative levels. Advanced technologies like
Remote Sensing and Geographic Information Systems
9
Changes in vegetation cover refer to variations in the distribution, composition, and overall
structure of plant communities across different temporal and spatial scales. These changes can
occur natural.
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
Chapter wise All Notes of First year Basic Civil Engineering.pptxDenish Jangid
Chapter wise All Notes of First year Basic Civil Engineering
Syllabus
Chapter-1
Introduction to objective, scope and outcome the subject
Chapter 2
Introduction: Scope and Specialization of Civil Engineering, Role of civil Engineer in Society, Impact of infrastructural development on economy of country.
Chapter 3
Surveying: Object Principles & Types of Surveying; Site Plans, Plans & Maps; Scales & Unit of different Measurements.
Linear Measurements: Instruments used. Linear Measurement by Tape, Ranging out Survey Lines and overcoming Obstructions; Measurements on sloping ground; Tape corrections, conventional symbols. Angular Measurements: Instruments used; Introduction to Compass Surveying, Bearings and Longitude & Latitude of a Line, Introduction to total station.
Levelling: Instrument used Object of levelling, Methods of levelling in brief, and Contour maps.
Chapter 4
Buildings: Selection of site for Buildings, Layout of Building Plan, Types of buildings, Plinth area, carpet area, floor space index, Introduction to building byelaws, concept of sun light & ventilation. Components of Buildings & their functions, Basic concept of R.C.C., Introduction to types of foundation
Chapter 5
Transportation: Introduction to Transportation Engineering; Traffic and Road Safety: Types and Characteristics of Various Modes of Transportation; Various Road Traffic Signs, Causes of Accidents and Road Safety Measures.
Chapter 6
Environmental Engineering: Environmental Pollution, Environmental Acts and Regulations, Functional Concepts of Ecology, Basics of Species, Biodiversity, Ecosystem, Hydrological Cycle; Chemical Cycles: Carbon, Nitrogen & Phosphorus; Energy Flow in Ecosystems.
Water Pollution: Water Quality standards, Introduction to Treatment & Disposal of Waste Water. Reuse and Saving of Water, Rain Water Harvesting. Solid Waste Management: Classification of Solid Waste, Collection, Transportation and Disposal of Solid. Recycling of Solid Waste: Energy Recovery, Sanitary Landfill, On-Site Sanitation. Air & Noise Pollution: Primary and Secondary air pollutants, Harmful effects of Air Pollution, Control of Air Pollution. . Noise Pollution Harmful Effects of noise pollution, control of noise pollution, Global warming & Climate Change, Ozone depletion, Greenhouse effect
Text Books:
1. Palancharmy, Basic Civil Engineering, McGraw Hill publishers.
2. Satheesh Gopi, Basic Civil Engineering, Pearson Publishers.
3. Ketki Rangwala Dalal, Essentials of Civil Engineering, Charotar Publishing House.
4. BCP, Surveying volume 1
How to Setup Warehouse & Location in Odoo 17 InventoryCeline George
In this slide, we'll explore how to set up warehouses and locations in Odoo 17 Inventory. This will help us manage our stock effectively, track inventory levels, and streamline warehouse operations.
How to Manage Your Lost Opportunities in Odoo 17 CRMCeline George
Odoo 17 CRM allows us to track why we lose sales opportunities with "Lost Reasons." This helps analyze our sales process and identify areas for improvement. Here's how to configure lost reasons in Odoo 17 CRM
Strategies for Effective Upskilling is a presentation by Chinwendu Peace in a Your Skill Boost Masterclass organisation by the Excellence Foundation for South Sudan on 08th and 09th June 2024 from 1 PM to 3 PM on each day.
This document provides an overview of wound healing, its functions, stages, mechanisms, factors affecting it, and complications.
A wound is a break in the integrity of the skin or tissues, which may be associated with disruption of the structure and function.
Healing is the body’s response to injury in an attempt to restore normal structure and functions.
Healing can occur in two ways: Regeneration and Repair
There are 4 phases of wound healing: hemostasis, inflammation, proliferation, and remodeling. This document also describes the mechanism of wound healing. Factors that affect healing include infection, uncontrolled diabetes, poor nutrition, age, anemia, the presence of foreign bodies, etc.
Complications of wound healing like infection, hyperpigmentation of scar, contractures, and keloid formation.
How to Make a Field Mandatory in Odoo 17Celine George
In Odoo, making a field required can be done through both Python code and XML views. When you set the required attribute to True in Python code, it makes the field required across all views where it's used. Conversely, when you set the required attribute in XML views, it makes the field required only in the context of that particular view.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
2. 2
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
3. 3
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
4. 4
Lecture 4: Reconstruction from two views
4.1 Shape from X
Techniques based on:
– Modifying the intrinsic camera parameters
i.e. Depth from Focus/Defocus and Depth from Zooming
– Considering an additional source of light onto the scene
i.e. Shape from Structured Light and Shape from Photometric
Stereo
– Considering additional surface information
i.e. Shape from Shading, Shape from Texture and Shape from
Geometric Constraints
– Multiple views
i.e. Shape from Stereo and Shape from Motion
Shape from Focus/Defocus
5. 5
Lecture 4: Reconstruction from two views
4.1 Shape from X
Techniques based on:
– Modifying the intrinsic camera parameters
i.e. Depth from Focus/Defocus and Depth from Zooming
– Considering an additional source of light onto the scene
i.e. Shape from Structured Light and Shape from Photometric
Stereo
– Considering additional surface information
i.e. Shape from Shading, Shape from Texture and Shape from
Geometric Constraints
– Multiple views
i.e. Shape from Stereo and Shape from Motion
Shape from Structured Light
6. 6
Lecture 4: Reconstruction from two views
4.1 Shape from X
Techniques based on:
– Modifying the intrinsic camera parameters
i.e. Depth from Focus/Defocus and Depth from Zooming
– Considering an additional source of light onto the scene
i.e. Shape from Structured Light and Shape from Photometric
Stereo
– Considering additional surface information
i.e. Shape from Shading, Shape from Texture and Shape from
Geometric Constraints
– Multiple views
i.e. Shape from Stereo and Shape from Motion
Shape from Shading
7. 7
Lecture 4: Reconstruction from two views
4.1 Shape from X
Techniques based on:
– Modifying the intrinsic camera parameters
i.e. Depth from Focus/Defocus and Depth from Zooming
– Considering an additional source of light onto the scene
i.e. Shape from Structured Light and Shape from Photometric
Stereo
– Considering additional surface information
i.e. Shape from Shading, Shape from Texture and Shape from
Geometric Constraints
– Multiple views
i.e. Shape from Stereo and Shape from Motion
Shape from Stereo
8. 8
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
9. 9
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
10. 10
Lecture 4: Reconstruction from two views
4.2 Triangulation principle
World
coordinate
system
WZ
WY WX
WO W
Camera’
coordinate
system
'CY
'CX
'CZ
'CO 'C
'IY
'IX
'IO
'I
'RY
'RX 'R
'RO
Camera
coordinate
system
CY
CX CZ
CO
C
IY
IXIO
I
RY
RX
RO
R
11. 11
Lecture 4: Reconstruction from two views
4.2 Triangulation principle
uP
'uP
Camera
coordinate
system
CY
CX CZ
CO
C
Camera’
coordinate
system
'CY
'CX
'CZ
'CO 'C
World
coordinate
system
WZ
WY WX
WO W
IY
IXIO
I
RY
RX RO
'IY
'IX
'IO
'I
'RY
'RX
'R
'RO
R
dP
'dP
wP
Pw = Pu + m u
Pw = P’u + m’ v
Steps:
1 - Pu + m u = P’u + m’ v
2 - Expand to x,y,z
3 - Get m and m’
4 – Compute Pw
v
u
12. 12
Lecture 4: Reconstruction from two views
4.2 Triangulation principle
rP
sP
u
v
ˆ
wP
wP
3D Object Point
Reconstructed
Point
ˆ
wP
uP
'uP
Camera
coordinate
system
CY
CX CZ
CO
C
Camera’
coordinate
system
'CY
'CX
'CZ
'CO 'C
World
coordinate
system
WZ
WY WX
WO W
IY
IXIO
I
RY
RX RO
'IY
'IX
'IO
'I
'RY
'RX
'R
'RO
R
dP
'dP
wP
Pw = Pu + m u
Pw = P’u + m’ v
13. 13
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
WOc2
WOc1
WP2D2
WP2D1
u
v
pq
WP3D
http://astronomy.swin.edu.au/~pbourke/geometry/lineline3d/
Pa = P1 + mua (P2 - P1)
Pb = P3 + mub (P4 - P3)
Two different ways:
Minimize the distance between points:
Min || Pb - Pa ||2
Min || P1 + mua (P2 - P1) - P3 - mub (P4 - P3) ||2
Finding mua and mub once expanded to (x,y and z)
Compute the dot product between vectors:
(Pa - Pb)T (P2 - P1) = 0
(Pa - Pb)T (P4 - P3) = 0
Because they are perpendicular.
Finding mua and mub once expanded to Pa, Pb
and (x,y and z)
4.2 Triangulation principle
Pb
Pa
14. 14
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
In practice we can use Least-Squares:
1
34333231
24232221
14131211
1
11
11
Z
Y
X
AAAA
AAAA
AAAA
s
vs
us
1
34333231
24232221
14131211
2
22
22
Z
Y
X
BBBB
BBBB
BBBB
s
vs
us
CQX
QXC
1
Z
Y
X
BvBBvBBvB
BuBBuBBuB
AvAAvAAvA
AuAAuAAuA
vBB
uBB
vAA
uAA
232332223221231
132331223211231
231332213221131
131331213211131
23424
23414
13424
13414
4.2 Triangulation principle
Add additional rows if we have additiional views of the same point
15. 15
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
16. 16
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
17. 17
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
I’I
OC’
OC
m
m’
e e’
M
OI
OI’
OW
CKC’
4.3 Epipolar Geometry – Modelling
• Focal points, epipoles and epipolar lines
• e is defined by OC’ in {I}, e’ is defined by OC in {I’}
• m defines an epipolar line in {I’}; m’ defines an epipolar line in {I}
• All epipolar lines intersect at the epipole
l’
m
lm’
18. 18
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
Epipole
Epipole
Epipolar lines
Epipolar lines
Area 1
Area 2
Correspondence
pointsZoom
Area 1
Zoom
Area 2
Epipolar geometry of Camera 1 Epipolar geometry of Camera 2
4.3 Epipolar Geometry – Modelling
20. 20
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
I’I
OC’
OC
m
m’
e e’
M
OI
OI’
OW
CKC’
•The Epipolar Geometry concerns the problem of computing the plane .
• a plane is defined by the cross product between two vectors
• M is unknown, m and m’ are knowns
• {W} is located at {C} or {C’} and can be computed at {C} or {C’}
4 solutions
4.3 Epipolar Geometry – Modelling
21. 21
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
I’I
OC’
OC
m
m’
e e’
M
OI
OI’
lm’
l’m
OW
C’K’C
•The Epipolar Geometry concerns the problem of computing the plane .
• a plane is defined by the cross product between two vectors
• M is unknown, m and m’ are knowns
• {W} is located at {C} or {C’} and can be computed at {C} or {C’}
4 solutions
CKC’
4.3 Epipolar Geometry – Modelling
22. 22
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
I’I
OC’
OC
m
m’
e e’
M
OI
OI’
lm’
l’m
OW
• Assume {W} at {C}
CKC’
P
P’
tRPKP
tRK
IP
'
0
tRtRRPe
t
t
I
C
Pe
ttt
1
0
1
0
''
1
0
1
'
1
4.3 Epipolar Geometry – Modelling
23. 23
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
• Assume {W} at {C}
I’I
OC’
OC
m
m’
e e’
M
O
I
OI’
lm’
l’m
OW
CKC
’
P
P’
tRPKP
tRK
IP
'
0
tRtRRPe
t
t
I
C
Pe
ttt
1
0
1
0
''
1
0
1
'
1
''''
)('''
'
1
RmtRmtmPel
mtRmtRmRtRmPel
xm
x
tttt
m
Since epipolar lines are contained in the plane , we can define the line by
a cross product of two vectors, obtaining the orthogonal vector of the line.
4.3 Epipolar Geometry – Modelling
24. 24
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
I’I
OC’
OC
m
m’
e e’
M
O
I
OI’
lm’
l’m
OW
CKC
’
P
P’
'0
'0
Rmtm
mtRm
x
t
x
tt
The Fundamental matrix is
defined by inner product of a
point with its epipolar line.
'
'
' Rmtl
mtRl
xm
x
t
m
'
'''''
'' Rmtmlmlm
mtRmlmlm
x
t
m
t
m
x
tt
m
t
m
Orthogonal, their cosinus is 0
4.3 Epipolar Geometry – Modelling
25. 25
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
I’I
OC’
OC
m
m’
e e’
M
O
I
OI’
lm’
l’m
OW
CKC
’
P
P’
l’
mlm’
'~'''''~
~~
1
1
mmmm
mmmm
AA
AA
Now we consider the
intrinsics. Points in pixels
instead of metrics
'~'~'~'~'0
~''~~'~''0
111
111
mRtmmRtmRmtm
mtRmmtRmmtRm
x
tt
x
t
x
t
x
ttt
x
tt
x
tt
AAAA
AAAA
tttttt
AAAABAB
11
1
1
''
'
AA
AA
RtF
tRF
x
t
x
tt
0'~'~
0~'~
mFm
mFm
t
t
100
0
0
0
0
v
u
v
u
A
4.3 Epipolar Geometry – Modelling
26. 26
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
FtRtRRtRtF
FRtRttRtRF
x
ttt
x
tttt
x
ttt
x
tt
x
tttt
x
tt
x
tttt
x
ttt
11
11
'''''
'''''
AAAAAAAA
AAAAAAAA
F and F’ are related by a transpose. So,
t
t
FF
FF
'
'
Demonstration:
1
1
''
'
AA
AA
RtF
tRF
x
t
x
tt
The same dissertation can be made assuming the origin at
{C’}, obtaining two more fundamental matrices that are also
equivalent to F and F’.
4.3 Epipolar Geometry – Modelling
27. 27
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
The Essential Matrix is the calibrated case of the Fundamental matrix.
• The Intrinsic parameters are known: A and A’ are known
The problem is reduced to estimate E or E’.
1
1
''
'
AA
AA
RtF
tRF
x
t
x
tt
RtE
tRE
x
x
t
'
The monocular stereo is a symplified version of F where A = A’, reducing the
complexity of computing F.
1
1
'
AA
AA
RtF
tRF
x
t
x
tt
4.3 Epipolar Geometry – Modelling
28. 28
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
29. 29
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
30. 30
Lecture 4: Reconstruction from two views
4.4 Epipolar Geometry – Calibration
' ' 0T
m m F 1 ' 0
1
i
i i i
x
x y y
F
Operating, we obtain:
0nU f
11 12 13 21 22 23 31 32 33, , , , , , , ,
t
f F F F F F F F F F
, , , , , , , ,1i i i i i i i i i i i i iu x x y x x x y y y y x y
The epipolar geometry is defined as:
1 2, ,...,
t
n nU u u u
The Eight Point Method
31. 31
Lecture 4: Reconstruction from two views
1n nU f
11 12 13 21 22 23 31 32, , , , , , ,
t
f F F F F F F F F
, , , , , , ,i i i i i i i i i i i i iu x x y x x x y y y y x y
F is defined up to a scale factor, so we can fix one of the
component to 1. Let’s fix F33 = 1.
First solution is :
0nU f
0f NOT WANTED
Then:
1 1
1n n n nU U f U
1
1n nf U
1
1
t t
n n n nf U U U
Least-Squares
1 2, ,...,
t
n nU u u u
4.4 Epipolar Geometry – Calibration
The Eight Point Method with Least Squares
32. 32
Lecture 4: Reconstruction from two views
1
'
AA x
tt
tRF
First solution is :
0nU f
0f NOT WANTED
0
0
0
xy
xz
yz
t x
F has to be rank-2 because [tx] is rank-2.
Any system of equations:
can be solved by SVD so that f lies in the nullspace of Un= UDVT .
[U,D,V] = svd (Un)
Hence f corresponds to a multiple of the column of V that belongs to the
unique singular value of D equal to 0.
Note that f is only known up to a scaling factor.
11 12 13 21 22 23 31 32 33, , , , , , , ,
t
f F F F F F F F F F0nU f
4.4 Epipolar Geometry – Calibration
The Eight Point Method with Eigen Analysis
33. 33
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
34. 34
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
35. 35
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
Extrinsics: Camera Pose
I I C W
C Ws m A K M
' ' '
' '' ' ' 'I I C W
C Ws m A K M
3D Reconstruction:
'
'; 'I I
C CA A
'
'; 'C C
W WK K
Intrinsics: Optics & Internal Geometry
I’I
OC’
OC
m
m’
M
OI
OI’
OW
4.5 Constraints in stereo vision
Constraints:
• The Correspondence Problem F/E matrix
• Stereo Configurations:
• Calibrated Stereo: Intrinsics and Extrinsics known Triangulation!
• Uncalibrated Stereo: Intrinsics and Extrinsics unknown F matrix
• Calibrated Monocular: Intrinsics known, Extrinsics unknown E matrix
• Uncalibrated Monocular: Intrinsics and Extrinsics unknown F matrix
36. 36
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
37. 37
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
38. 38
Lecture 4: Reconstruction from two views
4.6 Experimental comparison – methods
Linear Iterative Robust Optimisation Rank-2
Seven point (7p) X — yes
Eight point (8p) X LS or Eig. no
Rank-2 constraint X LS yes
Iterative Newton-
Raphson
X LS no
Linear iterative X LS no
Non-linear
minimization in
parameter space
X Eig. yes
Gradient technique X LS or Eig. no
FNS X AML no
CFNS X AML no
M-Estimator X LS or Eig. no / yes
LMedS X 7p / LS or Eig. no
RANSAC X 7p / Eig no
MLESAC X AML no
MAPSAC X AML no
LS: Least-Squares Eig: Eigen Analysis AML: Approximate Maximum Likelihood
Least-squares Eigen Analysis Approximate Maximum
Likelihood
39. 39
Lecture 4: Reconstruction from two views
Image plane camera 1 Image plane camera 2
4.6 Experimental comparison – Methodology
40. 40
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
Methods: 1.- 7-Point; 2.- 8-Point with Least-Squares;
3.- 8-Point with Eigen Analysis 4.- Rank-2 Constraint
* Mean and Std. in pixels
*
Linear methods: Good results if the points are well located and no outilers
mean
std
4.6 Experimental comparison – Synthetic images
41. 41
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views* Mean and Std. in pixels
*
Iterative methods: Can cope with noise but inefficient in the presence of outliers
Methods: 5.- Iterative Linear; 6.- Iterative Newton-Raphson;
7.- Minimization in parameter space;
8.- Gradient using LS; 9.- Gradient using Eigen;
10.- FNS; 11.- CFNS
4.6 Experimental comparison – Synthetic images
42. 42
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views* Mean and Std. in pixels
Robust methods: Cope with both noise and outliers
Methods: 12.- M-Estimator using LS; 13.- M-Estimator using Eigen;
14.- M-Estimator proposed by Torr;
15.- LMedS using LS; 16.- LMedS using Eigen;
17.- RANSAC; 18.- MLESAC; 19.- MAPSAC.
4.6 Experimental comparison – Synthetic images
43. 43
Lecture 4: Reconstruction from two viewsLecture 4: Reconstruction from two views
1.- 7-Point; 2.- 8-Point with Least-Squares; 3.- 8-Point with Eigen Analysis; 4.- Rank-2 Constraint;
5.- Iterative Linear; 6.- Iterative Newton-Raphson; 7.- Minimization in parameter space; 8.- Gradient using LS;
9.- Gradient using Eigen; 10.- FNS; 11.- CFNS; 12.- M-Estimator using LS; 13.- M-Estimator using Eigen;
14.- M-Estimator proposed by Torr; 15.- LMedS using LS; 16.- LMedS using Eigen; 17.- RANSAC;
18.- MLESAC; 19.- MAPSAC.
Linear Iterative Robust
Computing Time
4.6 Experimental comparison – Synthetic images
45. 45
Lecture 4: Reconstruction from two views
Methods: 1.- 7-Point; 2.- 8-Point with Least-Squares;
3.- 8-Point with Eigen Analysis 4.- Rank-2 Constraint
Methods: 5.- Iterative Linear; 6.- Iterative Newton-Raphson;
7.- Minimization in parameter space;
8.- Gradient using LS; 9.- Gradient using Eigen;
10.- FNS; 11.- CFNS
Methods: 12.- M-Estimator using LS; 13.- M-Estimator using Eigen;
14.- M-Estimator proposed by Torr;
15.- LMedS using LS; 16.- LMedS using Eigen;
17.- RANSAC; 18.- MLESAC; 19.- MAPSAC.* Mean and Std. in pixels
*
4.6 Experimental comparison – Real images
46. 46
Lecture 4: Reconstruction from two views
• Survey of 15 methods of computing F and up to 19 different
implementations
• Description of the estimators from an algorithmic point of view
• Conditions: Gaussian noise, outliers and real images
– Linear methods: Good results if the points are well located and
the correspondence problem previously solved (without outliers)
– Iterative methods: Can cope with noise but inefficient in the
presence of outliers
– Robust methods: Cope with both noise and outliers
• Least-squares is worse than eigen analysis and approximate
maximum likelihood
• Rank-2 matrices are preferred if a good geometry is required
• Better results when data are previously normalized
4.6 Experimental comparison – Conclusions
47. 47
Lecture 4: Reconstruction from two views
Publications
– X. Armangué and J. Salvi. Overall View Regarding Fundamental Matrix
Estimation. Image and Vision Computing, IVC, pp. 205-220, Vol. 21, Issue
2, February 2003.
– J. Salvi. An approach to coded structured light to obtain three dimensional
information. PhD Thesis. University of Girona, 1997. Chapter 3.
– J. Salvi, X. Armangué, J. Pagès. A survey addressing the fundamental
matrix estimation problem. IEEE International Conference on Image
Processing, ICIP 2001, Thessaloniki, Greece, October 2001.
More Information: http://eia.udg.es/~qsalvi/
4.6 Experimental comparison – Conclusions
48. 48
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
49. 49
Lecture 4: Reconstruction from two views
Contents
4. Reconstruction from two views
4.1 Shape from X
4.2 Triangulation principle
4.3 Epipolar geometry – Modelling
4.4 Epipolar geometry – Calibration
4.5 Constraints in stereo vision
4.6 Experimental comparison of methods
4.7 Sample: Mobile robot performing 3D mapping
50. 50
Lecture 4: Reconstruction from two views
4.7 Sample: Mobile robot performing 3D mapping
• Building a 3D map from an
unknown environment
using a stereo camera
system
• Localization of the robot in
the map
• Providing a new useful
sensor for the robot control
architecture GRILL Mobile robot with a
stereo camera system
51. 51
Lecture 4: Reconstruction from two views
Onboard Control Computer
Microcontroller
Onboard Vision Computer
Frame Grabber A Frame Grabber B
Motor Encoder Sonar
Camera A Camera B
Wireless Ethernet
Radio video emitter
Ethernet Card
PCI Bus
Ethernet Card
PCI Bus
Ethernet Link
PC104+PC104+
USB
RS-232
Control system
Vision system
Pioneer 2
Outside stereo vision systemInside stereo vision system
GRILL Mobile Robot
4.7 3D mapping – Robot components
52. 52
Lecture 4: Reconstruction from two views
Image Acquisition
Image Processing
Low Level
Image Processing
High Level
Description Level
Camera
LocalizationMap Building
3D Information Motion
Estimation
Correspondence
Problem
Feature
Extraction
A/D
Gradients
Filtering
Calibration
Remove
Distortion
Tracking
Camera Modelling
and Calibration
Localization and
Map Building
Stereo Vision and
Reconstruction
LocalizationMap Building
Calibration
Remove
Distortion
3D Information
Correspondence
Problem
Motion
Estimation
4.7 3D mapping – Data flow diagram
53. 53
Lecture 4: Reconstruction from two views
2D Image Processing
3D Image Processing
Map Building and
Localization
Camera A Camera B
Sequence A Sequence B
2D Points
3D Points
3D Points Position
3D Map Trajectory
Image Flow
2D Points Flow
3D Points Flow
Position Flow
4.7 3D mapping – Data flow diagram
54. 54
Lecture 4: Reconstruction from two views
2D Image Processing
3D Image Processing
Map Building and
Localization
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Data flow diagram
55. 55
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
• Cameras are calibrated
• Both stereo images are
obtained simultaneously
4.7 3D mapping – Input sequence
56. 56
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – RGB to I
• Description
– Converting a color image
to an intensity image
• Input
– Color image (RGB)
• Output
– Intensity image
57. 57
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
• Description
– Removing distortion of an
image using camera
calibration parameters
• Input
– Distorted image
• Output
– Undistorted image
4.7 3D mapping – Remove Distortion
Intensity ImagesUndistorted Images
58. 58
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Corners
• Description
– Detection of corners using
a variant of Harris corners
detector
• Input
– Undistorted image
• Output
– Corners list
Undistorted ImagesCorners Detected
59. 59
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Spatial Cross Correlation
• Description
– Spatial cross correlation
using fundamental matrix
obtained from camera
calibration parameters
• Input
– Undistorted image A
– Corners list A
– Undistorted image B
– Corners list B
• Output
– Spatial points list
– Spatial matches list
Corners DetectedPoints and matches list
60. 60
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Temporal Cross Correlation
• Description
– Temporal cross correlation
using small windows
search
• Input
– Previous undistorted
image
– Previous corners list
– Current undistorted image
– Current corners list
• Output
– Temporal points list
– Temporal matches list
Corners DetectedPoints and matches list
61. 61
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Stereo Reconstruction
• Description
– Stereo reconstruction by
triangulation using camera
calibration parameters
• Input
– Spatial points list
– Spatial matches list
• Output
– 3D points list
Points and matches list
3D points list
62. 62
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – 3D Tracker
• Description
– Tracking 3D points using
temporal cross correlation
• Input
– 3D points list
– Temporal points list A
– Temporal matches list A
– Temporal point list B
– Temporal matches list B
• Output
– 3D points history
– Points history A
– Matches history A
– Points history B
– Matches history B
Points and matches tracker
3D points
tracker
63. 63
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Outliers Detection
• Description
– Detection of outliers
comparing distance
between current and
previous 3D points list
• Input
– Odometry position
– Current 3D points list
– Previous 3D points list
• Output
– Outliers list
Current and previous 3D points with outliers
Outlier Example
Current and Previous 3D points without outliers
64. 64
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Local Localization
• Description
– Computing the absolute
position from the map by
minimizing distance
between the projection of
3D map points in cameras
and current 2D points and
matches.
• Input
– Odometry position as
initial guest
– Previous 3D points list
from the map
– Current 2D points list
– Current 2D matches list
• Output
– Absolute position
Map absolute position
65. 65
Lecture 4: Reconstruction from two views
4.7 3D mapping – Global Localization
• Description
– Computing the trajectory
effect by the robot
• Input
– Local position
• Output
– Global position
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
66. 66
Lecture 4: Reconstruction from two views
RGB to I RGB to I
Remove
Distortion
Remove
Distortion
Corners Corners
Spatial Cross
Correlation
Temp. Cross
Correlation
Temp. Cross
Correlation
Image Points
Buffer A
Image Points
Buffer B
Camera A Camera B
Image
Buffer B
Image
Buffer A
Stereo
Reconstruction
3D Tracker
Outliers
Detection
Local
Localization
Global
Localization
Build 3D Map
3D Map Trajectory
4.7 3D mapping – Building 3D Map
• Description
– Building the 3D map from
the 3D points with a
history longer than n times
• Input
– Global position
– Current 3D points list
– Previous 3D points list
• Output
– Global position
3D Map