SlideShare a Scribd company logo
1 of 62
Download to read offline
Introduction to Modern
LiDAR for Machine
Perception
Robert Laganière
CEO / Professor
Sensor Cortek inc. / University of Ottawa
Canada
• LiDAR fundementals
• LiDAR principles
• LiDAR taxonomy
• LiDAR processing (AI)
Presentation outline
2
© 2023 Sensor Cortek inc
• LiDAR (or CoLiDAR) name derived from the RADAR acronym
• LiDAR refers to the technology that uses a laser to sense the
environment
• Same fundamental principle as radar (or sonar)
• Emitting a signal and analyzing the bounced back signal
(the echo)
LiDAR: Light Detection and Ranging
3
© 2023 Sensor Cortek inc
Main benefit of using coherent collimated light
beam (i.e., laser)
4
© 2023 Sensor Cortek inc
radar signal
laser
light
• The laser light used by LiDAR has two interesting properties:
• It is coherent: all emitted light rays have the same
frequency and phase
• It is collimated: the beam of light has parallel rays and
spread minimally
LiDAR sequence
5
© 2023 Sensor Cortek inc
Corresponding camera sequence
6
© 2023 Sensor Cortek inc
did you perceive the same things?
• It produces direct 3D information
• It provides accurate 3D measurements
• From under a mm to few cm depending on the distance
• It’s an active sensor that operates day and night
• It can capture information at long range
• ~200 m
• It has a large FoV
• even 360o
What makes LiDAR an attractive sensor?
7
© 2023 Sensor Cortek inc
• It captures shape but not appearance
• It produces sparse data
• Sometimes only few points on an object
• It becomes noisy under fog, snow and rain
• It is still an expensive sensor
• Some are thousands of $
• While radars and cameras could be less than $100
• It often includes mechanical parts
However, LiDAR is not perfect
8
© 2023 Sensor Cortek inc
• A collection of
sparse 3D points
• A LiDAR frame
• Points not
captured at
exactly the
same time….
The LiDAR point cloud
9
© 2023 Sensor Cortek inc
• 1900s: Planck, Einstein and others “discovered” the photon
• May 16th 1960: first laser light produced (T. Maiman)
• A laser is a device that generates an intense beam
of coherent monochromatic light
• Light Amplification by Stimulated Emission of Radiation
• Laser differs from other light sources because it emits coherent
collimated light
• Laser light is very narrow, making it possible to see the smallest
details with high resolution at relatively long distances.
History of Laser
10
© 2023 Sensor Cortek inc
• 1960s: LiDAR initially developed for metrology and atmospheric
research
• But the idea of probing the atmosphere with light can be dated back
to the 1930s
• 1970s: LiDAR used for terrain mapping
• Apollo missions used laser to accurately measure Earth-Moon
distances
• 1980s: with the advent of GPS and inertial measurement units
(IMUs), LiDAR became very accurate
• 2005: first AV to complete the DARPA Grand Challenge (142 mile
desert course) was equipped with a LiDAR
History of LiDAR
© 2023 Sensor Cortek inc
Mt St-Helens
Wikimedia commons
11
1. The LiDAR emits pulsed light waves
2. The light potentially hits a surrounding object and bounces back
to the sensor
3. The sensor reads the bounced signal, estimates the time it took
to return to the LiDAR and measures the reflected light energy
• Simple technology; almost instantaneous!
• Potential interference from the sun and other LiDARs
How does LiDAR work? Pulsed LiDAR
12
© 2023 Sensor Cortek inc
How does LiDAR work?
13
© 2023 Sensor Cortek inc
blind region
1.
3.
2.
This configuration is called bistatic (most common and cheaper)
A monostatic optical system aligns the Tx and Rx for better detection
1. LiDAR can also use continuous waves
• Frequency Modulated Continuous Wave
2. The phase of the bounced back signal will differ from the emitted signal
3. The change in phase is used to extract the distance information
• This is done by mixing the emitted and received signals (as done in
radar)
• Velocity is a bonus!
• Virtually no interference
• You must read a longer signal (stay longer at each point)
How does LiDAR work? FMCW Lidar
14
© 2023 Sensor Cortek inc
15
© 2023 Sensor Cortek inc
How does LiDAR work? FMCW Lidar
www.bridgerphotonics.com/blog/frequency-
modulated-continuous-wave-fmcw-lidar
• LiDAR physics is governed by one simple equation
Distance of the object = (Speed of Light x Time of Flight) / 2
• But to be able to read the received light, you need power
(i.e., enough photons bouncing back)
Power received ≈ Power transmitted x Cross Section x Optic area
Distance2 Distance2
• The laser cross section is the average amount of optical
power returned by the target
LiDAR mathematics
16
© 2023 Sensor Cortek inc
LiDAR sensor taxonomy
© 2023 Sensor Cortek inc
LiDAR
Scanning
LiDAR
Mechanical
Scanners
Rotating
mirrors
Rotating
prisms
MEMS
mirrors
Electronic
Scanners
Optical
Phased
Array
Imaging
LiDAR
Flash LiDAR
Solid-state LiDAR
17
• Three strategies:
• Scanning LiDAR
• A laser scans the scene and a single photodetector is used to read
the returned photons
• Flash LiDAR
• The entire field of view is illuminated and a photodetector array
captures the received photons
• Optical Phased Arrays
• Several transmitters emitting laser light at different phases enabling
the steering of the beam (constructive/destructive interference)
Capturing a scene with LiDAR
18
© 2023 Sensor Cortek inc
• Shorter range – higher frame rate
• Costly focal plane array
• Limited FoV
• Light is distributed across the FoV – more noisy
• Pixels are small – more power required
• Angular resolution determined by the pixel density
• No motion distortion
• No moving parts
Flash LiDAR
19
© 2023 Sensor Cortek inc
• Longer range – smaller frame rate
• Expensive scanning mechanism
• e.g., spinning mirrors, MEMS mirrors
• or rotate everything
• Can be bulky
• LiDAR motion must be compensated
• Less tolerant to mechanical vibrations
Scanning LiDAR
20
© 2023 Sensor Cortek inc
• Today’s most popular solution
• Heavier than other solutions
• Vulnerable to vibrations
• Generally includes stack of
photodetectors to scan in
several horizontal layers
Scanning with rotating mirrors
(or rotate everything)
21
© 2023 Sensor Cortek inc
precisionlaserscanning.com/2017/12/mems-mirrors-vs-polygon-
scanners-for-lidar-in-autonomous-vehicles/
• Uses two (or more) sequential prisms
to steer the beam
• Shape of prism and speed of rotation
determines the scan pattern
• Limited FoV
Scanning with rotating prisms
22
© 2023 Sensor Cortek inc
Lidars for vehicles: from the requirements to the
technical evaluation, Z. Dai et al., Conference:
9th International Forum on Automotive Lighting,
2021
• Micro-Electro-Mechanical System
• Quasi solid-state
• Programmable scan patterns
• Limited FoV
• Requires careful calibration
Scanning with MEMS mirrors
23
© 2023 Sensor Cortek inc
preciseley.com/product/mems-scanning-mirror/
• Lower resolution – high frame rate
• Solid state
• No moving parts
• Smart zooming capability
• Interference from the antenna lobes
limits the angular resolution
• Complex design
• Lower production cost
• Probably the solution of the future
Optical phased array
24
© 2023 Sensor Cortek inc
MEMS Mirrors for LiDAR: A Review,
D. Wang, C. Watkins, H. Xie. Micromachines 2020, 11(5)
Rotating Prisms MEMS OPA Flash
Range Long Long Long Medium Low
Frame rate Low Low Low High High
FoV Large Limited Limited Limited Limited
Resolution High High High Adaptive Low
Power High High Low Low High
Solid Sate No No Quasi Yes Yes
Vulnerability High High High Low Low
Complexity Low Low High High Low
Cost High High High Low High
Lidar technologies
25
© 2023 Sensor Cortek inc
• LiDAR operates in the near IR
spectrum
• 780 nm to 3000 nm
• Be careful about eye safety!
• Typical field of view:
• 90°, 180° or 360°
• Depth resolution determined by the
temporal sampling frequency
• ∆D = c / 2f
• e.g., a depth resolution of 1 cm
requires a 1.5 GHz sampling rate
Some LiDAR specs
26
© 2023 Sensor Cortek inc
• Angular resolution determined by the
scanning point rate
• e.g., 0.1° corresponds to 18 cm at 100 m
• Pulse frequency
• e.g., 2 ns corresponds to a range
resolution of 3 cm
• Pulse frequency determines the
number of points per second per layer
• e.g., 40 kHz
Specs are usually given for 80%
Lambertian reflectivity
• When the vehicle moves, the LiDAR is scanning a moving scene
• Which will distort the point cloud
• This distortion is proportional to the vehicle’s speed and
inversely proportional to laser scanning rate
• Solution: using an inertial measurement unit in order to
compensate for the sensor motion
• IMUs provide acceleration and angular velocity
• Assumption: the ego vehicle motion has constant angular and
linear velocities
LiDAR motion compensation using IMU
27
© 2023 Sensor Cortek inc
LiDAR motion compensation using IMU
28
© 2023 Sensor Cortek inc
https://www.mathworks.com/help/lidar/ug/m
otion-compensation-in-lidar-point-cloud.html
Before motion compensation After motion compensation
• A LiDAR sensor produces a frame of 3D points
• A point cloud
• This point cloud has to be processed and analyzed in order to
interpret the scene
• e.g., detect objects on the road
• But LiDAR data is sparse and unstructured
• Signal processing (convolution) prefers regular grid sampling
• LiDAR needs to be preprocessed to build a more suitable
representation
Processing LiDAR data for detection
29
© 2023 Sensor Cortek inc
• Possible representations:
• Point sets
• Voxelization
• Bird’s eye view
• Point pillars encoding
• Frontal image generation
• Sparse convolution
LiDAR representation strategies
30
© 2023 Sensor Cortek inc
• Point sets
• Can we work directly on the point cloud?
• Point-based representation
• Voxelization
• Bird’s eye view
• Point pillars encoding
• Frontal image generation
• Sparse convolution
LiDAR representations
31
© 2023 Sensor Cortek inc
• The idea is to consume an unordered set of 3D points
• Transformations are learned to normalize the data
• Global point features are learned from the set
• Point cloud has to be segmented into small regions of
interest
• More difficult to apply in a complex scene composed of
many objects
Point-based representations
32
© 2023 Sensor Cortek inc
C. R. Qi, H. Su, K. Mo, and L. J. Guibas. Pointnet: Deep
learning on point sets for 3d classification and segmentation.
Proc. Computer Vision and Pattern Recognition (CVPR) 2017
Point-based representations
33
© 2023 Sensor Cortek inc
www.mdpi.com/1999-4907/12/2/131/htm
• The idea is to
consume an
unordered set of 3D
points
• Point sets
• Voxelization
• e.g., occupancy grid
• Bird’s eye view
• Point pillars encoding
• Frontal image generation
• Sparse convolution
LiDAR representations
34
© 2023 Sensor Cortek inc
Attribute Filtering of Urban Point Clouds Using Max-Tree on Voxel Data, F.
Guillotte, Mathematical Morphology and Its Applications to Signal and Image
Processing, 2019.
• A 3D voxel grid is created in which
each voxel contains:
• A scalar value
• A vector made of statistics
computed from the points inside
the voxel
• Mean, variance, reflectance, …
• By nature, the occupied voxels are
very sparse
• 3D convolution expensive and
inefficient
Voxelization
35
© 2023 Sensor Cortek inc
Wang, D.Z.; Posner, I. Voting for voting in online point cloud
object detection. In Proceedings of the Robotics: Science and
Systems, Rome, Italy, 13–17 July 2015
• Voxel feature encoding (VFE) is used
• from randomly sampled points in
each voxel
• using point coordinates and
reflectance
• and a fully connected network
transformation
• 3D convolution is applied on the VFE
Voxelization: VoxelNet
36
© 2023 Sensor Cortek inc
Y. Zhou and O. Tuzel. Voxelnet: End-to-end learning for
point cloud based 3d object detection. In CVPR, 2018
VFE
learning
Convolution
layer
Detection
network
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Vehicle Detection results – KITTI Dataset
37
© 2023 Sensor Cortek inc
• Point sets
• Voxelization
• Bird’s eye view
• Point pillars encoding
• Frontal image generation
• Sparse comvolution
LiDAR representations
38
© 2023 Sensor Cortek inc
• Computationally efficient
• Preserve the metric space for objects
on the road
• The representation produces an
image
• Height becomes a channel
• Image detection network can be
used
• Objects are not occluded
Bird’s eye view network : Pixor
39
© 2023 Sensor Cortek inc
Yang, B., Luo, W., Urtasun, R.: PIXOR: real-time 3D
object detection from point clouds. CVPR (2018)
Vehicle Detection results – KITTI Dataset
40
© 2023 Sensor Cortek inc
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
• Voxelization
• Bird’s eye view
• Pillars encoding
• Each pillar encodes point
distance to centroid and
reflectance
• Simplified PointNet is used
• Frontal image generation
• Sparse convolution
LiDAR representations
41
© 2023 Sensor Cortek inc
becominghuman.ai/pointpillars-3d-point-clouds-bounding-box-
detection-and-tracking-pointnet-pointnet-lasernet-67e26116de5a
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
PointPillar 92.07 87.74 86.65
Vehicle Detection results – KITTI Dataset
42
© 2023 Sensor Cortek inc
• Voxelization
• Bird’s eye view
• Point pillars encoding
• Frontal image generation
• Sparse convolution
LiDAR representations
43
© 2023 Sensor Cortek inc
• BEV + LiDAR project + Camera
• BEV is used to propose potential objects
• Multiview features are then uses to predict objects
Object-based feature extractor: MV3D
44
© 2023 Sensor Cortek inc
Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d
object detection network for autonomous driving. In
Proceedings of the IEEE Conference on Computer Vision
and Pattern Recognition, Honolulu, HI, USA, 21–26 July
2017
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
PointPillar 92.07 87.74 86.65
MV3D 86.55 78.1 76.67
Vehicle Detection results – KITTI Dataset
45
© 2023 Sensor Cortek inc
Frontal view fusion
46
© 2023 Sensor Cortek inc
• All images are fused together
• Including the camera image
• Lidar/camera fusion
• BGF fusion operator
Sensor Fusion Operators for Multimodal 2D Object Detection,
M.M. Pasandi, T. Liu, Y. Massoud, R, Laganiere, ISVC 2022
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
PointPillar 92.07 87.74 86.65
MV3D 86.55 78.1 76.67
BGF Fusion 94.90 88.40 78.38
Vehicle Detection results – KITTI Dataset
47
© 2023 Sensor Cortek inc
• Voxelization
• Bird’s eye view
• Pillars encoding
• Frontal image generation
• Sparse convolution
• When the data is very sparse, regular convolution becomes very
inefficient
• The idea is to compress the representation by ignoring zero values
• To this end, look-up tables are often used
LiDAR representations
48
© 2023 Sensor Cortek inc
SECOND: Sparse convolution on point features
49
© 2023 Sensor Cortek inc
SECOND: Sparsely Embedded Convolutional
Detection by Yan Yan, Yuxing Mao, Bo Li, Sensors
2018.
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
PointPillar 92.07 87.74 86.65
MV3D 86.55 78.1 76.67
BGF Fusion 94.90 88.40 78.38
SECOND 91.92 87.92 85.39
Vehicle Detection results – KITTI Dataset
50
© 2023 Sensor Cortek inc
PV-RCNN: Point-based + Voxels
51
© 2023 Sensor Cortek inc
PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection (2020)
Shaoshuai Shi, Chaoxu Guo, Li Jiang, Zhe Wang, Jianping Shi, Xiaogang Wang, Hongsheng Li, IEEE Conference on Computer
Vision and Pattern Recognition (CVPR).
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
PointPillar 92.07 87.74 86.65
MV3D 86.55 78.1 76.67
BGF Fusion 94.90 88.40 78.38
SECOND 91.92 87.92 85.39
PV-RCNN 92.86 88.93 88.74
Vehicle Detection results – KITTI Dataset
52
© 2023 Sensor Cortek inc
• Solving the LiDAR sparsity problem
• The farther the object, the sparser the point density
How to further improve detection?
53
© 2023 Sensor Cortek inc
• Depth completion network
• Can be used to densify the point cloud
Pseudo LiDAR
54
© 2023 Sensor Cortek inc
www.mdpi.com/1424-8220/22/18/6969
Depth Completion with Twin Surface Extrapolation at Occlusion
Boundaries, Saif Imran, Xiaoming Liu, Daniel Morris, CVPR 2021
BTC: using shape completion
55
© 2023 Sensor Cortek inc
Behind the Curtain: Learning Occluded
Shapes for 3D Object Detection, Qiangeng
Xu, Yiqi Zhong, Ulrich Neumann, AAAI 2022
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
PointPillar 92.07 87.74 86.65
MV3D 86.55 78.1 76.67
BGF Fusion 94.90 88.40 78.38
SECOND 91.92 87.92 85.39
PV-RCNN 92.86 88.93 88.74
BTC 93.46 89.53 87.44
Vehicle Detection results – KITTI Dataset
56
© 2023 Sensor Cortek inc
LiDAR and pseudo-LiDAR: SFD
57
© 2023 Sensor Cortek inc
• Lidar/camera fusion
Sparse Fuse Dense: Towards High Quality 3D Detection with Depth
Completion, Wu, Xiaopei and Peng, Liang and Yang, Honghui and Xie, Liang
and Huang, Chenxi and Deng, Chengqi and Liu, Haifeng and Cai, Deng}, 2022
Mean Average Precision (%)
Easy Medium Hard
VoxelNet 89.35 79.26 77.39
Pixor 81.7 77.05 72.95
PointPillar 92.07 87.74 86.65
MV3D 86.55 78.1 76.67
BGF Fusion 94.90 88.40 78.38
SECOND 91.92 87.92 85.39
PV-RCNN 92.86 88.93 88.74
BTC 93.46 89.53 87.44
SFD 95.85 91.92 91.41
58
© 2023 Sensor Cortek inc
Vehicle Detection results – KITTI Dataset
LiDAR densification: UYI
59
© 2023 Sensor Cortek inc
• Pseudo LiDAR module
Use Your Imagination: A Detector-Independent Approach For LiDAR
Quality Booster, Z. Zhang, T. Liu, R. Laganiere, 2023
LiDAR
detector
network
LiDAR densification: UYI
60
© 2023 Sensor Cortek inc
• Vehicle detection performance boosting
Use Your Imagination: A Detector-Independent Approach For LiDAR
Quality Booster, Z. Zhang, T. Liu, R. Laganiere, 2023
Conclusion
61
© 2023 Sensor Cortek inc
• LiDAR provides accurate 3D detection
• An essential component in ADAS/AV
• LiDAR technology will continue to evolve
• Improve density
• Lower power
• Solid-state
• Lower cost
• Radar / LiDAR Convergence
• C. R. Qi, H. Su, K. Mo, and L. J. Guibas. Pointnet: Deep learning on point sets for 3d classification and
segmentation. CVPR 2017.
• Yang, B., Luo, W., Urtasun, R.: PIXOR: real-time 3D object detection from point clouds. CVPR 2018.
• X. Chen, H. Ma, J. Wan, B. Li, T. Xia, Multi-view 3d object detection network for autonomous driving.
CVPR 2017.
• Yan Yan, Yuxing Mao, Bo Li. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018.
• S. Shi, C. Guo, L. Jiang, Z. Wang, J. Shi, X. Wang, H. Li. PV-RCNN: Point-Voxel Feature Set Abstraction
for 3D Object Detection. CVPR 2020.
• X. Wu, L. Peng, H. Yang, L. Xie, C. Huang, C. Deng, H. Liu, D. Cai. Sparse Fuse Dense: Towards High
Quality 3D Detection with Depth Completion, CVPR 2022.
• Q. Xu, Y. Zhong, U. Neumann. Behind the Curtain: Learning Occluded Shapes for 3D Object Detection.
AAAI 2022.
References
62
© 2023 Sensor Cortek inc
Reading list on LiDAR and AI

More Related Content

Similar to “Introduction to Modern LiDAR for Machine Perception,” a Presentation from the University of Ottawa

Lidar : light detection and rangeing
Lidar : light detection and rangeingLidar : light detection and rangeing
Lidar : light detection and rangeingRahul Bhagore
 
Lidar- light detection and ranging
Lidar- light detection and rangingLidar- light detection and ranging
Lidar- light detection and rangingKarthick Subramaniam
 
Airborne Laser Scanning Remote Sensing with LiDAR.ppt
Airborne Laser Scanning Remote Sensing with LiDAR.pptAirborne Laser Scanning Remote Sensing with LiDAR.ppt
Airborne Laser Scanning Remote Sensing with LiDAR.pptssuser6358fd
 
Introduction to Lasers
Introduction to LasersIntroduction to Lasers
Introduction to LasersAby Benz
 
Copy of LiDAR for Autonomous Driving _20231011_084847_0000.pdf
Copy of LiDAR for Autonomous Driving _20231011_084847_0000.pdfCopy of LiDAR for Autonomous Driving _20231011_084847_0000.pdf
Copy of LiDAR for Autonomous Driving _20231011_084847_0000.pdfNanDhu551039
 
DIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technology
DIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technologyDIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technology
DIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technologyDilshanDillu1
 
Laser communication
Laser communicationLaser communication
Laser communicationrahul dogra
 
Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...
Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...
Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...Aritra Sarkar
 
LiDAR Technology and Geospatial Services
LiDAR Technology and Geospatial Services LiDAR Technology and Geospatial Services
LiDAR Technology and Geospatial Services MattBethel1
 
IRJET - Lidar based Autonomous Robot
IRJET - Lidar based Autonomous RobotIRJET - Lidar based Autonomous Robot
IRJET - Lidar based Autonomous RobotIRJET Journal
 
Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)NopphawanTamkuan
 
Lidar final ppt
Lidar final pptLidar final ppt
Lidar final pptrsarnagat
 
Laser Communications
Laser CommunicationsLaser Communications
Laser CommunicationsSeminar Links
 

Similar to “Introduction to Modern LiDAR for Machine Perception,” a Presentation from the University of Ottawa (20)

Lidar
LidarLidar
Lidar
 
Lidar : light detection and rangeing
Lidar : light detection and rangeingLidar : light detection and rangeing
Lidar : light detection and rangeing
 
Lidar and sensing
Lidar and sensingLidar and sensing
Lidar and sensing
 
Lidar- light detection and ranging
Lidar- light detection and rangingLidar- light detection and ranging
Lidar- light detection and ranging
 
Airborne Laser Scanning Remote Sensing with LiDAR.ppt
Airborne Laser Scanning Remote Sensing with LiDAR.pptAirborne Laser Scanning Remote Sensing with LiDAR.ppt
Airborne Laser Scanning Remote Sensing with LiDAR.ppt
 
Lidar
LidarLidar
Lidar
 
Lidar
LidarLidar
Lidar
 
Introduction to Lasers
Introduction to LasersIntroduction to Lasers
Introduction to Lasers
 
Copy of LiDAR for Autonomous Driving _20231011_084847_0000.pdf
Copy of LiDAR for Autonomous Driving _20231011_084847_0000.pdfCopy of LiDAR for Autonomous Driving _20231011_084847_0000.pdf
Copy of LiDAR for Autonomous Driving _20231011_084847_0000.pdf
 
DIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technology
DIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technologyDIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technology
DIGITAL RADIOGRAPHY FOR bachelor of science in medical imaging technology
 
Laser communication
Laser communicationLaser communication
Laser communication
 
Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...
Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...
Elevation mapping using stereo vision enabled heterogeneous multi-agent robot...
 
LiDAR Technology and Geospatial Services
LiDAR Technology and Geospatial Services LiDAR Technology and Geospatial Services
LiDAR Technology and Geospatial Services
 
IRJET - Lidar based Autonomous Robot
IRJET - Lidar based Autonomous RobotIRJET - Lidar based Autonomous Robot
IRJET - Lidar based Autonomous Robot
 
Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)
 
Lidar final ppt
Lidar final pptLidar final ppt
Lidar final ppt
 
Laser Communications
Laser CommunicationsLaser Communications
Laser Communications
 
Laser Communications
Laser CommunicationsLaser Communications
Laser Communications
 
Eastern WV LiDAR Acquisition
Eastern WV LiDAR Acquisition Eastern WV LiDAR Acquisition
Eastern WV LiDAR Acquisition
 
Laser.pptx
Laser.pptxLaser.pptx
Laser.pptx
 

More from Edge AI and Vision Alliance

“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...Edge AI and Vision Alliance
 
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...Edge AI and Vision Alliance
 
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...Edge AI and Vision Alliance
 
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...Edge AI and Vision Alliance
 
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...Edge AI and Vision Alliance
 
“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...Edge AI and Vision Alliance
 
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsightsEdge AI and Vision Alliance
 
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...Edge AI and Vision Alliance
 
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...Edge AI and Vision Alliance
 
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...Edge AI and Vision Alliance
 
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...Edge AI and Vision Alliance
 
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...Edge AI and Vision Alliance
 
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...Edge AI and Vision Alliance
 
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...Edge AI and Vision Alliance
 
“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from Samsara“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from SamsaraEdge AI and Vision Alliance
 
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...Edge AI and Vision Alliance
 
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...Edge AI and Vision Alliance
 
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...Edge AI and Vision Alliance
 
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...Edge AI and Vision Alliance
 
“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...
“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...
“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...Edge AI and Vision Alliance
 

More from Edge AI and Vision Alliance (20)

“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
 
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
 
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
 
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
 
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
 
“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...
 
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
 
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
 
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
 
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
 
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
 
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
 
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
 
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
 
“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from Samsara“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from Samsara
 
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
 
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
 
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
 
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
 
“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...
“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...
“Tracking and Fusing Diverse Risk Factors to Drive a SAFER Future,” a Present...
 

Recently uploaded

Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...shyamraj55
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticscarlostorres15106
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsMiki Katsuragi
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDGMarianaLemus7
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024The Digital Insurer
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clashcharlottematthew16
 

Recently uploaded (20)

Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort ServiceHot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
 
Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering Tips
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDG
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clash
 

“Introduction to Modern LiDAR for Machine Perception,” a Presentation from the University of Ottawa

  • 1. Introduction to Modern LiDAR for Machine Perception Robert Laganière CEO / Professor Sensor Cortek inc. / University of Ottawa Canada
  • 2. • LiDAR fundementals • LiDAR principles • LiDAR taxonomy • LiDAR processing (AI) Presentation outline 2 © 2023 Sensor Cortek inc
  • 3. • LiDAR (or CoLiDAR) name derived from the RADAR acronym • LiDAR refers to the technology that uses a laser to sense the environment • Same fundamental principle as radar (or sonar) • Emitting a signal and analyzing the bounced back signal (the echo) LiDAR: Light Detection and Ranging 3 © 2023 Sensor Cortek inc
  • 4. Main benefit of using coherent collimated light beam (i.e., laser) 4 © 2023 Sensor Cortek inc radar signal laser light • The laser light used by LiDAR has two interesting properties: • It is coherent: all emitted light rays have the same frequency and phase • It is collimated: the beam of light has parallel rays and spread minimally
  • 5. LiDAR sequence 5 © 2023 Sensor Cortek inc
  • 6. Corresponding camera sequence 6 © 2023 Sensor Cortek inc did you perceive the same things?
  • 7. • It produces direct 3D information • It provides accurate 3D measurements • From under a mm to few cm depending on the distance • It’s an active sensor that operates day and night • It can capture information at long range • ~200 m • It has a large FoV • even 360o What makes LiDAR an attractive sensor? 7 © 2023 Sensor Cortek inc
  • 8. • It captures shape but not appearance • It produces sparse data • Sometimes only few points on an object • It becomes noisy under fog, snow and rain • It is still an expensive sensor • Some are thousands of $ • While radars and cameras could be less than $100 • It often includes mechanical parts However, LiDAR is not perfect 8 © 2023 Sensor Cortek inc
  • 9. • A collection of sparse 3D points • A LiDAR frame • Points not captured at exactly the same time…. The LiDAR point cloud 9 © 2023 Sensor Cortek inc
  • 10. • 1900s: Planck, Einstein and others “discovered” the photon • May 16th 1960: first laser light produced (T. Maiman) • A laser is a device that generates an intense beam of coherent monochromatic light • Light Amplification by Stimulated Emission of Radiation • Laser differs from other light sources because it emits coherent collimated light • Laser light is very narrow, making it possible to see the smallest details with high resolution at relatively long distances. History of Laser 10 © 2023 Sensor Cortek inc
  • 11. • 1960s: LiDAR initially developed for metrology and atmospheric research • But the idea of probing the atmosphere with light can be dated back to the 1930s • 1970s: LiDAR used for terrain mapping • Apollo missions used laser to accurately measure Earth-Moon distances • 1980s: with the advent of GPS and inertial measurement units (IMUs), LiDAR became very accurate • 2005: first AV to complete the DARPA Grand Challenge (142 mile desert course) was equipped with a LiDAR History of LiDAR © 2023 Sensor Cortek inc Mt St-Helens Wikimedia commons 11
  • 12. 1. The LiDAR emits pulsed light waves 2. The light potentially hits a surrounding object and bounces back to the sensor 3. The sensor reads the bounced signal, estimates the time it took to return to the LiDAR and measures the reflected light energy • Simple technology; almost instantaneous! • Potential interference from the sun and other LiDARs How does LiDAR work? Pulsed LiDAR 12 © 2023 Sensor Cortek inc
  • 13. How does LiDAR work? 13 © 2023 Sensor Cortek inc blind region 1. 3. 2. This configuration is called bistatic (most common and cheaper) A monostatic optical system aligns the Tx and Rx for better detection
  • 14. 1. LiDAR can also use continuous waves • Frequency Modulated Continuous Wave 2. The phase of the bounced back signal will differ from the emitted signal 3. The change in phase is used to extract the distance information • This is done by mixing the emitted and received signals (as done in radar) • Velocity is a bonus! • Virtually no interference • You must read a longer signal (stay longer at each point) How does LiDAR work? FMCW Lidar 14 © 2023 Sensor Cortek inc
  • 15. 15 © 2023 Sensor Cortek inc How does LiDAR work? FMCW Lidar www.bridgerphotonics.com/blog/frequency- modulated-continuous-wave-fmcw-lidar
  • 16. • LiDAR physics is governed by one simple equation Distance of the object = (Speed of Light x Time of Flight) / 2 • But to be able to read the received light, you need power (i.e., enough photons bouncing back) Power received ≈ Power transmitted x Cross Section x Optic area Distance2 Distance2 • The laser cross section is the average amount of optical power returned by the target LiDAR mathematics 16 © 2023 Sensor Cortek inc
  • 17. LiDAR sensor taxonomy © 2023 Sensor Cortek inc LiDAR Scanning LiDAR Mechanical Scanners Rotating mirrors Rotating prisms MEMS mirrors Electronic Scanners Optical Phased Array Imaging LiDAR Flash LiDAR Solid-state LiDAR 17
  • 18. • Three strategies: • Scanning LiDAR • A laser scans the scene and a single photodetector is used to read the returned photons • Flash LiDAR • The entire field of view is illuminated and a photodetector array captures the received photons • Optical Phased Arrays • Several transmitters emitting laser light at different phases enabling the steering of the beam (constructive/destructive interference) Capturing a scene with LiDAR 18 © 2023 Sensor Cortek inc
  • 19. • Shorter range – higher frame rate • Costly focal plane array • Limited FoV • Light is distributed across the FoV – more noisy • Pixels are small – more power required • Angular resolution determined by the pixel density • No motion distortion • No moving parts Flash LiDAR 19 © 2023 Sensor Cortek inc
  • 20. • Longer range – smaller frame rate • Expensive scanning mechanism • e.g., spinning mirrors, MEMS mirrors • or rotate everything • Can be bulky • LiDAR motion must be compensated • Less tolerant to mechanical vibrations Scanning LiDAR 20 © 2023 Sensor Cortek inc
  • 21. • Today’s most popular solution • Heavier than other solutions • Vulnerable to vibrations • Generally includes stack of photodetectors to scan in several horizontal layers Scanning with rotating mirrors (or rotate everything) 21 © 2023 Sensor Cortek inc precisionlaserscanning.com/2017/12/mems-mirrors-vs-polygon- scanners-for-lidar-in-autonomous-vehicles/
  • 22. • Uses two (or more) sequential prisms to steer the beam • Shape of prism and speed of rotation determines the scan pattern • Limited FoV Scanning with rotating prisms 22 © 2023 Sensor Cortek inc Lidars for vehicles: from the requirements to the technical evaluation, Z. Dai et al., Conference: 9th International Forum on Automotive Lighting, 2021
  • 23. • Micro-Electro-Mechanical System • Quasi solid-state • Programmable scan patterns • Limited FoV • Requires careful calibration Scanning with MEMS mirrors 23 © 2023 Sensor Cortek inc preciseley.com/product/mems-scanning-mirror/
  • 24. • Lower resolution – high frame rate • Solid state • No moving parts • Smart zooming capability • Interference from the antenna lobes limits the angular resolution • Complex design • Lower production cost • Probably the solution of the future Optical phased array 24 © 2023 Sensor Cortek inc MEMS Mirrors for LiDAR: A Review, D. Wang, C. Watkins, H. Xie. Micromachines 2020, 11(5)
  • 25. Rotating Prisms MEMS OPA Flash Range Long Long Long Medium Low Frame rate Low Low Low High High FoV Large Limited Limited Limited Limited Resolution High High High Adaptive Low Power High High Low Low High Solid Sate No No Quasi Yes Yes Vulnerability High High High Low Low Complexity Low Low High High Low Cost High High High Low High Lidar technologies 25 © 2023 Sensor Cortek inc
  • 26. • LiDAR operates in the near IR spectrum • 780 nm to 3000 nm • Be careful about eye safety! • Typical field of view: • 90°, 180° or 360° • Depth resolution determined by the temporal sampling frequency • ∆D = c / 2f • e.g., a depth resolution of 1 cm requires a 1.5 GHz sampling rate Some LiDAR specs 26 © 2023 Sensor Cortek inc • Angular resolution determined by the scanning point rate • e.g., 0.1° corresponds to 18 cm at 100 m • Pulse frequency • e.g., 2 ns corresponds to a range resolution of 3 cm • Pulse frequency determines the number of points per second per layer • e.g., 40 kHz Specs are usually given for 80% Lambertian reflectivity
  • 27. • When the vehicle moves, the LiDAR is scanning a moving scene • Which will distort the point cloud • This distortion is proportional to the vehicle’s speed and inversely proportional to laser scanning rate • Solution: using an inertial measurement unit in order to compensate for the sensor motion • IMUs provide acceleration and angular velocity • Assumption: the ego vehicle motion has constant angular and linear velocities LiDAR motion compensation using IMU 27 © 2023 Sensor Cortek inc
  • 28. LiDAR motion compensation using IMU 28 © 2023 Sensor Cortek inc https://www.mathworks.com/help/lidar/ug/m otion-compensation-in-lidar-point-cloud.html Before motion compensation After motion compensation
  • 29. • A LiDAR sensor produces a frame of 3D points • A point cloud • This point cloud has to be processed and analyzed in order to interpret the scene • e.g., detect objects on the road • But LiDAR data is sparse and unstructured • Signal processing (convolution) prefers regular grid sampling • LiDAR needs to be preprocessed to build a more suitable representation Processing LiDAR data for detection 29 © 2023 Sensor Cortek inc
  • 30. • Possible representations: • Point sets • Voxelization • Bird’s eye view • Point pillars encoding • Frontal image generation • Sparse convolution LiDAR representation strategies 30 © 2023 Sensor Cortek inc
  • 31. • Point sets • Can we work directly on the point cloud? • Point-based representation • Voxelization • Bird’s eye view • Point pillars encoding • Frontal image generation • Sparse convolution LiDAR representations 31 © 2023 Sensor Cortek inc
  • 32. • The idea is to consume an unordered set of 3D points • Transformations are learned to normalize the data • Global point features are learned from the set • Point cloud has to be segmented into small regions of interest • More difficult to apply in a complex scene composed of many objects Point-based representations 32 © 2023 Sensor Cortek inc C. R. Qi, H. Su, K. Mo, and L. J. Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. Proc. Computer Vision and Pattern Recognition (CVPR) 2017
  • 33. Point-based representations 33 © 2023 Sensor Cortek inc www.mdpi.com/1999-4907/12/2/131/htm • The idea is to consume an unordered set of 3D points
  • 34. • Point sets • Voxelization • e.g., occupancy grid • Bird’s eye view • Point pillars encoding • Frontal image generation • Sparse convolution LiDAR representations 34 © 2023 Sensor Cortek inc Attribute Filtering of Urban Point Clouds Using Max-Tree on Voxel Data, F. Guillotte, Mathematical Morphology and Its Applications to Signal and Image Processing, 2019.
  • 35. • A 3D voxel grid is created in which each voxel contains: • A scalar value • A vector made of statistics computed from the points inside the voxel • Mean, variance, reflectance, … • By nature, the occupied voxels are very sparse • 3D convolution expensive and inefficient Voxelization 35 © 2023 Sensor Cortek inc Wang, D.Z.; Posner, I. Voting for voting in online point cloud object detection. In Proceedings of the Robotics: Science and Systems, Rome, Italy, 13–17 July 2015
  • 36. • Voxel feature encoding (VFE) is used • from randomly sampled points in each voxel • using point coordinates and reflectance • and a fully connected network transformation • 3D convolution is applied on the VFE Voxelization: VoxelNet 36 © 2023 Sensor Cortek inc Y. Zhou and O. Tuzel. Voxelnet: End-to-end learning for point cloud based 3d object detection. In CVPR, 2018 VFE learning Convolution layer Detection network
  • 37. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Vehicle Detection results – KITTI Dataset 37 © 2023 Sensor Cortek inc
  • 38. • Point sets • Voxelization • Bird’s eye view • Point pillars encoding • Frontal image generation • Sparse comvolution LiDAR representations 38 © 2023 Sensor Cortek inc
  • 39. • Computationally efficient • Preserve the metric space for objects on the road • The representation produces an image • Height becomes a channel • Image detection network can be used • Objects are not occluded Bird’s eye view network : Pixor 39 © 2023 Sensor Cortek inc Yang, B., Luo, W., Urtasun, R.: PIXOR: real-time 3D object detection from point clouds. CVPR (2018)
  • 40. Vehicle Detection results – KITTI Dataset 40 © 2023 Sensor Cortek inc Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95
  • 41. • Voxelization • Bird’s eye view • Pillars encoding • Each pillar encodes point distance to centroid and reflectance • Simplified PointNet is used • Frontal image generation • Sparse convolution LiDAR representations 41 © 2023 Sensor Cortek inc becominghuman.ai/pointpillars-3d-point-clouds-bounding-box- detection-and-tracking-pointnet-pointnet-lasernet-67e26116de5a
  • 42. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95 PointPillar 92.07 87.74 86.65 Vehicle Detection results – KITTI Dataset 42 © 2023 Sensor Cortek inc
  • 43. • Voxelization • Bird’s eye view • Point pillars encoding • Frontal image generation • Sparse convolution LiDAR representations 43 © 2023 Sensor Cortek inc
  • 44. • BEV + LiDAR project + Camera • BEV is used to propose potential objects • Multiview features are then uses to predict objects Object-based feature extractor: MV3D 44 © 2023 Sensor Cortek inc Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3d object detection network for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017
  • 45. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95 PointPillar 92.07 87.74 86.65 MV3D 86.55 78.1 76.67 Vehicle Detection results – KITTI Dataset 45 © 2023 Sensor Cortek inc
  • 46. Frontal view fusion 46 © 2023 Sensor Cortek inc • All images are fused together • Including the camera image • Lidar/camera fusion • BGF fusion operator Sensor Fusion Operators for Multimodal 2D Object Detection, M.M. Pasandi, T. Liu, Y. Massoud, R, Laganiere, ISVC 2022
  • 47. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95 PointPillar 92.07 87.74 86.65 MV3D 86.55 78.1 76.67 BGF Fusion 94.90 88.40 78.38 Vehicle Detection results – KITTI Dataset 47 © 2023 Sensor Cortek inc
  • 48. • Voxelization • Bird’s eye view • Pillars encoding • Frontal image generation • Sparse convolution • When the data is very sparse, regular convolution becomes very inefficient • The idea is to compress the representation by ignoring zero values • To this end, look-up tables are often used LiDAR representations 48 © 2023 Sensor Cortek inc
  • 49. SECOND: Sparse convolution on point features 49 © 2023 Sensor Cortek inc SECOND: Sparsely Embedded Convolutional Detection by Yan Yan, Yuxing Mao, Bo Li, Sensors 2018.
  • 50. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95 PointPillar 92.07 87.74 86.65 MV3D 86.55 78.1 76.67 BGF Fusion 94.90 88.40 78.38 SECOND 91.92 87.92 85.39 Vehicle Detection results – KITTI Dataset 50 © 2023 Sensor Cortek inc
  • 51. PV-RCNN: Point-based + Voxels 51 © 2023 Sensor Cortek inc PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection (2020) Shaoshuai Shi, Chaoxu Guo, Li Jiang, Zhe Wang, Jianping Shi, Xiaogang Wang, Hongsheng Li, IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
  • 52. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95 PointPillar 92.07 87.74 86.65 MV3D 86.55 78.1 76.67 BGF Fusion 94.90 88.40 78.38 SECOND 91.92 87.92 85.39 PV-RCNN 92.86 88.93 88.74 Vehicle Detection results – KITTI Dataset 52 © 2023 Sensor Cortek inc
  • 53. • Solving the LiDAR sparsity problem • The farther the object, the sparser the point density How to further improve detection? 53 © 2023 Sensor Cortek inc
  • 54. • Depth completion network • Can be used to densify the point cloud Pseudo LiDAR 54 © 2023 Sensor Cortek inc www.mdpi.com/1424-8220/22/18/6969 Depth Completion with Twin Surface Extrapolation at Occlusion Boundaries, Saif Imran, Xiaoming Liu, Daniel Morris, CVPR 2021
  • 55. BTC: using shape completion 55 © 2023 Sensor Cortek inc Behind the Curtain: Learning Occluded Shapes for 3D Object Detection, Qiangeng Xu, Yiqi Zhong, Ulrich Neumann, AAAI 2022
  • 56. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95 PointPillar 92.07 87.74 86.65 MV3D 86.55 78.1 76.67 BGF Fusion 94.90 88.40 78.38 SECOND 91.92 87.92 85.39 PV-RCNN 92.86 88.93 88.74 BTC 93.46 89.53 87.44 Vehicle Detection results – KITTI Dataset 56 © 2023 Sensor Cortek inc
  • 57. LiDAR and pseudo-LiDAR: SFD 57 © 2023 Sensor Cortek inc • Lidar/camera fusion Sparse Fuse Dense: Towards High Quality 3D Detection with Depth Completion, Wu, Xiaopei and Peng, Liang and Yang, Honghui and Xie, Liang and Huang, Chenxi and Deng, Chengqi and Liu, Haifeng and Cai, Deng}, 2022
  • 58. Mean Average Precision (%) Easy Medium Hard VoxelNet 89.35 79.26 77.39 Pixor 81.7 77.05 72.95 PointPillar 92.07 87.74 86.65 MV3D 86.55 78.1 76.67 BGF Fusion 94.90 88.40 78.38 SECOND 91.92 87.92 85.39 PV-RCNN 92.86 88.93 88.74 BTC 93.46 89.53 87.44 SFD 95.85 91.92 91.41 58 © 2023 Sensor Cortek inc Vehicle Detection results – KITTI Dataset
  • 59. LiDAR densification: UYI 59 © 2023 Sensor Cortek inc • Pseudo LiDAR module Use Your Imagination: A Detector-Independent Approach For LiDAR Quality Booster, Z. Zhang, T. Liu, R. Laganiere, 2023 LiDAR detector network
  • 60. LiDAR densification: UYI 60 © 2023 Sensor Cortek inc • Vehicle detection performance boosting Use Your Imagination: A Detector-Independent Approach For LiDAR Quality Booster, Z. Zhang, T. Liu, R. Laganiere, 2023
  • 61. Conclusion 61 © 2023 Sensor Cortek inc • LiDAR provides accurate 3D detection • An essential component in ADAS/AV • LiDAR technology will continue to evolve • Improve density • Lower power • Solid-state • Lower cost • Radar / LiDAR Convergence
  • 62. • C. R. Qi, H. Su, K. Mo, and L. J. Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. CVPR 2017. • Yang, B., Luo, W., Urtasun, R.: PIXOR: real-time 3D object detection from point clouds. CVPR 2018. • X. Chen, H. Ma, J. Wan, B. Li, T. Xia, Multi-view 3d object detection network for autonomous driving. CVPR 2017. • Yan Yan, Yuxing Mao, Bo Li. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018. • S. Shi, C. Guo, L. Jiang, Z. Wang, J. Shi, X. Wang, H. Li. PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection. CVPR 2020. • X. Wu, L. Peng, H. Yang, L. Xie, C. Huang, C. Deng, H. Liu, D. Cai. Sparse Fuse Dense: Towards High Quality 3D Detection with Depth Completion, CVPR 2022. • Q. Xu, Y. Zhong, U. Neumann. Behind the Curtain: Learning Occluded Shapes for 3D Object Detection. AAAI 2022. References 62 © 2023 Sensor Cortek inc Reading list on LiDAR and AI