SlideShare a Scribd company logo
AIRBORNE LIDAR POINT DENSITY, MORE TO THE POINT
July 23, 2019
Matt Bethel
Director of Operations and Technology
Merrick & Company
How Is Airborne LiDAR Density Measured?
0.5
meter
0.5
meter
0.5 meter
ground sample
distance (GSD) or
nominal point
spacing (NPS)
How Is Airborne LiDAR Density Measured?
Type equation here.
1 meter
1 meter
4 points per
square meter
(PPSM)
=
𝐍𝐏𝐒 =
𝟏
𝑫𝒆𝒏𝒔𝒊𝒕𝒚
𝑫𝒆𝒏𝒔𝒊𝒕𝒚 =
𝟏
𝑵𝑷𝑺 𝟐0.5 meter
ground sample
distance (GSD) or
nominal point
spacing (NPS)
𝑫𝒆𝒏𝒔𝒊𝒕𝒚 =
𝐟𝐢𝐫𝐬𝐭 𝐨𝐫 𝐥𝐚𝐬𝐭 𝐫𝐞𝐭𝐮𝐫𝐧 𝐩𝐨𝐢𝐧𝐭 𝐜𝐨𝐮𝐧𝐭
𝒂𝒓𝒆𝒂
How Is Airborne LiDAR Density Measured?
1 meter
1 meter
9 or maybe 1
but not 4 PPSM
=
4 points per
square meter
(PPSM)
0.5 meter
ground sample
distance (GSD) or
nominal point
spacing (NPS)
How Is Airborne LiDAR Density Measured?
confusing=
Q: Are points on cell
boundaries shared for density
calculations?
A: No, they must be counted
within only one cell.
REMEMBER
THIS!!!
0.5 meter
ground sample
distance (GSD) or
nominal point
spacing (NPS)
Ways of Measuring Airborne LiDAR Density
• Representative samples
Pros
• Fast and easy to calculate
• Good for areas of interests
Cons
• Biased by many factors such
as sidelap, patches,
turbulence, etc.
• Very localized, not
representative of swaths or
project extents
• Cannot automatedly find
problem areas that could be
considered failures /
specification violations
• Difficult to use for reporting
Ways of Measuring Airborne LiDAR Density
• Representative samples
• Per swath
Pros
• Ideal to compare against
planned swath density
• Relatively easy to compute
• Reasonably batchable – one
process per flightline
• Good to use for reporting
• Is not biased (inflated) by sidelap
• No hidden problems with use or
reporting – very straightforward
Cons
• Needs interpretation if flying
>50% sidelap to achieve planned
density
Ways of Measuring Airborne LiDAR Density
• Representative samples
• Per swath
• Aggregate / project
wide
Pros
• Considers all collected points
• Good for reporting
• Straightforward approach
(number of first or last return
points / area of project boundary)
Cons
• Crosslines, sidelap, collection
block overlap, and patches can
inflate density results
• Tabular reporting only will not
identify localized density failures.
A thematic raster is needed for
locating localized density failures.
Area of Project Boundary (m2)
Ways of Measuring Airborne LiDAR Density
• Representative samples
• Per swath
• Aggregate / project wide
• Voronoi / Thiessen
polygon
Pros
• Most accurate representation of point
density
• Measurement is an area of point
influence
• Density can be derived by 1/Voronoi
area
• Unlike grids, polygons share
representative edges with neighboring
polygons. With grid cells, four of the
eight neighboring cells are connected
through only a point.
Cons
• Cumbersome to work with
• If used, typically only in representative
sample areas or around check points -
not project wide
Ways of Measuring Airborne LiDAR Density
• Representative samples
• Per swath
• Aggregate / project wide
• Voronoi / Thiessen
polygon
Square
meters
Pros
• Most accurate representation of point
density
• Measurement is an area of point
influence
• Density can be derived by 1/Voronoi
area
• Unlike grids, polygons share
representative edges with neighboring
polygons. With grid cells, four of the
eight neighboring cells are connected
through only a point.
Cons
• Cumbersome to work with
• If used, typically only in representative
sample areas or around check points -
not project wide
Ways of Measuring Airborne LiDAR Density
• Representative samples
• Per swath
• Aggregate / project wide
• Voronoi / Thiessen
polygon
PPSM
Pros
• Most accurate representation of point
density
• Measurement is an area of point
influence
• Density can be derived by 1/Voronoi
area
• Unlike grids, polygons share
representative edges with neighboring
polygons. With grid cells, four of the
eight neighboring cells are connected
through only a point.
Cons
• Cumbersome to work with
• If used, typically only in representative
sample areas or around check points -
not project wide
Ways of Measuring Airborne LiDAR Density
• Representative samples
• Per swath
• Aggregate / project wide
• Voronoi / Thiessen
polygon
• Grid / raster / tile
Pros
• Fast and easy to calculate
• Seemingly straightforward approach – use grid or tile scheme to count
points and report on lowest divisible points per grid/tile area
• Easy to use for reporting – pass fail percentage results and graphic
Cons
• The results are in pass/fail cell counts yet there are no establish
parameters for use or analysis (no passing thresholds)
• User selected processing cell size changes the results
• Inherent with major problems
• Results are severely misunderstood yet widely used and relied upon
There is a better way…
• In digital signal processing, the minimum sampling rate is limited to 2X the
maximum frequency. The purpose for this limit is to preserve important
information throughout the transformation. This is known as the Nyquist-
Shannon sampling theorem. This 2X law can be applied to LiDAR density
measurement and derivative grid creation. This is known as the Nyquist
sampling criteria.
• The Nyquist sampling criteria states that we must sample at no less than twice
the resolution of the smallest detail we intend to measure or model.
• This means that grids used for density calculation or raster product creation
must have a cell size no less than 2 x NPS.
Third Party Reports Showing
Varying Pass/Fail Percentage Results
These first return density reports were generated using third party software ran on two sample swaths using
different cell sizes (highlighted in each screenshot). Note the percentages widely vary with each cell size test.
That was all real LiDAR data with
random point spacing.
Let’s test synthetically created,
perfectly spaced point data.
If we take three LiDAR swaths
Export to an LAS grid file at exactly 2
PPSM / 0.7071067811865470 GSD
Then test and report on density using the
grid method.
We expect 100% passing of all tests.
If we take three LiDAR swaths
Export to an LAS grid file at exactly 2
PPSM / 0.7071067811865470 meter GSD
Then test and report on density using the
grid method.
We expect 100% passing of all tests.
What is going on here?
Less than Ideal Scan Pattern
Cross track point spacing = 0.5 m
Along track point spacing = 1 m
Parallel scanline pattern (Riegl)
(Minimum required point count per cell = 2)
190 cells contain 2 points
10 cells contain only 1 point
95% pass / 5% fail this test
Assume each cell is 1 m X 1 m (1 m2)
40 points across by 10 scan lines = 400 points / 200 cells = 2 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: Data with poor point spacing geometry can yield a high passing percentage “grade” using the grid
density test method.
Less than Ideal Scan Pattern
Cross track point spacing = 0.5 m
Along track point spacing = 1 m
Zigzag scanline pattern (Optech)
(Minimum required point count per cell = 2)
190 cells contain 2 points
10 cells contain only 1 point
95% pass / 5% fail this test
Assume each cell is 1 m X 1 m (1 m2)
40 points across by 10 scan lines = 400 points / 200 cells = 2 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: Data with poor point spacing geometry can yield a high passing percentage “grade” using the grid
density test method. Zigzag scanline patterns result in the same “grade” as parallel line scanning patterns.
Ideal Scan Pattern
Cross track point spacing = 0.71 m
Along track point spacing = 0.71 m
Parallel scanline pattern (Riegl)
(Minimum required point count per cell = 2)
100 cells contain 2 points
60 cells contain only 1 point
40 Cells contain more than 2 points
70% pass / 30% fail this test using 20 more points than the previous
example and a much improved point spacing distribution
Assume each cell is 1 m X 1 m (1 m2)
28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: Data with ideal point spacing geometry results in a much lower passing percentage “grade” using the grid density test method.
This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
Ideal Scan Pattern
Cross track point spacing = 0.71 m
Along track point spacing (at nadir) = 0.71 m
Zigzag scanline pattern (Optech)
(Minimum required point count per cell = 2)
98 cells contain 2 points
60 cells contain only 1 point
42 Cells contain more than 2 points
70% pass / 30% fail this test using 20 more points than the previous
example and a much improved point spacing distribution
Assume each cell is 1 m X 1 m (1 m2)
28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: Data with ideal point spacing geometry results in a much lower passing percentage “grade” using the grid density test method. Zigzag scanline patterns result in the same “grade” as
parallel line scanning patterns. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
What happens when we increase
the size of the testing tile?
Ideal Scan Pattern
Cross track point spacing = 0.71 m
Along track point spacing = 0.71 m
Parallel scanline pattern (Riegl)
Cell size increased to 2m x 2m (Minimum required point count per cell = 8)
0 cells contain 8 points
10 cells contain less than 8 points
40 Cells contain more than 8 points
80% pass / 20% fail this test using 20 more points than the previous
example and a much improved point spacing distribution
Assume each cell is 2 m X 2 m (4 m2)
28 points across by 15 scan lines = 420 points / 50 cells / 4 m2 = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which
invalidates the results. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
Ideal Scan Pattern
Cross track point spacing = 0.71 m
Along track point spacing (at nadir) = 0.71 m
Zigzag scanline pattern (Optech)
Cell size increased to 2m x 2m (Minimum required point count per cell = 8)
0 cells contain 8 points
10 cells contain less than 8 points
40 Cells contain more than 8 points
80% pass / 20% fail this test using 20 more points than the previous
example and a much improved point spacing distribution
Assume each cell is 2 m X 2 m (4 m2)
28 points across by 15 scan lines = 420 points / 50 cells / 4 m2 = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which invalidates the results. Zigzag
scanline patterns result in the same “grade” as parallel line scanning patterns. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a
PASS/FAIL percentage “grade”.
Ideal Scan Pattern
Cross track point spacing = 0.71 m
Along track point spacing = 0.71 m
Parallel scanline pattern (Riegl)
Cell size increased to 5m x 5m (Minimum required point count per cell = 50)
0 cells contain 50 points
4 cells contain less than 50 points
4 Cells contain more than 50 points
50% pass / 50% fail this test using 20 more points than the previous
example and a much improved point spacing distribution
Assume each cell is 5 m X 5 m (25 m2)
28 points across by 15 scan lines = 420 points / 8 cells / 25 m2 = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which
invalidates the results. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
Ideal Scan Pattern
Cross track point spacing = 0.71 m
Along track point spacing (at nadir) = 0.71 m
Zigzag scanline pattern (Optech)
Cell size increased 5m x 5m (Minimum required point count per cell = 50)
0 cells contain 50 points
3 cells contain less than 50 points
5 Cells contain more than 50 points
60% pass / 40% fail this test using 20 more points than the previous
example and a much improved point spacing distribution
Assume each cell is 5 m X 5 m (25 m2)
28 points across by 15 scan lines = 420 points / 8 cells / 25 m2 = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which invalidates the results.
Zigzag scanline patterns result in a similar “grade” as parallel line scanning patterns. This method is severely flawed for analyzing and reporting LiDAR point density, especially as
a PASS/FAIL percentage “grade”.
Ideal Scan Pattern With Nyquist Sampling Criteria
Cross track point spacing = 0.71 m
Along track point spacing = 0.71 m
Parallel scanline pattern (Riegl)
(Minimum required point count per cell = 1)
420 Cells contain more than 1 point
100% pass / 0% fail this test using
Assume each cell is 1.42m X 1.42 m (2.02 m2)
28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Takeaway: Data with ideal point spacing geometry results in a 100% passing percentage “grade” using the Nyquist sampling criteria grid
density test method.
Ideal Scan Pattern With Nyquist Sampling Criteria
Cross track point spacing = 0.71 m
Along track point spacing (at nadir) = 0.71 m
Zigzag scanline pattern (Optech)
(Minimum required point count per cell = 1)
Takeaway: Data with ideal point spacing geometry results in a 100% passing percentage “grade” using the Nyquist sampling criteria grid density
test method. Zigzag scanline patterns result in the same “grade” as parallel line scanning patterns.
420 Cells contain more than 1 point
100% pass / 0% fail this test using
Assume each cell is 1.42m X 1.42 m (2.02 m2)
28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm
This passes the USGS LBS v1.x NPS/density requirement
Test
Cross track
spacing (m)
Along track
spacing (m)
at nadir
Raster cell
size (m)
Minimum
required
point count
per cell
Number of
points in
test
Percentage
of passing
cells
Percentage
of failing
cells
Test
Cross track
spacing (m)
Along track
spacing (m)
at nadir
Raster cell
size (m)
Minimum
required
point count
per cell
Number of
points in
test
Percentage
of passing
cells
Percentage
of failing
cells
Parallel scan pattern 0.5 1 1 2 400 95% 5%
Zigzag scan pattern 0.5 1 1 2 400 95% 5%
Test
Cross track
spacing (m)
Along track
spacing (m)
at nadir
Raster cell
size (m)
Minimum
required
point count
per cell
Number of
points in
test
Percentage
of passing
cells
Percentage
of failing
cells
Parallel scan pattern 0.5 1 1 2 400 95% 5%
Zigzag scan pattern 0.5 1 1 2 400 95% 5%
Parallel scan pattern 0.71 0.71 1 2 420 70% 30%
Zigzag scan pattern 0.71 0.71 1 2 420 70% 30%
Test
Cross track
spacing (m)
Along track
spacing (m)
at nadir
Raster cell
size (m)
Minimum
required
point count
per cell
Number of
points in
test
Percentage
of passing
cells
Percentage
of failing
cells
Parallel scan pattern 0.5 1 1 2 400 95% 5%
Zigzag scan pattern 0.5 1 1 2 400 95% 5%
Parallel scan pattern 0.71 0.71 1 2 420 70% 30%
Zigzag scan pattern 0.71 0.71 1 2 420 70% 30%
Parallel scan pattern 0.71 0.71 2 8 420 80% 20%
Zigzag scan pattern 0.71 0.71 2 8 420 80% 20%
Test
Cross track
spacing (m)
Along track
spacing (m)
at nadir
Raster cell
size (m)
Minimum
required
point count
per cell
Number of
points in
test
Percentage
of passing
cells
Percentage
of failing
cells
Parallel scan pattern 0.5 1 1 2 400 95% 5%
Zigzag scan pattern 0.5 1 1 2 400 95% 5%
Parallel scan pattern 0.71 0.71 1 2 420 70% 30%
Zigzag scan pattern 0.71 0.71 1 2 420 70% 30%
Parallel scan pattern 0.71 0.71 2 8 420 80% 20%
Zigzag scan pattern 0.71 0.71 2 8 420 80% 20%
Parallel scan pattern 0.71 0.71 5 50 420 50% 50%
Zigzag scan pattern 0.71 0.71 5 50 420 60% 40%
Test
Cross track
spacing (m)
Along track
spacing (m)
at nadir
Raster cell
size (m)
Minimum
required
point count
per cell
Number of
points in
test
Percentage
of passing
cells
Percentage
of failing
cells
Parallel scan pattern 0.5 1 1 2 400 95% 5%
Zigzag scan pattern 0.5 1 1 2 400 95% 5%
Parallel scan pattern 0.71 0.71 1 2 420 70% 30%
Zigzag scan pattern 0.71 0.71 1 2 420 70% 30%
Parallel scan pattern 0.71 0.71 2 8 420 80% 20%
Zigzag scan pattern 0.71 0.71 2 8 420 80% 20%
Parallel scan pattern 0.71 0.71 5 50 420 50% 50%
Zigzag scan pattern 0.71 0.71 5 50 420 60% 40%
Parallel scan pattern 0.71 0.71 1.42 1 420 100% 0%
Zigzag scan pattern 0.71 0.71 1.42 1 420 100% 0%
Conclusions and Recommendations
• Representative samples are too limiting for project analysis and reporting.
• Swath density analysis is straightforward, reliable, well understood, and very representative.
• Aggregate is too generalizing. A supplemental raster is required to identify localized failures.
• Voronoi is the most accurate but too cumbersome to use for thorough project analysis.
• Grid/raster/tile density method is very effective only if using the Nyquist sampling criteria. A
qualifying pass/fail threshold is required. Also, the results of this method are consistent with
the results from the reliable swath density analysis.
• Never use the flawed grid method with a simple points per square area calculation.
Thank You
Matt Bethel
Director of Operations and Technology
Merrick & Company
http://www.merrick.com/Geospatial
matt.bethel@merrick.com
(303) 353-3662

More Related Content

What's hot

An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...
An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...
An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...
3D Laser Mapping
 
Rankin LiDAR presentation
Rankin LiDAR presentationRankin LiDAR presentation
Rankin LiDAR presentationJustin Farrow
 
Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)
NopphawanTamkuan
 
Lidar
LidarLidar
Introduction to LiDAR
Introduction to LiDARIntroduction to LiDAR
Introduction to LiDAR
Bob Champoux
 
Object Based Image Analysis
Object Based Image Analysis Object Based Image Analysis
Object Based Image Analysis
Kabir Uddin
 
GIS Standards and Interoperability
GIS Standards and InteroperabilityGIS Standards and Interoperability
GIS Standards and InteroperabilityNasr Khashoggi
 
Digital image classification22oct
Digital image classification22octDigital image classification22oct
Digital image classification22oct
Aleemuddin Abbasi
 
How to better understand SAR, interpret SAR products and realize the limitations
How to better understand SAR, interpret SAR products and realize the limitationsHow to better understand SAR, interpret SAR products and realize the limitations
How to better understand SAR, interpret SAR products and realize the limitations
NopphawanTamkuan
 
SBAS-DInSAR processing on the ESA Geohazards Exploitation Platform
SBAS-DInSAR processing on the ESA Geohazards Exploitation PlatformSBAS-DInSAR processing on the ESA Geohazards Exploitation Platform
SBAS-DInSAR processing on the ESA Geohazards Exploitation Platform
Emmanuel Mathot
 
Synthetic Aperture Radar (SAR) Basics.pptx
Synthetic Aperture Radar (SAR) Basics.pptxSynthetic Aperture Radar (SAR) Basics.pptx
Synthetic Aperture Radar (SAR) Basics.pptx
sapna kinattinkara
 
Introduction to sar-marjolaine_rouault
Introduction to sar-marjolaine_rouaultIntroduction to sar-marjolaine_rouault
Introduction to sar-marjolaine_rouault
Naivedya Mishra
 
Utilização de imagens de satélites em diferentes aplicações - Casos Reais
Utilização de imagens de satélites em diferentes aplicações - Casos ReaisUtilização de imagens de satélites em diferentes aplicações - Casos Reais
Utilização de imagens de satélites em diferentes aplicações - Casos ReaisGlobalGeo Geotecnologias
 
Introduction to LiDAR presentation.
Introduction to LiDAR presentation.Introduction to LiDAR presentation.
Introduction to LiDAR presentation.
Bob Champoux
 
Landsat.pptx
Landsat.pptxLandsat.pptx
Landsat.pptx
Reshma Jeswani
 
Spectrum Overview November 2010
Spectrum Overview November 2010Spectrum Overview November 2010
Spectrum Overview November 2010
keithlarose
 
GIS
GISGIS
An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...
An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...
An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...
3D Laser Mapping
 
Terrestrial laser scanning
Terrestrial laser scanningTerrestrial laser scanning
Terrestrial laser scanning
Iris Kramer
 

What's hot (20)

An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...
An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...
An Introduction to Laser Scanning - Part 3: Mobile mapping and accuracy chall...
 
Rankin LiDAR presentation
Rankin LiDAR presentationRankin LiDAR presentation
Rankin LiDAR presentation
 
Wide field sensor
Wide field sensorWide field sensor
Wide field sensor
 
Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)Introduction to Synthetic Aperture Radar (SAR)
Introduction to Synthetic Aperture Radar (SAR)
 
Lidar
LidarLidar
Lidar
 
Introduction to LiDAR
Introduction to LiDARIntroduction to LiDAR
Introduction to LiDAR
 
Object Based Image Analysis
Object Based Image Analysis Object Based Image Analysis
Object Based Image Analysis
 
GIS Standards and Interoperability
GIS Standards and InteroperabilityGIS Standards and Interoperability
GIS Standards and Interoperability
 
Digital image classification22oct
Digital image classification22octDigital image classification22oct
Digital image classification22oct
 
How to better understand SAR, interpret SAR products and realize the limitations
How to better understand SAR, interpret SAR products and realize the limitationsHow to better understand SAR, interpret SAR products and realize the limitations
How to better understand SAR, interpret SAR products and realize the limitations
 
SBAS-DInSAR processing on the ESA Geohazards Exploitation Platform
SBAS-DInSAR processing on the ESA Geohazards Exploitation PlatformSBAS-DInSAR processing on the ESA Geohazards Exploitation Platform
SBAS-DInSAR processing on the ESA Geohazards Exploitation Platform
 
Synthetic Aperture Radar (SAR) Basics.pptx
Synthetic Aperture Radar (SAR) Basics.pptxSynthetic Aperture Radar (SAR) Basics.pptx
Synthetic Aperture Radar (SAR) Basics.pptx
 
Introduction to sar-marjolaine_rouault
Introduction to sar-marjolaine_rouaultIntroduction to sar-marjolaine_rouault
Introduction to sar-marjolaine_rouault
 
Utilização de imagens de satélites em diferentes aplicações - Casos Reais
Utilização de imagens de satélites em diferentes aplicações - Casos ReaisUtilização de imagens de satélites em diferentes aplicações - Casos Reais
Utilização de imagens de satélites em diferentes aplicações - Casos Reais
 
Introduction to LiDAR presentation.
Introduction to LiDAR presentation.Introduction to LiDAR presentation.
Introduction to LiDAR presentation.
 
Landsat.pptx
Landsat.pptxLandsat.pptx
Landsat.pptx
 
Spectrum Overview November 2010
Spectrum Overview November 2010Spectrum Overview November 2010
Spectrum Overview November 2010
 
GIS
GISGIS
GIS
 
An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...
An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...
An Introduction to Laser Scanning - Part 2: Airborne LiDAR, the basics of tra...
 
Terrestrial laser scanning
Terrestrial laser scanningTerrestrial laser scanning
Terrestrial laser scanning
 

Similar to AIRBORNE LIDAR POINT DENSITY

AIRBORNE LIDAR POINT DENSITY
AIRBORNE LIDAR POINT DENSITYAIRBORNE LIDAR POINT DENSITY
AIRBORNE LIDAR POINT DENSITY
MattBethel1
 
ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...
ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...
ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...
MattBethel1
 
2023_ASPRS_LiDAR_Division_Update.pdf
2023_ASPRS_LiDAR_Division_Update.pdf2023_ASPRS_LiDAR_Division_Update.pdf
2023_ASPRS_LiDAR_Division_Update.pdf
MattBethel1
 
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Sathishkumar Samiappan
 
Sampling-SDM2012_Jun
Sampling-SDM2012_JunSampling-SDM2012_Jun
Sampling-SDM2012_Jun
MDO_Lab
 
Introduction to sampling
Introduction to samplingIntroduction to sampling
Introduction to samplingSituo Liu
 
Measurement Procedures for Design and Enforcement of Harm Claim Thresholds
Measurement Procedures for Design and Enforcement of Harm Claim ThresholdsMeasurement Procedures for Design and Enforcement of Harm Claim Thresholds
Measurement Procedures for Design and Enforcement of Harm Claim Thresholds
Pierre de Vries
 
Chapter 6 image quality in ct
Chapter 6 image quality in ct Chapter 6 image quality in ct
Chapter 6 image quality in ct
Muntaser S.Ahmad
 
Adaptive Geographical Search in Networks
Adaptive Geographical Search in NetworksAdaptive Geographical Search in Networks
Adaptive Geographical Search in Networks
Andrea Wiggins
 
Automatic Visualization - Leland Wilkinson, Chief Scientist, H2O.ai
Automatic Visualization - Leland Wilkinson, Chief Scientist, H2O.aiAutomatic Visualization - Leland Wilkinson, Chief Scientist, H2O.ai
Automatic Visualization - Leland Wilkinson, Chief Scientist, H2O.ai
Sri Ambati
 
Disparity Estimation Using A Color Segmentation V3
Disparity Estimation Using A Color Segmentation V3Disparity Estimation Using A Color Segmentation V3
Disparity Estimation Using A Color Segmentation V3
thomaswangxin
 
TERN Surveillance Training 2019 - Day 1, Session 3
TERN Surveillance Training 2019 - Day 1, Session 3TERN Surveillance Training 2019 - Day 1, Session 3
TERN Surveillance Training 2019 - Day 1, Session 3
bensparrowau
 
Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.
Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.
Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.
Wuhyun Rico Shin
 
Clustering
ClusteringClustering
Clustering
Smrutiranjan Sahu
 
Outlier Detection.pptx
Outlier Detection.pptxOutlier Detection.pptx
Outlier Detection.pptx
AtulSVivek2
 
PEMF2_SDM_2012_Ali
PEMF2_SDM_2012_AliPEMF2_SDM_2012_Ali
PEMF2_SDM_2012_Ali
MDO_Lab
 
07 dimensionality reduction
07 dimensionality reduction07 dimensionality reduction
07 dimensionality reduction
Marco Quartulli
 
Bf4102414417
Bf4102414417Bf4102414417
Bf4102414417
IJERA Editor
 
-Particle-Size-Analysis-pptx.pptx
-Particle-Size-Analysis-pptx.pptx-Particle-Size-Analysis-pptx.pptx
-Particle-Size-Analysis-pptx.pptx
prashantingole13
 

Similar to AIRBORNE LIDAR POINT DENSITY (20)

AIRBORNE LIDAR POINT DENSITY
AIRBORNE LIDAR POINT DENSITYAIRBORNE LIDAR POINT DENSITY
AIRBORNE LIDAR POINT DENSITY
 
ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...
ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...
ASPRS LiDAR Division Update with a focus on quantifying horizontal sampling d...
 
2023_ASPRS_LiDAR_Division_Update.pdf
2023_ASPRS_LiDAR_Division_Update.pdf2023_ASPRS_LiDAR_Division_Update.pdf
2023_ASPRS_LiDAR_Division_Update.pdf
 
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
 
Sampling-SDM2012_Jun
Sampling-SDM2012_JunSampling-SDM2012_Jun
Sampling-SDM2012_Jun
 
Introduction to sampling
Introduction to samplingIntroduction to sampling
Introduction to sampling
 
Measurement Procedures for Design and Enforcement of Harm Claim Thresholds
Measurement Procedures for Design and Enforcement of Harm Claim ThresholdsMeasurement Procedures for Design and Enforcement of Harm Claim Thresholds
Measurement Procedures for Design and Enforcement of Harm Claim Thresholds
 
LiDAR_Project
LiDAR_ProjectLiDAR_Project
LiDAR_Project
 
Chapter 6 image quality in ct
Chapter 6 image quality in ct Chapter 6 image quality in ct
Chapter 6 image quality in ct
 
Adaptive Geographical Search in Networks
Adaptive Geographical Search in NetworksAdaptive Geographical Search in Networks
Adaptive Geographical Search in Networks
 
Automatic Visualization - Leland Wilkinson, Chief Scientist, H2O.ai
Automatic Visualization - Leland Wilkinson, Chief Scientist, H2O.aiAutomatic Visualization - Leland Wilkinson, Chief Scientist, H2O.ai
Automatic Visualization - Leland Wilkinson, Chief Scientist, H2O.ai
 
Disparity Estimation Using A Color Segmentation V3
Disparity Estimation Using A Color Segmentation V3Disparity Estimation Using A Color Segmentation V3
Disparity Estimation Using A Color Segmentation V3
 
TERN Surveillance Training 2019 - Day 1, Session 3
TERN Surveillance Training 2019 - Day 1, Session 3TERN Surveillance Training 2019 - Day 1, Session 3
TERN Surveillance Training 2019 - Day 1, Session 3
 
Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.
Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.
Paper review: Measuring the Intrinsic Dimension of Objective Landscapes.
 
Clustering
ClusteringClustering
Clustering
 
Outlier Detection.pptx
Outlier Detection.pptxOutlier Detection.pptx
Outlier Detection.pptx
 
PEMF2_SDM_2012_Ali
PEMF2_SDM_2012_AliPEMF2_SDM_2012_Ali
PEMF2_SDM_2012_Ali
 
07 dimensionality reduction
07 dimensionality reduction07 dimensionality reduction
07 dimensionality reduction
 
Bf4102414417
Bf4102414417Bf4102414417
Bf4102414417
 
-Particle-Size-Analysis-pptx.pptx
-Particle-Size-Analysis-pptx.pptx-Particle-Size-Analysis-pptx.pptx
-Particle-Size-Analysis-pptx.pptx
 

Recently uploaded

Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...
Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...
Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...
OECD Directorate for Financial and Enterprise Affairs
 
Announcement of 18th IEEE International Conference on Software Testing, Verif...
Announcement of 18th IEEE International Conference on Software Testing, Verif...Announcement of 18th IEEE International Conference on Software Testing, Verif...
Announcement of 18th IEEE International Conference on Software Testing, Verif...
Sebastiano Panichella
 
Doctoral Symposium at the 17th IEEE International Conference on Software Test...
Doctoral Symposium at the 17th IEEE International Conference on Software Test...Doctoral Symposium at the 17th IEEE International Conference on Software Test...
Doctoral Symposium at the 17th IEEE International Conference on Software Test...
Sebastiano Panichella
 
Acorn Recovery: Restore IT infra within minutes
Acorn Recovery: Restore IT infra within minutesAcorn Recovery: Restore IT infra within minutes
Acorn Recovery: Restore IT infra within minutes
IP ServerOne
 
International Workshop on Artificial Intelligence in Software Testing
International Workshop on Artificial Intelligence in Software TestingInternational Workshop on Artificial Intelligence in Software Testing
International Workshop on Artificial Intelligence in Software Testing
Sebastiano Panichella
 
Bitcoin Lightning wallet and tic-tac-toe game XOXO
Bitcoin Lightning wallet and tic-tac-toe game XOXOBitcoin Lightning wallet and tic-tac-toe game XOXO
Bitcoin Lightning wallet and tic-tac-toe game XOXO
Matjaž Lipuš
 
0x01 - Newton's Third Law: Static vs. Dynamic Abusers
0x01 - Newton's Third Law:  Static vs. Dynamic Abusers0x01 - Newton's Third Law:  Static vs. Dynamic Abusers
0x01 - Newton's Third Law: Static vs. Dynamic Abusers
OWASP Beja
 
Bonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdf
Bonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdfBonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdf
Bonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdf
khadija278284
 
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
Orkestra
 
Eureka, I found it! - Special Libraries Association 2021 Presentation
Eureka, I found it! - Special Libraries Association 2021 PresentationEureka, I found it! - Special Libraries Association 2021 Presentation
Eureka, I found it! - Special Libraries Association 2021 Presentation
Access Innovations, Inc.
 
Getting started with Amazon Bedrock Studio and Control Tower
Getting started with Amazon Bedrock Studio and Control TowerGetting started with Amazon Bedrock Studio and Control Tower
Getting started with Amazon Bedrock Studio and Control Tower
Vladimir Samoylov
 
somanykidsbutsofewfathers-140705000023-phpapp02.pptx
somanykidsbutsofewfathers-140705000023-phpapp02.pptxsomanykidsbutsofewfathers-140705000023-phpapp02.pptx
somanykidsbutsofewfathers-140705000023-phpapp02.pptx
Howard Spence
 
Obesity causes and management and associated medical conditions
Obesity causes and management and associated medical conditionsObesity causes and management and associated medical conditions
Obesity causes and management and associated medical conditions
Faculty of Medicine And Health Sciences
 

Recently uploaded (13)

Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...
Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...
Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...
 
Announcement of 18th IEEE International Conference on Software Testing, Verif...
Announcement of 18th IEEE International Conference on Software Testing, Verif...Announcement of 18th IEEE International Conference on Software Testing, Verif...
Announcement of 18th IEEE International Conference on Software Testing, Verif...
 
Doctoral Symposium at the 17th IEEE International Conference on Software Test...
Doctoral Symposium at the 17th IEEE International Conference on Software Test...Doctoral Symposium at the 17th IEEE International Conference on Software Test...
Doctoral Symposium at the 17th IEEE International Conference on Software Test...
 
Acorn Recovery: Restore IT infra within minutes
Acorn Recovery: Restore IT infra within minutesAcorn Recovery: Restore IT infra within minutes
Acorn Recovery: Restore IT infra within minutes
 
International Workshop on Artificial Intelligence in Software Testing
International Workshop on Artificial Intelligence in Software TestingInternational Workshop on Artificial Intelligence in Software Testing
International Workshop on Artificial Intelligence in Software Testing
 
Bitcoin Lightning wallet and tic-tac-toe game XOXO
Bitcoin Lightning wallet and tic-tac-toe game XOXOBitcoin Lightning wallet and tic-tac-toe game XOXO
Bitcoin Lightning wallet and tic-tac-toe game XOXO
 
0x01 - Newton's Third Law: Static vs. Dynamic Abusers
0x01 - Newton's Third Law:  Static vs. Dynamic Abusers0x01 - Newton's Third Law:  Static vs. Dynamic Abusers
0x01 - Newton's Third Law: Static vs. Dynamic Abusers
 
Bonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdf
Bonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdfBonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdf
Bonzo subscription_hjjjjjjjj5hhhhhhh_2024.pdf
 
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...
 
Eureka, I found it! - Special Libraries Association 2021 Presentation
Eureka, I found it! - Special Libraries Association 2021 PresentationEureka, I found it! - Special Libraries Association 2021 Presentation
Eureka, I found it! - Special Libraries Association 2021 Presentation
 
Getting started with Amazon Bedrock Studio and Control Tower
Getting started with Amazon Bedrock Studio and Control TowerGetting started with Amazon Bedrock Studio and Control Tower
Getting started with Amazon Bedrock Studio and Control Tower
 
somanykidsbutsofewfathers-140705000023-phpapp02.pptx
somanykidsbutsofewfathers-140705000023-phpapp02.pptxsomanykidsbutsofewfathers-140705000023-phpapp02.pptx
somanykidsbutsofewfathers-140705000023-phpapp02.pptx
 
Obesity causes and management and associated medical conditions
Obesity causes and management and associated medical conditionsObesity causes and management and associated medical conditions
Obesity causes and management and associated medical conditions
 

AIRBORNE LIDAR POINT DENSITY

  • 1. AIRBORNE LIDAR POINT DENSITY, MORE TO THE POINT July 23, 2019 Matt Bethel Director of Operations and Technology Merrick & Company
  • 2. How Is Airborne LiDAR Density Measured? 0.5 meter 0.5 meter 0.5 meter ground sample distance (GSD) or nominal point spacing (NPS)
  • 3. How Is Airborne LiDAR Density Measured? Type equation here. 1 meter 1 meter 4 points per square meter (PPSM) = 𝐍𝐏𝐒 = 𝟏 𝑫𝒆𝒏𝒔𝒊𝒕𝒚 𝑫𝒆𝒏𝒔𝒊𝒕𝒚 = 𝟏 𝑵𝑷𝑺 𝟐0.5 meter ground sample distance (GSD) or nominal point spacing (NPS) 𝑫𝒆𝒏𝒔𝒊𝒕𝒚 = 𝐟𝐢𝐫𝐬𝐭 𝐨𝐫 𝐥𝐚𝐬𝐭 𝐫𝐞𝐭𝐮𝐫𝐧 𝐩𝐨𝐢𝐧𝐭 𝐜𝐨𝐮𝐧𝐭 𝒂𝒓𝒆𝒂
  • 4. How Is Airborne LiDAR Density Measured? 1 meter 1 meter 9 or maybe 1 but not 4 PPSM = 4 points per square meter (PPSM) 0.5 meter ground sample distance (GSD) or nominal point spacing (NPS)
  • 5. How Is Airborne LiDAR Density Measured? confusing= Q: Are points on cell boundaries shared for density calculations? A: No, they must be counted within only one cell. REMEMBER THIS!!! 0.5 meter ground sample distance (GSD) or nominal point spacing (NPS)
  • 6. Ways of Measuring Airborne LiDAR Density • Representative samples Pros • Fast and easy to calculate • Good for areas of interests Cons • Biased by many factors such as sidelap, patches, turbulence, etc. • Very localized, not representative of swaths or project extents • Cannot automatedly find problem areas that could be considered failures / specification violations • Difficult to use for reporting
  • 7. Ways of Measuring Airborne LiDAR Density • Representative samples • Per swath Pros • Ideal to compare against planned swath density • Relatively easy to compute • Reasonably batchable – one process per flightline • Good to use for reporting • Is not biased (inflated) by sidelap • No hidden problems with use or reporting – very straightforward Cons • Needs interpretation if flying >50% sidelap to achieve planned density
  • 8. Ways of Measuring Airborne LiDAR Density • Representative samples • Per swath • Aggregate / project wide Pros • Considers all collected points • Good for reporting • Straightforward approach (number of first or last return points / area of project boundary) Cons • Crosslines, sidelap, collection block overlap, and patches can inflate density results • Tabular reporting only will not identify localized density failures. A thematic raster is needed for locating localized density failures. Area of Project Boundary (m2)
  • 9. Ways of Measuring Airborne LiDAR Density • Representative samples • Per swath • Aggregate / project wide • Voronoi / Thiessen polygon Pros • Most accurate representation of point density • Measurement is an area of point influence • Density can be derived by 1/Voronoi area • Unlike grids, polygons share representative edges with neighboring polygons. With grid cells, four of the eight neighboring cells are connected through only a point. Cons • Cumbersome to work with • If used, typically only in representative sample areas or around check points - not project wide
  • 10. Ways of Measuring Airborne LiDAR Density • Representative samples • Per swath • Aggregate / project wide • Voronoi / Thiessen polygon Square meters Pros • Most accurate representation of point density • Measurement is an area of point influence • Density can be derived by 1/Voronoi area • Unlike grids, polygons share representative edges with neighboring polygons. With grid cells, four of the eight neighboring cells are connected through only a point. Cons • Cumbersome to work with • If used, typically only in representative sample areas or around check points - not project wide
  • 11. Ways of Measuring Airborne LiDAR Density • Representative samples • Per swath • Aggregate / project wide • Voronoi / Thiessen polygon PPSM Pros • Most accurate representation of point density • Measurement is an area of point influence • Density can be derived by 1/Voronoi area • Unlike grids, polygons share representative edges with neighboring polygons. With grid cells, four of the eight neighboring cells are connected through only a point. Cons • Cumbersome to work with • If used, typically only in representative sample areas or around check points - not project wide
  • 12. Ways of Measuring Airborne LiDAR Density • Representative samples • Per swath • Aggregate / project wide • Voronoi / Thiessen polygon • Grid / raster / tile Pros • Fast and easy to calculate • Seemingly straightforward approach – use grid or tile scheme to count points and report on lowest divisible points per grid/tile area • Easy to use for reporting – pass fail percentage results and graphic Cons • The results are in pass/fail cell counts yet there are no establish parameters for use or analysis (no passing thresholds) • User selected processing cell size changes the results • Inherent with major problems • Results are severely misunderstood yet widely used and relied upon
  • 13.
  • 14.
  • 15.
  • 16.
  • 17. There is a better way… • In digital signal processing, the minimum sampling rate is limited to 2X the maximum frequency. The purpose for this limit is to preserve important information throughout the transformation. This is known as the Nyquist- Shannon sampling theorem. This 2X law can be applied to LiDAR density measurement and derivative grid creation. This is known as the Nyquist sampling criteria. • The Nyquist sampling criteria states that we must sample at no less than twice the resolution of the smallest detail we intend to measure or model. • This means that grids used for density calculation or raster product creation must have a cell size no less than 2 x NPS.
  • 18.
  • 19.
  • 20.
  • 21. Third Party Reports Showing Varying Pass/Fail Percentage Results These first return density reports were generated using third party software ran on two sample swaths using different cell sizes (highlighted in each screenshot). Note the percentages widely vary with each cell size test.
  • 22. That was all real LiDAR data with random point spacing. Let’s test synthetically created, perfectly spaced point data.
  • 23. If we take three LiDAR swaths Export to an LAS grid file at exactly 2 PPSM / 0.7071067811865470 GSD Then test and report on density using the grid method. We expect 100% passing of all tests.
  • 24. If we take three LiDAR swaths Export to an LAS grid file at exactly 2 PPSM / 0.7071067811865470 meter GSD Then test and report on density using the grid method. We expect 100% passing of all tests.
  • 25.
  • 26.
  • 27.
  • 28.
  • 29.
  • 30. What is going on here?
  • 31. Less than Ideal Scan Pattern Cross track point spacing = 0.5 m Along track point spacing = 1 m Parallel scanline pattern (Riegl) (Minimum required point count per cell = 2) 190 cells contain 2 points 10 cells contain only 1 point 95% pass / 5% fail this test Assume each cell is 1 m X 1 m (1 m2) 40 points across by 10 scan lines = 400 points / 200 cells = 2 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: Data with poor point spacing geometry can yield a high passing percentage “grade” using the grid density test method.
  • 32. Less than Ideal Scan Pattern Cross track point spacing = 0.5 m Along track point spacing = 1 m Zigzag scanline pattern (Optech) (Minimum required point count per cell = 2) 190 cells contain 2 points 10 cells contain only 1 point 95% pass / 5% fail this test Assume each cell is 1 m X 1 m (1 m2) 40 points across by 10 scan lines = 400 points / 200 cells = 2 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: Data with poor point spacing geometry can yield a high passing percentage “grade” using the grid density test method. Zigzag scanline patterns result in the same “grade” as parallel line scanning patterns.
  • 33. Ideal Scan Pattern Cross track point spacing = 0.71 m Along track point spacing = 0.71 m Parallel scanline pattern (Riegl) (Minimum required point count per cell = 2) 100 cells contain 2 points 60 cells contain only 1 point 40 Cells contain more than 2 points 70% pass / 30% fail this test using 20 more points than the previous example and a much improved point spacing distribution Assume each cell is 1 m X 1 m (1 m2) 28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: Data with ideal point spacing geometry results in a much lower passing percentage “grade” using the grid density test method. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
  • 34. Ideal Scan Pattern Cross track point spacing = 0.71 m Along track point spacing (at nadir) = 0.71 m Zigzag scanline pattern (Optech) (Minimum required point count per cell = 2) 98 cells contain 2 points 60 cells contain only 1 point 42 Cells contain more than 2 points 70% pass / 30% fail this test using 20 more points than the previous example and a much improved point spacing distribution Assume each cell is 1 m X 1 m (1 m2) 28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: Data with ideal point spacing geometry results in a much lower passing percentage “grade” using the grid density test method. Zigzag scanline patterns result in the same “grade” as parallel line scanning patterns. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
  • 35. What happens when we increase the size of the testing tile?
  • 36. Ideal Scan Pattern Cross track point spacing = 0.71 m Along track point spacing = 0.71 m Parallel scanline pattern (Riegl) Cell size increased to 2m x 2m (Minimum required point count per cell = 8) 0 cells contain 8 points 10 cells contain less than 8 points 40 Cells contain more than 8 points 80% pass / 20% fail this test using 20 more points than the previous example and a much improved point spacing distribution Assume each cell is 2 m X 2 m (4 m2) 28 points across by 15 scan lines = 420 points / 50 cells / 4 m2 = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which invalidates the results. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
  • 37. Ideal Scan Pattern Cross track point spacing = 0.71 m Along track point spacing (at nadir) = 0.71 m Zigzag scanline pattern (Optech) Cell size increased to 2m x 2m (Minimum required point count per cell = 8) 0 cells contain 8 points 10 cells contain less than 8 points 40 Cells contain more than 8 points 80% pass / 20% fail this test using 20 more points than the previous example and a much improved point spacing distribution Assume each cell is 2 m X 2 m (4 m2) 28 points across by 15 scan lines = 420 points / 50 cells / 4 m2 = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which invalidates the results. Zigzag scanline patterns result in the same “grade” as parallel line scanning patterns. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
  • 38. Ideal Scan Pattern Cross track point spacing = 0.71 m Along track point spacing = 0.71 m Parallel scanline pattern (Riegl) Cell size increased to 5m x 5m (Minimum required point count per cell = 50) 0 cells contain 50 points 4 cells contain less than 50 points 4 Cells contain more than 50 points 50% pass / 50% fail this test using 20 more points than the previous example and a much improved point spacing distribution Assume each cell is 5 m X 5 m (25 m2) 28 points across by 15 scan lines = 420 points / 8 cells / 25 m2 = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which invalidates the results. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
  • 39. Ideal Scan Pattern Cross track point spacing = 0.71 m Along track point spacing (at nadir) = 0.71 m Zigzag scanline pattern (Optech) Cell size increased 5m x 5m (Minimum required point count per cell = 50) 0 cells contain 50 points 3 cells contain less than 50 points 5 Cells contain more than 50 points 60% pass / 40% fail this test using 20 more points than the previous example and a much improved point spacing distribution Assume each cell is 5 m X 5 m (25 m2) 28 points across by 15 scan lines = 420 points / 8 cells / 25 m2 = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: When using the grid density test, increasing the cell size used to analyze point density CHANGES the passing percentage “grade” which invalidates the results. Zigzag scanline patterns result in a similar “grade” as parallel line scanning patterns. This method is severely flawed for analyzing and reporting LiDAR point density, especially as a PASS/FAIL percentage “grade”.
  • 40. Ideal Scan Pattern With Nyquist Sampling Criteria Cross track point spacing = 0.71 m Along track point spacing = 0.71 m Parallel scanline pattern (Riegl) (Minimum required point count per cell = 1) 420 Cells contain more than 1 point 100% pass / 0% fail this test using Assume each cell is 1.42m X 1.42 m (2.02 m2) 28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement Takeaway: Data with ideal point spacing geometry results in a 100% passing percentage “grade” using the Nyquist sampling criteria grid density test method.
  • 41. Ideal Scan Pattern With Nyquist Sampling Criteria Cross track point spacing = 0.71 m Along track point spacing (at nadir) = 0.71 m Zigzag scanline pattern (Optech) (Minimum required point count per cell = 1) Takeaway: Data with ideal point spacing geometry results in a 100% passing percentage “grade” using the Nyquist sampling criteria grid density test method. Zigzag scanline patterns result in the same “grade” as parallel line scanning patterns. 420 Cells contain more than 1 point 100% pass / 0% fail this test using Assume each cell is 1.42m X 1.42 m (2.02 m2) 28 points across by 15 scan lines = 420 points / 200 cells = 2.1 ppsm This passes the USGS LBS v1.x NPS/density requirement
  • 42. Test Cross track spacing (m) Along track spacing (m) at nadir Raster cell size (m) Minimum required point count per cell Number of points in test Percentage of passing cells Percentage of failing cells
  • 43. Test Cross track spacing (m) Along track spacing (m) at nadir Raster cell size (m) Minimum required point count per cell Number of points in test Percentage of passing cells Percentage of failing cells Parallel scan pattern 0.5 1 1 2 400 95% 5% Zigzag scan pattern 0.5 1 1 2 400 95% 5%
  • 44. Test Cross track spacing (m) Along track spacing (m) at nadir Raster cell size (m) Minimum required point count per cell Number of points in test Percentage of passing cells Percentage of failing cells Parallel scan pattern 0.5 1 1 2 400 95% 5% Zigzag scan pattern 0.5 1 1 2 400 95% 5% Parallel scan pattern 0.71 0.71 1 2 420 70% 30% Zigzag scan pattern 0.71 0.71 1 2 420 70% 30%
  • 45. Test Cross track spacing (m) Along track spacing (m) at nadir Raster cell size (m) Minimum required point count per cell Number of points in test Percentage of passing cells Percentage of failing cells Parallel scan pattern 0.5 1 1 2 400 95% 5% Zigzag scan pattern 0.5 1 1 2 400 95% 5% Parallel scan pattern 0.71 0.71 1 2 420 70% 30% Zigzag scan pattern 0.71 0.71 1 2 420 70% 30% Parallel scan pattern 0.71 0.71 2 8 420 80% 20% Zigzag scan pattern 0.71 0.71 2 8 420 80% 20%
  • 46. Test Cross track spacing (m) Along track spacing (m) at nadir Raster cell size (m) Minimum required point count per cell Number of points in test Percentage of passing cells Percentage of failing cells Parallel scan pattern 0.5 1 1 2 400 95% 5% Zigzag scan pattern 0.5 1 1 2 400 95% 5% Parallel scan pattern 0.71 0.71 1 2 420 70% 30% Zigzag scan pattern 0.71 0.71 1 2 420 70% 30% Parallel scan pattern 0.71 0.71 2 8 420 80% 20% Zigzag scan pattern 0.71 0.71 2 8 420 80% 20% Parallel scan pattern 0.71 0.71 5 50 420 50% 50% Zigzag scan pattern 0.71 0.71 5 50 420 60% 40%
  • 47. Test Cross track spacing (m) Along track spacing (m) at nadir Raster cell size (m) Minimum required point count per cell Number of points in test Percentage of passing cells Percentage of failing cells Parallel scan pattern 0.5 1 1 2 400 95% 5% Zigzag scan pattern 0.5 1 1 2 400 95% 5% Parallel scan pattern 0.71 0.71 1 2 420 70% 30% Zigzag scan pattern 0.71 0.71 1 2 420 70% 30% Parallel scan pattern 0.71 0.71 2 8 420 80% 20% Zigzag scan pattern 0.71 0.71 2 8 420 80% 20% Parallel scan pattern 0.71 0.71 5 50 420 50% 50% Zigzag scan pattern 0.71 0.71 5 50 420 60% 40% Parallel scan pattern 0.71 0.71 1.42 1 420 100% 0% Zigzag scan pattern 0.71 0.71 1.42 1 420 100% 0%
  • 48. Conclusions and Recommendations • Representative samples are too limiting for project analysis and reporting. • Swath density analysis is straightforward, reliable, well understood, and very representative. • Aggregate is too generalizing. A supplemental raster is required to identify localized failures. • Voronoi is the most accurate but too cumbersome to use for thorough project analysis. • Grid/raster/tile density method is very effective only if using the Nyquist sampling criteria. A qualifying pass/fail threshold is required. Also, the results of this method are consistent with the results from the reliable swath density analysis. • Never use the flawed grid method with a simple points per square area calculation.
  • 49. Thank You Matt Bethel Director of Operations and Technology Merrick & Company http://www.merrick.com/Geospatial matt.bethel@merrick.com (303) 353-3662