SlideShare a Scribd company logo
Fast and High-Precision 3D Tracking and Position Measurement with MEMS Micromirrors
Veljko Milanović and Wing Kin Lo
Mirrorcle Technologies, Inc.
828 San Pablo Ave., Ste. 109, Albany, CA 94706
veljko@mirrorcletech.com
Abstract - We demonstrate real-time fast-motion tracking of an object in a 3D volume, while obtaining its precise XYZ co-ordinates.
Two separate scanning MEMS micromirror sub-systems track the object in a 20 kHz closed-loop. A demonstration system capable
of tracking full-speed human hand motion provides position information at up to 5m distance with 16-bit precision, or <=20µm
precision on the X and Y axes (up/down, left/right,) and precision on the depth (Z-axis) from 10µm to 1.5mm, depending on distance.
INTRODUCTION
Obtaining real-time 3D co-ordinates of a moving object has
many applications such as gaming [1], robotics and human-
computer interaction applications [2-4], industrial applications
etc. Various technologies have been investigated for and used in
these applications, including sensing via wire-interfaces [2],
ultrasound, and laser interferometry. However a simple and low
cost solution that can provide enough precision and flexibility
has not been available. Recent proliferation of low-cost inertial
sensors has not addressed the problem of position tracking.
Cassinelli et al demonstrated a scanning mirror-based tracking
solution [3-4], however their system does not solve the problem
of object searching/selecting and does not have adequate depth
(Z-axis) measurements.
The objective of this work was to develop and demonstrate
an optical-MEMS based, very low cost and versatile platform
for tracking and position measurement in a variety of situations.
Use of MEMS mirrors [5] with potential for use of wide-angle
lenses provides the possibility of tracking in a very large
volume, and very far distances. E.g. use of remote-control IR
source-detector modules can provide a range of 50m or more.
MULTIPLE TRACKING OPTIONS
We have developed several beam-steering based techniques
to track an object inside a conic volume, as depicted in Fig. 1a.
A. Tracking a photo-detector or a retro-reflector
As depicted in Fig. 1b, there are two laser beams scanned
by two MEMS mirrors into a common volume. Both systems
are pointed in a parallel direction, but are spaced a known
distance d apart (Fig. 2a.) The devices run a spiral search
pattern from origin to maximum angles until they encounter a
photo-detector which synchronously relays its readings to the
control FPGA. From this point forward the devices renew a
search but with an updated origin at the last known position of
the photo-detector. The system is therefore in a perpetual search
mode, although only in a very small neighborhood of the photo-
detector. Full motion tracking (Fig. 2b) was achieved with fast
MEMS devices giving at least 2 kHz of motion bandwidth.
Since only one device can illuminate the target at a time, we
time-multiplex the sub-systems by laser modulation.
In our best-performing setup we use a quadrant photo-
detector which provides additional information for tracking,
specifically the needed adjustments in X and Y to get centered
on the target. Here there are clearly distinct modes: search
(spiraling) and tracking. Tracking is a proportional control
closed-loop based on the quad-detector X and Y inputs as loop
errors. We also implemented a small beam-motion dither on the
MEMS scanners that allows us to measure the tilt-orientation of
the quad-detector (Fig. 2c,) therefore giving us 4 DoF of the
detector and allowing us to use it at any rotational position.
In a nearly identical setup, we placed 2 photo-detectors in
close proximity with the MEMS mirrors. The object being
searched in the 3D volume is a retro-reflector (“cats eye”) or a
corner-cube reflector (both were used in our experiments.) In
this manner both devices can simultaneously illuminate the
target and operate independently.
B. Tracking an LED
As depicted in Fig. 1c, there is a photo-detector near each
one of the MEMS scanning units. An optical source such as a
near-IR LED is the target object that illuminates the
micromirrors. When the mirrors are properly pointed, that
illumination is reflected onto each detector. Therefore no time-
multiplexing or communication to the target is necessary.
3D POSITION MEASUREMENT
Both devices X and Y axes are driven by separate channels
of a 16-bit FPGA system. They achieve angle (negative and
positive) maxima (–θmax, + θmax) when the system sends –K to
+K to its output DAC, where K=215
-1. In most of our
experiments we calibrate our devices to provide θmax=10°, giving
a total scan angle of 20°. When device 1 successfully tracks the
target, the FPGA system records the angle of the device’s x-axis
and y-axis in terms of the open-loop output values OX1 and OY1.
Second device provides knowledge of its open-loop angles OX2
and OY2. The devices are level in y but spaced a known distance
d in x. Therefore when both devices are tracking the object they
see nearly identical Y readings OY1 and OY2, but due to motion
parallax the X readings are different and depend on the distance
of the object. We utilize the X readings to obtain a true distance
of the object to the origin (a point directly between the two
micromirrors) as:
( ) )(
1
tan 21max XX OO
Kd
Z
−
⋅
⋅
=
θ
.
With Z known, X and Y are found from known parameters
and by averaging from two devices’ readings:
( ) ( )
( ) ( )
)(
)(
2
2/tan)(
)(
)(
2
2/tan)(
21
12
max12
21
12
max12
XX
YY
YY
XX
XX
XX
OO
OOd
KZOOY
OO
OOd
KZOOX
−
+
=⋅⋅+=
−
+
=⋅⋅+=
θ
θ
.
RESULTS
Our MEMS devices provided pointing precision >= the
DAC’s 16-bit resolution, and therefore our overall system results
all demonstrated this 16-bit limitation. When target object was
not moving, no single digit of X,Y,Z was changing. Movements
of 1mm on an optical-bench micrometer were easily recorded at
5m distance. With the loop-gain and bandwidth capable of
tracking full-speed human hand motion, the system provides
position information at up to 5m distance with <=20µm
precision on the X and Y axes (up, down, left, right,) and
precision on the depth (Z-axis) from 10µm to 1.5mm, depending
on the distance. Precision can be greatly increased with slower
tracking settings and lower loop-gain in different applications.
[1] J. Brophy-Warren, “Magic Wand: How Hackers Make Use
Of Their Wii-motes,” The Wall Street Journal, Apr. 28th
, 2007.
[2] P. Arcara, et al, “Perception of Depth Information by Means
of a Wire-Actuated Haptic Interface,” Proc. of 2000 IEEE Int.
Conf. on Robotics and Automation, Apr. 2000.
[3] A. Cassinelli, et al, “Smart Laser-Scanner for 3D Human-
Machine Interface,” Int. Conf. on Human Factors in Computing
Systems, Portland, OR, Apr. 02 - 07, 2005, pp. 1138 - 1139.
[4] S. Perrin, et al, “Laser-Based Finger Tracking System
Suitable for MOEMS Integration,” Image and Vision
Computing, New Zealand, 26-28 Nov. 2003, pp.131-136.
[5] V. Milanović, et al, "Gimbal-less Monolithic Silicon
Actuators For Tip-Tilt-Piston Micromirror Applications," IEEE
J. of Select Topics in Quantum Electronics, vol. 10(3), Jun 2004.
Photo Detector
Lasers
y
d
z
x
MEMS
Mirror 1
MEMS
Mirror 2
X = 288.25 mm
Y = -167.48 mm
Z = 3500.7 mm
20°
FOV
MEMS Tracking System
IR Emitter
OR
Reflector
OR
Photo-Detector
(a) (b) (c)
Near IR
source
MEMS
Mirror 1
Photo Detectors
d
y
z
x
MEMS
Mirror 2
Figure 1. (a) Schematic diagram of 3D tracking of a hand-held object in a 3D volume. (b) Schematic of a 3D Tracking setup with
two beam-steering MEMS mirrors aiming their laser sources onto the target. (c) Schematic diagram of 3D tracking and
measurement setup with two MEMS devices steering incident light from a (near-IR) source onto their respective photo-detector.
Sensor Tilt
Sensor Z-Position (distance)
Sensor XY-Position
(a) (b) (c)
Figure 2. (a)Photograph of the two MEMS scanners and amplifiers. The devices are d=75mm apart and aimed in the same
direction. Each amplifier in the background is driven by the FPGA closed-loop controller. (b) A 2s long exposure photograph of
quad-detector tracking. Both laser spots are on the detector, and both devices successfully track the target. (c) GUI screen capture
showing the measured 4 DoF of the detector: position X [mm], position Y [mm], position Z [mm], and tilt of the quad-detector [deg.]
(a) (b) (c)
Figure 3. Gimbal-less dual-axis 4-quadrant devices used in this work: (a) typical device which reaches mechanical tilt from -8° to
+8° on both axes. Device has a 2mm mirror, this larger aperture being more suitable for the setup of Fig. 1c. (b) Voltage vs.
Mechanical tilt angle measurements of a typical 4-quadrant device, linearized by our 4-channel amplifier driving scheme. (c) Small-
signal characteristics of fast devices with 0.8mm mirror used in the setup of Fig. 1b, where larger aperture size is not required.

More Related Content

What's hot

ICRA2019 Poster
ICRA2019 PosterICRA2019 Poster
ICRA2019 Poster
Jack Collins
 
Foot-mounted Inertial Navigation Made Easy
Foot-mounted Inertial Navigation Made EasyFoot-mounted Inertial Navigation Made Easy
Foot-mounted Inertial Navigation Made Easy
oblu.io
 
Inertial Sensor Array Calibration Made Easy !
Inertial Sensor Array Calibration Made Easy !Inertial Sensor Array Calibration Made Easy !
Inertial Sensor Array Calibration Made Easy !
oblu.io
 
An Experimental Study on a Pedestrian Tracking Device
An Experimental Study on a Pedestrian Tracking DeviceAn Experimental Study on a Pedestrian Tracking Device
An Experimental Study on a Pedestrian Tracking Device
oblu.io
 
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
CSCJournals
 
Primal-Dual Coding to Probe Light Transport
Primal-Dual Coding to Probe Light TransportPrimal-Dual Coding to Probe Light Transport
Primal-Dual Coding to Probe Light Transport
Matthew O'Toole
 
Evolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensor
Evolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensorEvolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensor
Evolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensor
oblu.io
 
Foot Mounted Pedestrian Navigation Systems
Foot Mounted Pedestrian Navigation SystemsFoot Mounted Pedestrian Navigation Systems
Foot Mounted Pedestrian Navigation Systems
oblu.io
 
Monocular LSD-SLAM integreation within AR System
Monocular LSD-SLAM integreation within AR SystemMonocular LSD-SLAM integreation within AR System
Monocular LSD-SLAM integreation within AR SystemMarkus Höll
 
Multi Inertial Measurement Units (MIMU) Platforms: Designs & Applications
Multi Inertial Measurement Units (MIMU)  Platforms: Designs & ApplicationsMulti Inertial Measurement Units (MIMU)  Platforms: Designs & Applications
Multi Inertial Measurement Units (MIMU) Platforms: Designs & Applications
oblu.io
 
CHARACTERISTICSOFASTERGDEMVERSION2.pptx
CHARACTERISTICSOFASTERGDEMVERSION2.pptxCHARACTERISTICSOFASTERGDEMVERSION2.pptx
CHARACTERISTICSOFASTERGDEMVERSION2.pptxgrssieee
 
Compressive Light Field Photography using Overcomplete Dictionaries and Optim...
Compressive Light Field Photography using Overcomplete Dictionaries and Optim...Compressive Light Field Photography using Overcomplete Dictionaries and Optim...
Compressive Light Field Photography using Overcomplete Dictionaries and Optim...
Ankit Thiranh
 
C05131525
C05131525C05131525
C05131525
IOSR-JEN
 
Optical interferometery to detect sound waves as an analogue for gravitationa...
Optical interferometery to detect sound waves as an analogue for gravitationa...Optical interferometery to detect sound waves as an analogue for gravitationa...
Optical interferometery to detect sound waves as an analogue for gravitationa...Thomas Actn
 
FV_IGARSS11.ppt
FV_IGARSS11.pptFV_IGARSS11.ppt
FV_IGARSS11.pptgrssieee
 

What's hot (17)

AbratenkoTalk
AbratenkoTalkAbratenkoTalk
AbratenkoTalk
 
ICRA2019 Poster
ICRA2019 PosterICRA2019 Poster
ICRA2019 Poster
 
Foot-mounted Inertial Navigation Made Easy
Foot-mounted Inertial Navigation Made EasyFoot-mounted Inertial Navigation Made Easy
Foot-mounted Inertial Navigation Made Easy
 
Inertial Sensor Array Calibration Made Easy !
Inertial Sensor Array Calibration Made Easy !Inertial Sensor Array Calibration Made Easy !
Inertial Sensor Array Calibration Made Easy !
 
An Experimental Study on a Pedestrian Tracking Device
An Experimental Study on a Pedestrian Tracking DeviceAn Experimental Study on a Pedestrian Tracking Device
An Experimental Study on a Pedestrian Tracking Device
 
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
 
Primal-Dual Coding to Probe Light Transport
Primal-Dual Coding to Probe Light TransportPrimal-Dual Coding to Probe Light Transport
Primal-Dual Coding to Probe Light Transport
 
Evolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensor
Evolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensorEvolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensor
Evolution of a shoe-mounted multi-IMU pedestrian dead reckoning PDR sensor
 
Foot Mounted Pedestrian Navigation Systems
Foot Mounted Pedestrian Navigation SystemsFoot Mounted Pedestrian Navigation Systems
Foot Mounted Pedestrian Navigation Systems
 
Monocular LSD-SLAM integreation within AR System
Monocular LSD-SLAM integreation within AR SystemMonocular LSD-SLAM integreation within AR System
Monocular LSD-SLAM integreation within AR System
 
Multi Inertial Measurement Units (MIMU) Platforms: Designs & Applications
Multi Inertial Measurement Units (MIMU)  Platforms: Designs & ApplicationsMulti Inertial Measurement Units (MIMU)  Platforms: Designs & Applications
Multi Inertial Measurement Units (MIMU) Platforms: Designs & Applications
 
CHARACTERISTICSOFASTERGDEMVERSION2.pptx
CHARACTERISTICSOFASTERGDEMVERSION2.pptxCHARACTERISTICSOFASTERGDEMVERSION2.pptx
CHARACTERISTICSOFASTERGDEMVERSION2.pptx
 
1886 1892
1886 18921886 1892
1886 1892
 
Compressive Light Field Photography using Overcomplete Dictionaries and Optim...
Compressive Light Field Photography using Overcomplete Dictionaries and Optim...Compressive Light Field Photography using Overcomplete Dictionaries and Optim...
Compressive Light Field Photography using Overcomplete Dictionaries and Optim...
 
C05131525
C05131525C05131525
C05131525
 
Optical interferometery to detect sound waves as an analogue for gravitationa...
Optical interferometery to detect sound waves as an analogue for gravitationa...Optical interferometery to detect sound waves as an analogue for gravitationa...
Optical interferometery to detect sound waves as an analogue for gravitationa...
 
FV_IGARSS11.ppt
FV_IGARSS11.pptFV_IGARSS11.ppt
FV_IGARSS11.ppt
 

Similar to Fast and High-Precision 3D Tracking and Position Measurement with MEMS Micromirrors

Sensors for remote sensing
Sensors for remote sensingSensors for remote sensing
Sensors for remote sensing
Mohsin Siddique
 
Motion capture document
Motion capture documentMotion capture document
Motion capture document
harini501
 
Sensors optimized for 3 d digitization
Sensors optimized for 3 d digitizationSensors optimized for 3 d digitization
Sensors optimized for 3 d digitization
Basavaraj Patted
 
Types of scanners
Types of scannersTypes of scanners
Types of scanners
Pramoda Raj
 
Mumma_Radar_Lab_Posters
Mumma_Radar_Lab_PostersMumma_Radar_Lab_Posters
Mumma_Radar_Lab_PostersDr. Ali Nassib
 
Application of lasers
Application of lasersApplication of lasers
Application of lasers
Alexander Decker
 
OE Instrumentation_02_Interferometry_1.pdf
OE Instrumentation_02_Interferometry_1.pdfOE Instrumentation_02_Interferometry_1.pdf
OE Instrumentation_02_Interferometry_1.pdf
JamesWalter40
 
A Fast Single-Pixel Laser Imager for VR/AR Headset Tracking
A Fast Single-Pixel Laser Imager for VR/AR Headset TrackingA Fast Single-Pixel Laser Imager for VR/AR Headset Tracking
A Fast Single-Pixel Laser Imager for VR/AR Headset Tracking
Ping Hsu
 
MOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdf
MOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdfMOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdf
MOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdf
BhuvanaN12
 
mems based optical coherence tomography imaging
mems based optical coherence tomography imagingmems based optical coherence tomography imaging
mems based optical coherence tomography imaging
Gayathri Pv
 
Ultimate astronomicalimaging
Ultimate astronomicalimagingUltimate astronomicalimaging
Ultimate astronomicalimagingClifford Stone
 
MIRAS: the instrument aboard SMOS
MIRAS: the instrument aboard SMOSMIRAS: the instrument aboard SMOS
MIRAS: the instrument aboard SMOS
adrianocamps
 
Missile detection and automatic destroy system
Missile detection and automatic destroy system Missile detection and automatic destroy system
Missile detection and automatic destroy system
LokeshLavakusha
 
Solar cell powering with integrated global positioning system for
Solar cell powering with integrated global positioning system forSolar cell powering with integrated global positioning system for
Solar cell powering with integrated global positioning system forAlexander Decker
 
Measurement techniques in micro machining PDF by badebhau4@gmail.com
Measurement techniques in micro machining PDF by badebhau4@gmail.comMeasurement techniques in micro machining PDF by badebhau4@gmail.com
Measurement techniques in micro machining PDF by badebhau4@gmail.com
Er. Bade Bhausaheb
 
A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...
A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...
A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...
Ping Hsu
 

Similar to Fast and High-Precision 3D Tracking and Position Measurement with MEMS Micromirrors (20)

Sensors for remote sensing
Sensors for remote sensingSensors for remote sensing
Sensors for remote sensing
 
Motion capture document
Motion capture documentMotion capture document
Motion capture document
 
Remote Sensin
Remote SensinRemote Sensin
Remote Sensin
 
Sensors optimized for 3 d digitization
Sensors optimized for 3 d digitizationSensors optimized for 3 d digitization
Sensors optimized for 3 d digitization
 
Types of scanners
Types of scannersTypes of scanners
Types of scanners
 
Mumma_Radar_Lab_Posters
Mumma_Radar_Lab_PostersMumma_Radar_Lab_Posters
Mumma_Radar_Lab_Posters
 
Epid
EpidEpid
Epid
 
Application of lasers
Application of lasersApplication of lasers
Application of lasers
 
OE Instrumentation_02_Interferometry_1.pdf
OE Instrumentation_02_Interferometry_1.pdfOE Instrumentation_02_Interferometry_1.pdf
OE Instrumentation_02_Interferometry_1.pdf
 
Robotic sensors
Robotic sensorsRobotic sensors
Robotic sensors
 
A Fast Single-Pixel Laser Imager for VR/AR Headset Tracking
A Fast Single-Pixel Laser Imager for VR/AR Headset TrackingA Fast Single-Pixel Laser Imager for VR/AR Headset Tracking
A Fast Single-Pixel Laser Imager for VR/AR Headset Tracking
 
MOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdf
MOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdfMOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdf
MOD 5 SVIT NOTES VTU SYLLABUS 2018 SCHEME.pdf
 
mems based optical coherence tomography imaging
mems based optical coherence tomography imagingmems based optical coherence tomography imaging
mems based optical coherence tomography imaging
 
Project
ProjectProject
Project
 
Ultimate astronomicalimaging
Ultimate astronomicalimagingUltimate astronomicalimaging
Ultimate astronomicalimaging
 
MIRAS: the instrument aboard SMOS
MIRAS: the instrument aboard SMOSMIRAS: the instrument aboard SMOS
MIRAS: the instrument aboard SMOS
 
Missile detection and automatic destroy system
Missile detection and automatic destroy system Missile detection and automatic destroy system
Missile detection and automatic destroy system
 
Solar cell powering with integrated global positioning system for
Solar cell powering with integrated global positioning system forSolar cell powering with integrated global positioning system for
Solar cell powering with integrated global positioning system for
 
Measurement techniques in micro machining PDF by badebhau4@gmail.com
Measurement techniques in micro machining PDF by badebhau4@gmail.comMeasurement techniques in micro machining PDF by badebhau4@gmail.com
Measurement techniques in micro machining PDF by badebhau4@gmail.com
 
A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...
A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...
A MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY P...
 

More from Ping Hsu

Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...
Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...
Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...
Ping Hsu
 
Two-Axis Scanning Mirror for Free-Space Optical Communication between UAVs
Two-Axis Scanning Mirror for Free-Space Optical Communication between UAVsTwo-Axis Scanning Mirror for Free-Space Optical Communication between UAVs
Two-Axis Scanning Mirror for Free-Space Optical Communication between UAVs
Ping Hsu
 
FULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAY
FULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAYFULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAY
FULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAY
Ping Hsu
 
MEMS Mirror Based Dynamic Solid State Lighting Module
MEMS Mirror Based Dynamic Solid State Lighting ModuleMEMS Mirror Based Dynamic Solid State Lighting Module
MEMS Mirror Based Dynamic Solid State Lighting Module
Ping Hsu
 
MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...
MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...
MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...
Ping Hsu
 
UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability
UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability
UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability
Ping Hsu
 

More from Ping Hsu (6)

Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...
Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...
Closed Loop Control of Gimbal-less MEMS Mirrors for Increased Bandwidth in Li...
 
Two-Axis Scanning Mirror for Free-Space Optical Communication between UAVs
Two-Axis Scanning Mirror for Free-Space Optical Communication between UAVsTwo-Axis Scanning Mirror for Free-Space Optical Communication between UAVs
Two-Axis Scanning Mirror for Free-Space Optical Communication between UAVs
 
FULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAY
FULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAYFULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAY
FULLY-FUNCTIONAL TIP-TILT-PISTON MICROMIRROR ARRAY
 
MEMS Mirror Based Dynamic Solid State Lighting Module
MEMS Mirror Based Dynamic Solid State Lighting ModuleMEMS Mirror Based Dynamic Solid State Lighting Module
MEMS Mirror Based Dynamic Solid State Lighting Module
 
MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...
MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...
MEMS BASED OPTICAL COHERENCE TOMOGRAPHY IMAGING SYSTEM AND OPTICAL BIOPSY PRO...
 
UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability
UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability
UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability
 

Recently uploaded

20240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 202420240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 2024
Matthew Sinclair
 
How to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptxHow to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptx
danishmna97
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
Matthew Sinclair
 
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptxSecstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
nkrafacyberclub
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
Guy Korland
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Aggregage
 
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdfUni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems S.M.S.A.
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
Alan Dix
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
DanBrown980551
 
Large Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial ApplicationsLarge Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial Applications
Rohit Gautam
 
Microsoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdfMicrosoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdf
Uni Systems S.M.S.A.
 
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
Neo4j
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
mikeeftimakis1
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Albert Hoitingh
 
Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
Alpen-Adria-Universität
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
Neo4j
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Nexer Digital
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Paige Cruz
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
DianaGray10
 
Free Complete Python - A step towards Data Science
Free Complete Python - A step towards Data ScienceFree Complete Python - A step towards Data Science
Free Complete Python - A step towards Data Science
RinaMondal9
 

Recently uploaded (20)

20240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 202420240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 2024
 
How to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptxHow to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptx
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
 
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptxSecstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
 
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdfUni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdf
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
 
Large Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial ApplicationsLarge Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial Applications
 
Microsoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdfMicrosoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdf
 
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
 
Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
 
Free Complete Python - A step towards Data Science
Free Complete Python - A step towards Data ScienceFree Complete Python - A step towards Data Science
Free Complete Python - A step towards Data Science
 

Fast and High-Precision 3D Tracking and Position Measurement with MEMS Micromirrors

  • 1. Fast and High-Precision 3D Tracking and Position Measurement with MEMS Micromirrors Veljko Milanović and Wing Kin Lo Mirrorcle Technologies, Inc. 828 San Pablo Ave., Ste. 109, Albany, CA 94706 veljko@mirrorcletech.com Abstract - We demonstrate real-time fast-motion tracking of an object in a 3D volume, while obtaining its precise XYZ co-ordinates. Two separate scanning MEMS micromirror sub-systems track the object in a 20 kHz closed-loop. A demonstration system capable of tracking full-speed human hand motion provides position information at up to 5m distance with 16-bit precision, or <=20µm precision on the X and Y axes (up/down, left/right,) and precision on the depth (Z-axis) from 10µm to 1.5mm, depending on distance. INTRODUCTION Obtaining real-time 3D co-ordinates of a moving object has many applications such as gaming [1], robotics and human- computer interaction applications [2-4], industrial applications etc. Various technologies have been investigated for and used in these applications, including sensing via wire-interfaces [2], ultrasound, and laser interferometry. However a simple and low cost solution that can provide enough precision and flexibility has not been available. Recent proliferation of low-cost inertial sensors has not addressed the problem of position tracking. Cassinelli et al demonstrated a scanning mirror-based tracking solution [3-4], however their system does not solve the problem of object searching/selecting and does not have adequate depth (Z-axis) measurements. The objective of this work was to develop and demonstrate an optical-MEMS based, very low cost and versatile platform for tracking and position measurement in a variety of situations. Use of MEMS mirrors [5] with potential for use of wide-angle lenses provides the possibility of tracking in a very large volume, and very far distances. E.g. use of remote-control IR source-detector modules can provide a range of 50m or more. MULTIPLE TRACKING OPTIONS We have developed several beam-steering based techniques to track an object inside a conic volume, as depicted in Fig. 1a. A. Tracking a photo-detector or a retro-reflector As depicted in Fig. 1b, there are two laser beams scanned by two MEMS mirrors into a common volume. Both systems are pointed in a parallel direction, but are spaced a known distance d apart (Fig. 2a.) The devices run a spiral search pattern from origin to maximum angles until they encounter a photo-detector which synchronously relays its readings to the control FPGA. From this point forward the devices renew a search but with an updated origin at the last known position of the photo-detector. The system is therefore in a perpetual search mode, although only in a very small neighborhood of the photo- detector. Full motion tracking (Fig. 2b) was achieved with fast MEMS devices giving at least 2 kHz of motion bandwidth. Since only one device can illuminate the target at a time, we time-multiplex the sub-systems by laser modulation. In our best-performing setup we use a quadrant photo- detector which provides additional information for tracking, specifically the needed adjustments in X and Y to get centered on the target. Here there are clearly distinct modes: search (spiraling) and tracking. Tracking is a proportional control closed-loop based on the quad-detector X and Y inputs as loop errors. We also implemented a small beam-motion dither on the MEMS scanners that allows us to measure the tilt-orientation of the quad-detector (Fig. 2c,) therefore giving us 4 DoF of the detector and allowing us to use it at any rotational position. In a nearly identical setup, we placed 2 photo-detectors in close proximity with the MEMS mirrors. The object being searched in the 3D volume is a retro-reflector (“cats eye”) or a corner-cube reflector (both were used in our experiments.) In this manner both devices can simultaneously illuminate the target and operate independently. B. Tracking an LED As depicted in Fig. 1c, there is a photo-detector near each one of the MEMS scanning units. An optical source such as a near-IR LED is the target object that illuminates the micromirrors. When the mirrors are properly pointed, that illumination is reflected onto each detector. Therefore no time- multiplexing or communication to the target is necessary. 3D POSITION MEASUREMENT Both devices X and Y axes are driven by separate channels of a 16-bit FPGA system. They achieve angle (negative and positive) maxima (–θmax, + θmax) when the system sends –K to +K to its output DAC, where K=215 -1. In most of our experiments we calibrate our devices to provide θmax=10°, giving a total scan angle of 20°. When device 1 successfully tracks the target, the FPGA system records the angle of the device’s x-axis and y-axis in terms of the open-loop output values OX1 and OY1. Second device provides knowledge of its open-loop angles OX2 and OY2. The devices are level in y but spaced a known distance d in x. Therefore when both devices are tracking the object they see nearly identical Y readings OY1 and OY2, but due to motion parallax the X readings are different and depend on the distance of the object. We utilize the X readings to obtain a true distance of the object to the origin (a point directly between the two micromirrors) as: ( ) )( 1 tan 21max XX OO Kd Z − ⋅ ⋅ = θ . With Z known, X and Y are found from known parameters and by averaging from two devices’ readings: ( ) ( ) ( ) ( ) )( )( 2 2/tan)( )( )( 2 2/tan)( 21 12 max12 21 12 max12 XX YY YY XX XX XX OO OOd KZOOY OO OOd KZOOX − + =⋅⋅+= − + =⋅⋅+= θ θ . RESULTS Our MEMS devices provided pointing precision >= the DAC’s 16-bit resolution, and therefore our overall system results all demonstrated this 16-bit limitation. When target object was not moving, no single digit of X,Y,Z was changing. Movements of 1mm on an optical-bench micrometer were easily recorded at 5m distance. With the loop-gain and bandwidth capable of tracking full-speed human hand motion, the system provides position information at up to 5m distance with <=20µm precision on the X and Y axes (up, down, left, right,) and
  • 2. precision on the depth (Z-axis) from 10µm to 1.5mm, depending on the distance. Precision can be greatly increased with slower tracking settings and lower loop-gain in different applications. [1] J. Brophy-Warren, “Magic Wand: How Hackers Make Use Of Their Wii-motes,” The Wall Street Journal, Apr. 28th , 2007. [2] P. Arcara, et al, “Perception of Depth Information by Means of a Wire-Actuated Haptic Interface,” Proc. of 2000 IEEE Int. Conf. on Robotics and Automation, Apr. 2000. [3] A. Cassinelli, et al, “Smart Laser-Scanner for 3D Human- Machine Interface,” Int. Conf. on Human Factors in Computing Systems, Portland, OR, Apr. 02 - 07, 2005, pp. 1138 - 1139. [4] S. Perrin, et al, “Laser-Based Finger Tracking System Suitable for MOEMS Integration,” Image and Vision Computing, New Zealand, 26-28 Nov. 2003, pp.131-136. [5] V. Milanović, et al, "Gimbal-less Monolithic Silicon Actuators For Tip-Tilt-Piston Micromirror Applications," IEEE J. of Select Topics in Quantum Electronics, vol. 10(3), Jun 2004. Photo Detector Lasers y d z x MEMS Mirror 1 MEMS Mirror 2 X = 288.25 mm Y = -167.48 mm Z = 3500.7 mm 20° FOV MEMS Tracking System IR Emitter OR Reflector OR Photo-Detector (a) (b) (c) Near IR source MEMS Mirror 1 Photo Detectors d y z x MEMS Mirror 2 Figure 1. (a) Schematic diagram of 3D tracking of a hand-held object in a 3D volume. (b) Schematic of a 3D Tracking setup with two beam-steering MEMS mirrors aiming their laser sources onto the target. (c) Schematic diagram of 3D tracking and measurement setup with two MEMS devices steering incident light from a (near-IR) source onto their respective photo-detector. Sensor Tilt Sensor Z-Position (distance) Sensor XY-Position (a) (b) (c) Figure 2. (a)Photograph of the two MEMS scanners and amplifiers. The devices are d=75mm apart and aimed in the same direction. Each amplifier in the background is driven by the FPGA closed-loop controller. (b) A 2s long exposure photograph of quad-detector tracking. Both laser spots are on the detector, and both devices successfully track the target. (c) GUI screen capture showing the measured 4 DoF of the detector: position X [mm], position Y [mm], position Z [mm], and tilt of the quad-detector [deg.] (a) (b) (c) Figure 3. Gimbal-less dual-axis 4-quadrant devices used in this work: (a) typical device which reaches mechanical tilt from -8° to +8° on both axes. Device has a 2mm mirror, this larger aperture being more suitable for the setup of Fig. 1c. (b) Voltage vs. Mechanical tilt angle measurements of a typical 4-quadrant device, linearized by our 4-channel amplifier driving scheme. (c) Small- signal characteristics of fast devices with 0.8mm mirror used in the setup of Fig. 1b, where larger aperture size is not required.