SlideShare a Scribd company logo
1 of 30
Detection, Recognition, and Identification (DRI)
What is DRI?
DRI is a universally accepted set of standards providing a means of measuring the
distance whereby a thermal sensor can produce an image of a specific target. This
takes into consideration many different criterion such as noise, array size, optical
blur, lens depletion, aperture, atmospheric depletion, detector pitch and many
more.
Detection
Detection means that you will be able to see the target, however it will be little more
than a spec. Specifically it means that the target is visible on at least two pixels, and
that there is a good chance that the target is actually something of suspicion.
Detection: Ability to distinguish an object from the background.
Recognition
Recognition does not mean that you can recognize an individual. Recognition
simply means that you are able to recognize an object’s class (is it a human or a car,
is it a truck or a tank, etc).
Identification
Identification of an object means that you are able to differentiate between objects.
For example, being able to identify the type of vehicle not just its class.
Identification: ability to describe the object in details (a man with a hat, a deer, a
Jeep.
The following pictures illustrate these definitions:
Left image: Detection – At several kms, 2 targets are detected out of the background
Center image: Recognition - a human is walking along the fence
Right image: Identification – 2 males with trousers and jackets are identified –
one is smoking.
• Thermal sensors have a resolution of 640×480 which is over 300,000 pixels.
Human “detection” only requires 3.6 of those pixels and “identification” only
requires 230 pixels, which is an extraordinarily small amount on the screen that can
easily go unnoticed by the human eye.
The Johnson criteria
Johnson defined thresholds, known as the Johnson criteria, as the minimum number
of line pairs to detect, recognize or identify targets captured by scene imagers. The
Johnson criteria was based on experiments using observers. It estimates the number
of line pairs required across a target, and indicates that there is at least a 50% chance
that an observer can accurately achieve the criteria.
The Johnson criteria state the following thresholds that are achieved 50% of the time:
• Detection: We need 1.0 +/− 0.25 line pairs to detect an object 50% of the time.
• Recognition: The type object can be discerned, a person versus a car (4 +/− 0.8 line
pairs)
• Identification: A specific object can be discerned, a woman versus a man, the
specific car (6.4 +/− 1.5 line pairs)
• He found that the minimum resolution required for a particular "decision response“
was nearly constant for a group of nine military targets. The results of these
experiments are tabulated in Table 1. The data show that the minimum resolution
required for a particular decision activity is a constant for nine military targets
within a maximum error excursion of +25%.
• The methodology developed by Johnson was simple and straight forward. A target
was moved to a range where it was just barely detectable. A bar pattern was placed
in the field of view and its spatial frequency was increased until it could barely be
resolved at the same range, i.e. the number of lines on the bar pattern was
increased until they could no longer be distinguished. The spatial frequency of the
pattern was specified in terms of the number of lines in the pattern subtended by
the object's minimum dimension as illustrated in Figure 1. The same methodology
was used for orientation, recognition and identification.
FACTORS INFLUENCING TARGET DETECTION AND RECOGNITION
The scene
1. The size of the picture or displayed image.
2. Numbers, sizes, shapes, and scene distribution of areas contextually likely to contain
the target object.
3. Scene objects: numbers, shapes and patterns, achromatic and color contrasts, colors,
(hue, saturation, lightness), acutance, amount of resolved details, all both absolutely
and relative to the target object.
4. Scene distribution of objects.
5. Granularity, noise.
6. Total available information content and amount of each type of information.
7. Average image brightness or lightness.
8. Contextual cues to target object location.
The target
1. Location in the image format.
2. Location in the scene
3. Shape and pattern.
4. Size, color, resolution(s), acutance, lightness or brightness.
5. Type and degree of isolation from background and objects.
The observer
1. Training
2. Experience
3. Native ability
4. Instructions and task briefing
5. Search habits
6. Motivation
7. Compromise on speed versus accuracy
8. Assumptions.
Putting it all together
If we ignore the effects of the atmosphere, normal recognition would be 25% of
the detection and identification would be 12.5% of the detection range. Weather
is almost never ideal so in reality these percentages are almost always reduced.
“Ideal Conditions” Another thing that is often not mentioned is that these
ratings are based on what is termed “ideal conditions” which rarely happen in
the real world. The average environmental application will get 25% less than the
distance that the thermal camera is rated for and in extreme conditions can be
less than 10% of the rated distance.
Thermal camera characteristics influencing detection range
• With Johnson criteria in mind, it is easy to understand that the spatial resolution of a
camera is an important parameter determining the quality of the security system.
The spatial resolution: is the camera’s ability to capture very small details at great
distances. It is closely related to the number of camera pixels. Clearly, the more
pixels there are, the better the spatial resolution will be and the greater the detection
range will be.
• However, the thermal sensitivity has also a high impact on detection performance.
The thermal sensitivity is the ability to distinguish very small differences in
temperature. The “cooled” IR cameras offer better thermal sensitivity than
“uncooled” IR cameras. For the same number of pixels, the image of a “cooled”
thermal camera is of better quality (i.e. less “noisy”) than an “uncooled” camera’s
image.
Thermal camera characteristics influencing detection range
• Left: image from a thermal imager with poor thermal sensitivity
• Right: image from a cooled thermal camera with high thermal sensitivity
Another important characteristic of an infrared camera is the field of view (FOV). A
large panoramic FOV is necessary in order to enhance the probability to detect a single
event into a complete 360-degree scene. See in the picture below the image captured by
a thermal sensor with a wide panoramic horizontal FOV, and, in green in the center, the
area screened by a PTZ camera zoomed to a 2° FOV. It is clear that this PTZ camera
has no chance at all to detect any event in this scenery.
Thermal camera characteristics influencing detection range
Converting Line Pair to Pixels
• How many line pairs do we need to detect something 100% of the time? First,
let’s convert line pairs to pixels because the resolution of today’s cameras is
specified in pixels rather than lines.
• Line pair resolution is defined as at least one black and one white detection.
Theoretically, you could have one pixel per line, but in reality, it requires more
than 2 pixels. We need at least one pixel that detects black, a second that detects
white, and a third that detects another black. At a minimum, we need at least 3
pixels to detect a small white object. However, there’s more to this. We need to
take into account a certain amount of statistical uncertainty of a sensor picking
up the signal.
The Kell Factor
In the real world of detection, the white spot could appear somewhere in between
the pixels of the sensor. This means we need more pixels to actually detect the
small object. The concept of the uncertainty of where an object appears on a sensor
was introduced in 1934 by Raymond Kell. It was first applied to sampling an image
to avoid beat frequencies affecting the image. The Kell factor is a fudge factor that
attempts to adjust for the uncertainty of analog sensors. The number has changed
depending on how it is used. Kell said the number should be 0.64, but today people
usually use 0.7.
The Kell factor increases the number of pixels required. We adjust the pixels using
the formula: pixels/ kell factor , 3 / 0.7 = 4.29 vertical pixels.
The Mesnik Factors
 Besides the Kell factor, there are other real-world factors that can affect what we
see. For example, long-range lenses sometimes do not have the clarity they require
to accommodate the higher resolution sensors. Other factors include low-light
operation and atmospheric conditions that affect visibility.
 The atmospheric conditions can be very obvious like rain or fog or can be subtle
like heat waves or higher humidity. The view can be reduced dramatically by fog,
haze (smoke), rain, or snow. Unfortunately, the amount of reduced vision can vary
depending on the intensity of fog or rain.
 Long range cameras must also be mounted to very stable platforms. Placing a
camera on top of a tower can cause vibration that affects the clarity of the video.
Some long-range cameras include image stabilization. There are electronic and
mechanical mechanisms for stabilizing the cameras.
 We can adjust for some of these real-world visual problems. We have come up
with the following factors that result in the need for increased pixel count for
meeting the visual criteria.
• Lens factor: From 1 to 0.6 depending on the actual lens clarity selected
• Low light noise factor: Thermal cameras are affected by the difference in
temperature between the object and the background.
• Atmospheric Factors: Heat waves and other clear day factors such as humidity
increase with distance.
• Atmospheric Distance Factor. These are all estimates of atmospheric effects on
visibility. Actual visibility is affected by the difference in contrast between the object
and the background.
For distance less than 1000 m (0.62 miles) = 1.0
for distance between 1001m – 2000 m (1.24 miles) = 0.9,
for distance 2001m to 3000 m (1.86 miles) = 0.85,
for distance over 3001 m to 5000 m = 0.8
for distance 5,001 m to 10,000 m = 0.75
for distance 10,001 m (6.2 miles) to 15000 m (9.3 miles) = 0.7
for distance 15001m to 20000 m (12.4 miles) = 0.65
An example of using the factors:
• The following example is a camera system used at night, that will observe objects
about 2-miles away. We excluded severe atmospheric conditions such as fog that
could reduce visibility dramatically. We calculated the pixel counts for a thermal
camera with similar lens magnification.
Reasonably good lens (costs a lot) = 0.9
Used at night: = 1.0 for thermal camera
Distance = 3219 m (2 miles) = 0.8
– Total for Thermal = 0.9 x 1.0 x 0.8 = 0.72
– Total Pixels for Thermal Camera = 4.29 / 0.72 = 5.96 pixels
Here’s the summary of the new DRI standard.
Detection: 8 vertical pixels across the target
Recognition: we require eight times the 4 line pairs or 32 vertical pixels across
the target
Identification: we need eight times the 8 line pairs = 64 vertical pixels across
the target
Spatial Frequency Concept
• Spatial frequency is typically measured per cycles (or line pairs) per a unit angle or
a unit length
Evaluation of ranges of effective surveillance
• It is possible to measure directly detection, recognition, and identification ranges of
a target of interest and to evaluate the tested thermal camera on the basis of the test
results. However, it is a risky solution. The ranges vary with observation conditions
(atmosphere, background) and it is relatively easy to manipulate with the detection,
recognition and identification ranges at real conditions if the observation conditions
are not very precisely specified. Next, it is difficult to compare test results of
different thermal cameras tested at different time periods and at different
observation conditions.
• The safest way to evaluate surveillance thermal cameras is to measure MRTD of
this thermal imager and calculate the detection, recognition, and identification
ranges of a standard NATO target using methodology proposed by this standard:
STANAG 4347.
• The detection, recognition, and identification ranges of the standard NATO target
are potentially a good criterion to be used in requirements on surveillance thermal
imagers. It is apparently a good idea to present requirements on a surveillance
thermal imager by presenting requirements on the performance ranges (Table)
In order to calculate the performance ranges of a thermal imager we must
know its MRTD function. Therefore a more common way to specify requirements on
surveillance thermal cameras is to present requirements on MRTD characteristic in a
form shown in Table.
• Now, let us discuss the way to calculate the detection and recognition ranges on the
basis of the measured and known MRTD function of the tested thermal camera
using the recommendation from STANAG 4347. The summary of the
recommendations is shown in Table.
• In details, the detection, recognition and identification ranges of the standard NATO
targets can be calculated using the below presented algorithm.
1. Convert MRTD characteristic into a new one by changing variable from spatial
frequency n [mrad-1] to the range R [km] using the following formulas
2. Calculate decrease in the initial temperature difference DT (it was assumed
that initial = = 2 K) between the target and the background due to limited
atmospheric transmission.
• Determine the respective nominal static ranges as the intersections of and the
converted MRTD functions.
calculation of the detection ranges of the standard NATO target using a thermal camera
of MRTD function shown in Fig. The calculation results are shown in Fig. and we can
conclude that the detection ranges are the following:
 7.2 km at good atmospheric transmission.
 2.8 km at limited atmospheric transmission.
Calculation of the recognition and identification ranges can be done in the same manner.
A comparison between the ACQUIRE Model and the Johnson Criteria assuming standard N50 values for each
discrimination level. The distance difference is measured at the 50% probability of detect and compares the
difference between the Johnson Criteria and the AQUIRE Metric. These values were calculated for a system viewing
a target of 0.75 meter height, with an imaging system having a 10 micron pixel and a 10 millimeter focal length.
Determine range for camouflaged targets
• In 2004, the ACQUIRE-LC metric was introduced to determine range for
camouflaged targets. Camouflage developers showed that camouflage did indeed
have an impact on the range at which a target could be detected. New equations
were set to govern the type of camouflage and amount of clutter in the scene. These
can be seen in Equations.
THANK YOU…

More Related Content

What's hot

Visual Product Identification For Blind Peoples
Visual Product Identification For Blind PeoplesVisual Product Identification For Blind Peoples
Visual Product Identification For Blind PeoplesIRJET Journal
 
Image Classification And Skin cancer detection
Image Classification And Skin cancer detectionImage Classification And Skin cancer detection
Image Classification And Skin cancer detectionEman Othman
 
Multi modal biometric system
Multi modal biometric systemMulti modal biometric system
Multi modal biometric systemAalaa Khattab
 
SEPARATION OF OVERLAPPED FINGERPRINTS
SEPARATION OF OVERLAPPED FINGERPRINTSSEPARATION OF OVERLAPPED FINGERPRINTS
SEPARATION OF OVERLAPPED FINGERPRINTSpadma gade
 
Pattern recognition palm print authentication system
Pattern recognition palm print authentication systemPattern recognition palm print authentication system
Pattern recognition palm print authentication systemMazin Alwaaly
 
implementation of finger vein authentication technique
implementation of finger vein authentication techniqueimplementation of finger vein authentication technique
implementation of finger vein authentication techniqueViraj Rajopadhye
 
Using Ultraviolet and Infrared Imaging to Capture Fingerprints
Using Ultraviolet and Infrared Imaging to Capture Fingerprints Using Ultraviolet and Infrared Imaging to Capture Fingerprints
Using Ultraviolet and Infrared Imaging to Capture Fingerprints Darren Corbett
 
Biometrics/fingerprint sensors
Biometrics/fingerprint sensorsBiometrics/fingerprint sensors
Biometrics/fingerprint sensorsJeffrey Funk
 
Image receptors for radiology seminar
Image receptors   for radiology seminarImage receptors   for radiology seminar
Image receptors for radiology seminarPuneet Choudhary
 
Dental Radiation Protection .pptx
Dental Radiation Protection .pptxDental Radiation Protection .pptx
Dental Radiation Protection .pptxOsama Albedaiwi
 
Radiation safty & protection
Radiation safty & protectionRadiation safty & protection
Radiation safty & protectionMohammed Sa'ad
 
Introduction to DIGITAL IMAGE PROCESSING - DAY 1
Introduction to DIGITAL IMAGE PROCESSING - DAY 1Introduction to DIGITAL IMAGE PROCESSING - DAY 1
Introduction to DIGITAL IMAGE PROCESSING - DAY 1vijayanand Kandaswamy
 
Smart surveillance
Smart surveillanceSmart surveillance
Smart surveillancePuneet soni
 
Radiation: Effects and Dose Calculations
Radiation: Effects and Dose CalculationsRadiation: Effects and Dose Calculations
Radiation: Effects and Dose Calculationsrowelganzon
 

What's hot (20)

BRAIN FINGERPRINTING TECHNOLOGY
BRAIN FINGERPRINTING TECHNOLOGYBRAIN FINGERPRINTING TECHNOLOGY
BRAIN FINGERPRINTING TECHNOLOGY
 
Visual Product Identification For Blind Peoples
Visual Product Identification For Blind PeoplesVisual Product Identification For Blind Peoples
Visual Product Identification For Blind Peoples
 
Image Classification And Skin cancer detection
Image Classification And Skin cancer detectionImage Classification And Skin cancer detection
Image Classification And Skin cancer detection
 
Multi modal biometric system
Multi modal biometric systemMulti modal biometric system
Multi modal biometric system
 
SEPARATION OF OVERLAPPED FINGERPRINTS
SEPARATION OF OVERLAPPED FINGERPRINTSSEPARATION OF OVERLAPPED FINGERPRINTS
SEPARATION OF OVERLAPPED FINGERPRINTS
 
Pattern recognition palm print authentication system
Pattern recognition palm print authentication systemPattern recognition palm print authentication system
Pattern recognition palm print authentication system
 
implementation of finger vein authentication technique
implementation of finger vein authentication techniqueimplementation of finger vein authentication technique
implementation of finger vein authentication technique
 
Iris recognition
Iris recognitionIris recognition
Iris recognition
 
Using Ultraviolet and Infrared Imaging to Capture Fingerprints
Using Ultraviolet and Infrared Imaging to Capture Fingerprints Using Ultraviolet and Infrared Imaging to Capture Fingerprints
Using Ultraviolet and Infrared Imaging to Capture Fingerprints
 
Cone beam ct
Cone beam ctCone beam ct
Cone beam ct
 
Biometrics/fingerprint sensors
Biometrics/fingerprint sensorsBiometrics/fingerprint sensors
Biometrics/fingerprint sensors
 
Image receptors for radiology seminar
Image receptors   for radiology seminarImage receptors   for radiology seminar
Image receptors for radiology seminar
 
Dental Radiation Protection .pptx
Dental Radiation Protection .pptxDental Radiation Protection .pptx
Dental Radiation Protection .pptx
 
How Fingerprint Scanner Work
How Fingerprint Scanner WorkHow Fingerprint Scanner Work
How Fingerprint Scanner Work
 
POROSCOPY AND EDGEOSCOPY
POROSCOPY AND EDGEOSCOPYPOROSCOPY AND EDGEOSCOPY
POROSCOPY AND EDGEOSCOPY
 
Radiation safty & protection
Radiation safty & protectionRadiation safty & protection
Radiation safty & protection
 
Introduction to DIGITAL IMAGE PROCESSING - DAY 1
Introduction to DIGITAL IMAGE PROCESSING - DAY 1Introduction to DIGITAL IMAGE PROCESSING - DAY 1
Introduction to DIGITAL IMAGE PROCESSING - DAY 1
 
Brain fingerprinting
Brain fingerprintingBrain fingerprinting
Brain fingerprinting
 
Smart surveillance
Smart surveillanceSmart surveillance
Smart surveillance
 
Radiation: Effects and Dose Calculations
Radiation: Effects and Dose CalculationsRadiation: Effects and Dose Calculations
Radiation: Effects and Dose Calculations
 

Similar to DRI Basic

Chapter 5: Remote sensing
Chapter 5: Remote sensingChapter 5: Remote sensing
Chapter 5: Remote sensingShankar Gangaju
 
Advanced Photojournalism BMT III Sem.pptx
Advanced Photojournalism BMT III Sem.pptxAdvanced Photojournalism BMT III Sem.pptx
Advanced Photojournalism BMT III Sem.pptxBajracharyaSunilC
 
Ee 417 Senior Design
Ee 417 Senior DesignEe 417 Senior Design
Ee 417 Senior Designcrouchj1
 
The challenges of minimal illumination
The challenges of minimal illuminationThe challenges of minimal illumination
The challenges of minimal illuminationcnssources
 
RDT-112-PRELIM-LESSON-2-NOTES.docx
RDT-112-PRELIM-LESSON-2-NOTES.docxRDT-112-PRELIM-LESSON-2-NOTES.docx
RDT-112-PRELIM-LESSON-2-NOTES.docxJianSoliman2
 
Ambient light sensors
Ambient light sensorsAmbient light sensors
Ambient light sensorsDavid M.Blass
 
Computer Vision for Measurement & FR
Computer Vision for Measurement & FRComputer Vision for Measurement & FR
Computer Vision for Measurement & FRRekaNext Capital
 
Parameterized Image Filtering Using fuzzy Logic
Parameterized Image Filtering Using fuzzy LogicParameterized Image Filtering Using fuzzy Logic
Parameterized Image Filtering Using fuzzy LogicEditor IJCATR
 
Remote sensing
 Remote sensing Remote sensing
Remote sensingFidy Zegge
 
Testo 885 Super Resolution
Testo 885 Super ResolutionTesto 885 Super Resolution
Testo 885 Super Resolutionvideoborescope
 
IRJET- Object Recognization from Structural Information using Perception
IRJET- Object Recognization from Structural Information using PerceptionIRJET- Object Recognization from Structural Information using Perception
IRJET- Object Recognization from Structural Information using PerceptionIRJET Journal
 
Photography 101
Photography 101Photography 101
Photography 101Rica Camus
 

Similar to DRI Basic (20)

Chapter 5: Remote sensing
Chapter 5: Remote sensingChapter 5: Remote sensing
Chapter 5: Remote sensing
 
Advanced Photojournalism BMT III Sem.pptx
Advanced Photojournalism BMT III Sem.pptxAdvanced Photojournalism BMT III Sem.pptx
Advanced Photojournalism BMT III Sem.pptx
 
TMP100 U03: Exposure Theory
TMP100 U03: Exposure TheoryTMP100 U03: Exposure Theory
TMP100 U03: Exposure Theory
 
Ee 417 Senior Design
Ee 417 Senior DesignEe 417 Senior Design
Ee 417 Senior Design
 
Thermal imagingtech tn_0305-508
Thermal imagingtech tn_0305-508Thermal imagingtech tn_0305-508
Thermal imagingtech tn_0305-508
 
The challenges of minimal illumination
The challenges of minimal illuminationThe challenges of minimal illumination
The challenges of minimal illumination
 
Chemclubposter
ChemclubposterChemclubposter
Chemclubposter
 
Digital imaging
Digital imagingDigital imaging
Digital imaging
 
RDT-112-PRELIM-LESSON-2-NOTES.docx
RDT-112-PRELIM-LESSON-2-NOTES.docxRDT-112-PRELIM-LESSON-2-NOTES.docx
RDT-112-PRELIM-LESSON-2-NOTES.docx
 
Ambient light sensors
Ambient light sensorsAmbient light sensors
Ambient light sensors
 
Computer Vision for Measurement & FR
Computer Vision for Measurement & FRComputer Vision for Measurement & FR
Computer Vision for Measurement & FR
 
Testo 885-2
Testo 885-2Testo 885-2
Testo 885-2
 
Testo 882
Testo 882Testo 882
Testo 882
 
Parameterized Image Filtering Using fuzzy Logic
Parameterized Image Filtering Using fuzzy LogicParameterized Image Filtering Using fuzzy Logic
Parameterized Image Filtering Using fuzzy Logic
 
Remote sensing
 Remote sensing Remote sensing
Remote sensing
 
Testo 885 Super Resolution
Testo 885 Super ResolutionTesto 885 Super Resolution
Testo 885 Super Resolution
 
IRJET- Object Recognization from Structural Information using Perception
IRJET- Object Recognization from Structural Information using PerceptionIRJET- Object Recognization from Structural Information using Perception
IRJET- Object Recognization from Structural Information using Perception
 
Testo 885-2
Testo 885-2Testo 885-2
Testo 885-2
 
Photography 101
Photography 101Photography 101
Photography 101
 
09 David Wiles - (Media Cybernetics) Introduction to Image Analysis
09 David Wiles - (Media Cybernetics) Introduction to Image Analysis09 David Wiles - (Media Cybernetics) Introduction to Image Analysis
09 David Wiles - (Media Cybernetics) Introduction to Image Analysis
 

Recently uploaded

OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...Soham Mondal
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝soniya singh
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ
 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learningmisbanausheenparvam
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escortsranjana rawat
 
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfCCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfAsst.prof M.Gokilavani
 
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2RajaP95
 
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escortsranjana rawat
 
power system scada applications and uses
power system scada applications and usespower system scada applications and uses
power system scada applications and usesDevarapalliHaritha
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...srsj9000
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfAsst.prof M.Gokilavani
 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx959SahilShah
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Dr.Costas Sachpazis
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024hassan khalil
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidNikhilNagaraju
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxbritheesh05
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxwendy cai
 
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...ZTE
 
Microscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptxMicroscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptxpurnimasatapathy1234
 

Recently uploaded (20)

OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
OSVC_Meta-Data based Simulation Automation to overcome Verification Challenge...
 
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learning
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
 
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfCCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
 
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
 
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
 
power system scada applications and uses
power system scada applications and usespower system scada applications and uses
power system scada applications and uses
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfid
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptx
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptx
 
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
 
Microscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptxMicroscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptx
 

DRI Basic

  • 1. Detection, Recognition, and Identification (DRI)
  • 2. What is DRI? DRI is a universally accepted set of standards providing a means of measuring the distance whereby a thermal sensor can produce an image of a specific target. This takes into consideration many different criterion such as noise, array size, optical blur, lens depletion, aperture, atmospheric depletion, detector pitch and many more.
  • 3. Detection Detection means that you will be able to see the target, however it will be little more than a spec. Specifically it means that the target is visible on at least two pixels, and that there is a good chance that the target is actually something of suspicion. Detection: Ability to distinguish an object from the background. Recognition Recognition does not mean that you can recognize an individual. Recognition simply means that you are able to recognize an object’s class (is it a human or a car, is it a truck or a tank, etc). Identification Identification of an object means that you are able to differentiate between objects. For example, being able to identify the type of vehicle not just its class. Identification: ability to describe the object in details (a man with a hat, a deer, a Jeep.
  • 4. The following pictures illustrate these definitions: Left image: Detection – At several kms, 2 targets are detected out of the background Center image: Recognition - a human is walking along the fence Right image: Identification – 2 males with trousers and jackets are identified – one is smoking.
  • 5. • Thermal sensors have a resolution of 640×480 which is over 300,000 pixels. Human “detection” only requires 3.6 of those pixels and “identification” only requires 230 pixels, which is an extraordinarily small amount on the screen that can easily go unnoticed by the human eye.
  • 6. The Johnson criteria Johnson defined thresholds, known as the Johnson criteria, as the minimum number of line pairs to detect, recognize or identify targets captured by scene imagers. The Johnson criteria was based on experiments using observers. It estimates the number of line pairs required across a target, and indicates that there is at least a 50% chance that an observer can accurately achieve the criteria. The Johnson criteria state the following thresholds that are achieved 50% of the time: • Detection: We need 1.0 +/− 0.25 line pairs to detect an object 50% of the time. • Recognition: The type object can be discerned, a person versus a car (4 +/− 0.8 line pairs) • Identification: A specific object can be discerned, a woman versus a man, the specific car (6.4 +/− 1.5 line pairs)
  • 7. • He found that the minimum resolution required for a particular "decision response“ was nearly constant for a group of nine military targets. The results of these experiments are tabulated in Table 1. The data show that the minimum resolution required for a particular decision activity is a constant for nine military targets within a maximum error excursion of +25%.
  • 8. • The methodology developed by Johnson was simple and straight forward. A target was moved to a range where it was just barely detectable. A bar pattern was placed in the field of view and its spatial frequency was increased until it could barely be resolved at the same range, i.e. the number of lines on the bar pattern was increased until they could no longer be distinguished. The spatial frequency of the pattern was specified in terms of the number of lines in the pattern subtended by the object's minimum dimension as illustrated in Figure 1. The same methodology was used for orientation, recognition and identification.
  • 9. FACTORS INFLUENCING TARGET DETECTION AND RECOGNITION The scene 1. The size of the picture or displayed image. 2. Numbers, sizes, shapes, and scene distribution of areas contextually likely to contain the target object. 3. Scene objects: numbers, shapes and patterns, achromatic and color contrasts, colors, (hue, saturation, lightness), acutance, amount of resolved details, all both absolutely and relative to the target object. 4. Scene distribution of objects. 5. Granularity, noise. 6. Total available information content and amount of each type of information. 7. Average image brightness or lightness. 8. Contextual cues to target object location.
  • 10. The target 1. Location in the image format. 2. Location in the scene 3. Shape and pattern. 4. Size, color, resolution(s), acutance, lightness or brightness. 5. Type and degree of isolation from background and objects. The observer 1. Training 2. Experience 3. Native ability 4. Instructions and task briefing 5. Search habits 6. Motivation 7. Compromise on speed versus accuracy 8. Assumptions.
  • 11. Putting it all together If we ignore the effects of the atmosphere, normal recognition would be 25% of the detection and identification would be 12.5% of the detection range. Weather is almost never ideal so in reality these percentages are almost always reduced. “Ideal Conditions” Another thing that is often not mentioned is that these ratings are based on what is termed “ideal conditions” which rarely happen in the real world. The average environmental application will get 25% less than the distance that the thermal camera is rated for and in extreme conditions can be less than 10% of the rated distance.
  • 12. Thermal camera characteristics influencing detection range • With Johnson criteria in mind, it is easy to understand that the spatial resolution of a camera is an important parameter determining the quality of the security system. The spatial resolution: is the camera’s ability to capture very small details at great distances. It is closely related to the number of camera pixels. Clearly, the more pixels there are, the better the spatial resolution will be and the greater the detection range will be. • However, the thermal sensitivity has also a high impact on detection performance. The thermal sensitivity is the ability to distinguish very small differences in temperature. The “cooled” IR cameras offer better thermal sensitivity than “uncooled” IR cameras. For the same number of pixels, the image of a “cooled” thermal camera is of better quality (i.e. less “noisy”) than an “uncooled” camera’s image.
  • 13. Thermal camera characteristics influencing detection range • Left: image from a thermal imager with poor thermal sensitivity • Right: image from a cooled thermal camera with high thermal sensitivity
  • 14. Another important characteristic of an infrared camera is the field of view (FOV). A large panoramic FOV is necessary in order to enhance the probability to detect a single event into a complete 360-degree scene. See in the picture below the image captured by a thermal sensor with a wide panoramic horizontal FOV, and, in green in the center, the area screened by a PTZ camera zoomed to a 2° FOV. It is clear that this PTZ camera has no chance at all to detect any event in this scenery. Thermal camera characteristics influencing detection range
  • 15. Converting Line Pair to Pixels • How many line pairs do we need to detect something 100% of the time? First, let’s convert line pairs to pixels because the resolution of today’s cameras is specified in pixels rather than lines. • Line pair resolution is defined as at least one black and one white detection. Theoretically, you could have one pixel per line, but in reality, it requires more than 2 pixels. We need at least one pixel that detects black, a second that detects white, and a third that detects another black. At a minimum, we need at least 3 pixels to detect a small white object. However, there’s more to this. We need to take into account a certain amount of statistical uncertainty of a sensor picking up the signal.
  • 16. The Kell Factor In the real world of detection, the white spot could appear somewhere in between the pixels of the sensor. This means we need more pixels to actually detect the small object. The concept of the uncertainty of where an object appears on a sensor was introduced in 1934 by Raymond Kell. It was first applied to sampling an image to avoid beat frequencies affecting the image. The Kell factor is a fudge factor that attempts to adjust for the uncertainty of analog sensors. The number has changed depending on how it is used. Kell said the number should be 0.64, but today people usually use 0.7. The Kell factor increases the number of pixels required. We adjust the pixels using the formula: pixels/ kell factor , 3 / 0.7 = 4.29 vertical pixels.
  • 17. The Mesnik Factors  Besides the Kell factor, there are other real-world factors that can affect what we see. For example, long-range lenses sometimes do not have the clarity they require to accommodate the higher resolution sensors. Other factors include low-light operation and atmospheric conditions that affect visibility.  The atmospheric conditions can be very obvious like rain or fog or can be subtle like heat waves or higher humidity. The view can be reduced dramatically by fog, haze (smoke), rain, or snow. Unfortunately, the amount of reduced vision can vary depending on the intensity of fog or rain.  Long range cameras must also be mounted to very stable platforms. Placing a camera on top of a tower can cause vibration that affects the clarity of the video. Some long-range cameras include image stabilization. There are electronic and mechanical mechanisms for stabilizing the cameras.  We can adjust for some of these real-world visual problems. We have come up with the following factors that result in the need for increased pixel count for meeting the visual criteria.
  • 18. • Lens factor: From 1 to 0.6 depending on the actual lens clarity selected • Low light noise factor: Thermal cameras are affected by the difference in temperature between the object and the background. • Atmospheric Factors: Heat waves and other clear day factors such as humidity increase with distance. • Atmospheric Distance Factor. These are all estimates of atmospheric effects on visibility. Actual visibility is affected by the difference in contrast between the object and the background. For distance less than 1000 m (0.62 miles) = 1.0 for distance between 1001m – 2000 m (1.24 miles) = 0.9, for distance 2001m to 3000 m (1.86 miles) = 0.85, for distance over 3001 m to 5000 m = 0.8 for distance 5,001 m to 10,000 m = 0.75 for distance 10,001 m (6.2 miles) to 15000 m (9.3 miles) = 0.7 for distance 15001m to 20000 m (12.4 miles) = 0.65
  • 19. An example of using the factors: • The following example is a camera system used at night, that will observe objects about 2-miles away. We excluded severe atmospheric conditions such as fog that could reduce visibility dramatically. We calculated the pixel counts for a thermal camera with similar lens magnification. Reasonably good lens (costs a lot) = 0.9 Used at night: = 1.0 for thermal camera Distance = 3219 m (2 miles) = 0.8 – Total for Thermal = 0.9 x 1.0 x 0.8 = 0.72 – Total Pixels for Thermal Camera = 4.29 / 0.72 = 5.96 pixels Here’s the summary of the new DRI standard. Detection: 8 vertical pixels across the target Recognition: we require eight times the 4 line pairs or 32 vertical pixels across the target Identification: we need eight times the 8 line pairs = 64 vertical pixels across the target
  • 21. • Spatial frequency is typically measured per cycles (or line pairs) per a unit angle or a unit length
  • 22. Evaluation of ranges of effective surveillance • It is possible to measure directly detection, recognition, and identification ranges of a target of interest and to evaluate the tested thermal camera on the basis of the test results. However, it is a risky solution. The ranges vary with observation conditions (atmosphere, background) and it is relatively easy to manipulate with the detection, recognition and identification ranges at real conditions if the observation conditions are not very precisely specified. Next, it is difficult to compare test results of different thermal cameras tested at different time periods and at different observation conditions. • The safest way to evaluate surveillance thermal cameras is to measure MRTD of this thermal imager and calculate the detection, recognition, and identification ranges of a standard NATO target using methodology proposed by this standard: STANAG 4347. • The detection, recognition, and identification ranges of the standard NATO target are potentially a good criterion to be used in requirements on surveillance thermal imagers. It is apparently a good idea to present requirements on a surveillance thermal imager by presenting requirements on the performance ranges (Table)
  • 23. In order to calculate the performance ranges of a thermal imager we must know its MRTD function. Therefore a more common way to specify requirements on surveillance thermal cameras is to present requirements on MRTD characteristic in a form shown in Table.
  • 24. • Now, let us discuss the way to calculate the detection and recognition ranges on the basis of the measured and known MRTD function of the tested thermal camera using the recommendation from STANAG 4347. The summary of the recommendations is shown in Table.
  • 25. • In details, the detection, recognition and identification ranges of the standard NATO targets can be calculated using the below presented algorithm. 1. Convert MRTD characteristic into a new one by changing variable from spatial frequency n [mrad-1] to the range R [km] using the following formulas 2. Calculate decrease in the initial temperature difference DT (it was assumed that initial = = 2 K) between the target and the background due to limited atmospheric transmission.
  • 26. • Determine the respective nominal static ranges as the intersections of and the converted MRTD functions. calculation of the detection ranges of the standard NATO target using a thermal camera of MRTD function shown in Fig. The calculation results are shown in Fig. and we can conclude that the detection ranges are the following:  7.2 km at good atmospheric transmission.  2.8 km at limited atmospheric transmission. Calculation of the recognition and identification ranges can be done in the same manner.
  • 27.
  • 28. A comparison between the ACQUIRE Model and the Johnson Criteria assuming standard N50 values for each discrimination level. The distance difference is measured at the 50% probability of detect and compares the difference between the Johnson Criteria and the AQUIRE Metric. These values were calculated for a system viewing a target of 0.75 meter height, with an imaging system having a 10 micron pixel and a 10 millimeter focal length.
  • 29. Determine range for camouflaged targets • In 2004, the ACQUIRE-LC metric was introduced to determine range for camouflaged targets. Camouflage developers showed that camouflage did indeed have an impact on the range at which a target could be detected. New equations were set to govern the type of camouflage and amount of clutter in the scene. These can be seen in Equations.