SlideShare a Scribd company logo
1 of 26
Download to read offline
Presenter: Benyamin Moadab
Learning RGB-D Salient Object Detection
using
background enclosure, depth contrast, and top-down features
Supervisor: Dr. Vahid Rostami
Islamic Azad University of QAZVIN
IN THE NAME OF GOD
May - June 2021
Department of Electrical, Computer and Information Technology
1. Prerequisites
2. Introduction
3. Proposed architecture
4. RGB-D saliency detection system
5. Experimental Evaluation
6. Conclusion
CONTENTS
Prerequisites
Basic concepts and requirements in the field of image processing
Prerequisites
4
Top-down and Bottom-up information
Visual attention
Low-level Depth Features
High-level Depth Features
▪ Colors, textures, edges , bubbles, ….
▪ objects and events
Introduction
Some general information
Introduction
• In human visual saliency, top-down and bottom-up information are combined as a
basis of visual attention.
• The use of top-down information and the exploration of CNNs for RGB-D salient
object detection remains limited.
• Saliency can be used in the context of many computer vision problems such as
compression, object detection, visual tracking and etc.
Proposed architecture
In this section, we introduce our approach to
RGB-D salient object detection.
We combine the strengths of previous approaches to high-level and low-level feature
based deep CNN RGB salient object detection
top-down
and
bottom-up methods
Mid
level depth feature
low
level features
Proposed architecture
8
This paper makes three major contributions:
RGB-D saliency detection
system
RGB-D saliency detection system
10
2
3
5 Concatenation of
Color and Depth
Features
RGB low and high level
saliency from ELD-Net
BED Feature
• As in the previous section, features to identify the protrusion of the object introduced in this section,
each of them are introduced in detail, which are:
1
4
Low level
Depth Features
Nonlinear combination
of depth features
.
1 BED Feature
11
✓ BED is inspired by LBE
✓ LBE
✓ BED
✓ BED relies on learning
✓ BED feature captures
the enclosure distribution
properties of a patch.
❖ How the BED feature works
1) Descriptive
2) Math
𝑓 𝑃,𝑡 =
𝜃𝑎1+𝜃𝑎2
2𝜋
‫و‬ g P,t =
𝜃𝑎3
2𝜋
2. Low-level Depth Features
12
✓ Depth contrast
✓ Depth histogram distance
✓ SLIC Algorithm
✓ focused superpixel
✓ Procedures:
1) Calculate the average depth value
2) subdivide the image grid cells
3) subtracting the average depth value of each grid cell
from the average depth value for each super-pixel
4) calculate the difference between the depth histogram
of the focused super-pixel
5) divide the entire range of depth values into 8 intervals
3. RGB low and high level saliency from ELDNet
13
✓ To represent high-level and low-level features for RGB,
we make use of the extended version of ELD-Net
▪ VGG-Net
▪ GoogleNet
VGG Net
GoogleNet
4. Nonlinear combination of depth features
14
✓ As described in the previous sections on the low-level feature and the BED feature, objects must be combined to detect
protrusions.
✓ In order to better combine these features, we use three layers of convolution to create depth feature outputs.
5. Concatenation of Color and Depth Features
15
1) we make use of the pretrained caffe-model of ELD-Net.
2) The calculated 1 × 20 × 20 color feature layer is concatenated with the depth feature outputs.
3) We then connect the 1×20×20+1×20×20 concatenated output features with two fully connected layers and
calculate the saliency score for the focused super-pixel.
4) We calculate the cross entropy loss for a softmax classifier to evaluate the outputs as:
5.1. Preprocessing on depth and color images
16
✓ synchronize the scale of depth values with color values.
✓ Normalize the depth value to the same scale 0 to 255.
• Depth values of RGBD1000 are represented with greater bit depth and so require normalization.
• On NJUDS2000 the scale of depth values are already 0- 255, and so are not modified.
✓ After normalization, we resize the color and depth images to 324×324.
➢ The following figures show these processes:
Superpixel Segmentation
17
✓ We use gSLICr ,the GPU version of SLIC, to segment the images into super-pixels.
✓ We divide each image into approximately 18 × 18 super-pixels,
Experimental Evaluation
Experimental Evaluation
❖ We evaluate our architecture’s performance on two datasets:
1) RGBD1000 (600 =training set, 200 =validation set, 200 = test)
2) NJUDS2000 (1200 =training set, 385= validation set, 400 =test)
❖The results are compared against other state-of-the-art RGB-D saliency detection methods:
1) local background enclosure (LBE)
2) multi-scale depth-contrast (LMH)
3) saliency based on region contrast and background, depth, and an orientation prior (GP)
4) anisotropic center surround depth based saliency (ACSD)
5) discriminative regional feature integration approach (DRFI)
6) Saliency detection via dense and sparse reconstruction (DSR )
7) detection approaches: saliency from low and high level features (ELD)
8) the Deep hierarchical saliency network (DHS)
9) CNN-based RGB-D prominent object detection method (DF)
Evaluation Criteria
1) precision-recall curve
2) mean F-score
21
Discussion
Table 1. Comparing average F-measure score
with other state-of-the-art saliency methods on two datasets.
Table 2. Replacing the super-pixel histogram with mean depth improves results for
NJUDS2000 where depth data is noisy.
Table 3. Comparing scores with different input features on
RGBD1000 .Note that LD means Low-level Depth Features.
Table 4. Comparing scores with different input features on
NJUDS2000 .Note that LD means Low-level Depth Features.
Discussion
Conclusion
Conclusion
24
✓ As can be seen in the results, the proposed method was based on CNN and using the combination of features:
background space, contrast depth and top-down and bottom-up information achieved the best result in comparison
with other available methods.
Performance improvements (My suggestion)
25
▪ Because the architectures used in high-level feature recognition, VGG and Google Net, have disadvantages
(such as the following), it is recommended to use other architectures such as YOLO or RCNN to improve
speed and better performance.
Now Any Question??
FINISH
Email: Benyamin-moadab@qiau.ac.ir
Benyamin Moadab
Master of Computer Engineering (Artificial Intelligence and Robotics), Qazvin Branch, Islamic Azad University
Thank you for your attention

More Related Content

What's hot

COMPUTER GRAPHICS
COMPUTER GRAPHICSCOMPUTER GRAPHICS
COMPUTER GRAPHICSJagan Raja
 
illumination model in Computer Graphics by irru pychukar
illumination model in Computer Graphics by irru pychukarillumination model in Computer Graphics by irru pychukar
illumination model in Computer Graphics by irru pychukarsyedArr
 
Python in Computer Vision
Python in Computer VisionPython in Computer Vision
Python in Computer VisionBrian Thorne
 
Requirement analysis and specification
Requirement analysis and specificationRequirement analysis and specification
Requirement analysis and specificationM.E. at GTU- PG School
 
Future Directions for Compute-for-Graphics
Future Directions for Compute-for-GraphicsFuture Directions for Compute-for-Graphics
Future Directions for Compute-for-GraphicsElectronic Arts / DICE
 
Import and Export Big Data using R Studio
Import and Export Big Data using R StudioImport and Export Big Data using R Studio
Import and Export Big Data using R StudioRupak Roy
 
Migrating from OpenGL to Vulkan
Migrating from OpenGL to VulkanMigrating from OpenGL to Vulkan
Migrating from OpenGL to VulkanMark Kilgard
 
Chapter 2 - Matlab Environment
Chapter 2 - Matlab EnvironmentChapter 2 - Matlab Environment
Chapter 2 - Matlab EnvironmentSiva Gopal
 
Introduction to Image Processing with MATLAB
Introduction to Image Processing with MATLABIntroduction to Image Processing with MATLAB
Introduction to Image Processing with MATLABSriram Emarose
 
[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원
[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원
[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원강 민우
 
Color Image Processing
Color Image ProcessingColor Image Processing
Color Image Processingkiruthiammu
 
Software Architecture – Centric Methods and Agile Development
Software Architecture –   Centric Methods and   Agile DevelopmentSoftware Architecture –   Centric Methods and   Agile Development
Software Architecture – Centric Methods and Agile Developmentsathish sak
 
INCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP exam
INCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP examINCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP exam
INCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP examsystemsengineeringprep
 
Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...
Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...
Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...Unity Technologies
 

What's hot (20)

COMPUTER GRAPHICS
COMPUTER GRAPHICSCOMPUTER GRAPHICS
COMPUTER GRAPHICS
 
illumination model in Computer Graphics by irru pychukar
illumination model in Computer Graphics by irru pychukarillumination model in Computer Graphics by irru pychukar
illumination model in Computer Graphics by irru pychukar
 
Python in Computer Vision
Python in Computer VisionPython in Computer Vision
Python in Computer Vision
 
Requirement analysis and specification
Requirement analysis and specificationRequirement analysis and specification
Requirement analysis and specification
 
Future Directions for Compute-for-Graphics
Future Directions for Compute-for-GraphicsFuture Directions for Compute-for-Graphics
Future Directions for Compute-for-Graphics
 
Video coding standards ppt
Video coding standards pptVideo coding standards ppt
Video coding standards ppt
 
Import and Export Big Data using R Studio
Import and Export Big Data using R StudioImport and Export Big Data using R Studio
Import and Export Big Data using R Studio
 
Migrating from OpenGL to Vulkan
Migrating from OpenGL to VulkanMigrating from OpenGL to Vulkan
Migrating from OpenGL to Vulkan
 
Chapter 2 - Matlab Environment
Chapter 2 - Matlab EnvironmentChapter 2 - Matlab Environment
Chapter 2 - Matlab Environment
 
Digital Image Fundamentals - II
Digital Image Fundamentals - IIDigital Image Fundamentals - II
Digital Image Fundamentals - II
 
HEVC overview main
HEVC overview mainHEVC overview main
HEVC overview main
 
Introduction to Image Processing with MATLAB
Introduction to Image Processing with MATLABIntroduction to Image Processing with MATLAB
Introduction to Image Processing with MATLAB
 
H.264 vs HEVC
H.264 vs HEVCH.264 vs HEVC
H.264 vs HEVC
 
[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원
[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원
[IGC2018] 에픽게임즈 신광섭 - 언리얼엔진4 포트나이트 멀티플랫폼 개발 지원
 
Color Image Processing
Color Image ProcessingColor Image Processing
Color Image Processing
 
Software Architecture – Centric Methods and Agile Development
Software Architecture –   Centric Methods and   Agile DevelopmentSoftware Architecture –   Centric Methods and   Agile Development
Software Architecture – Centric Methods and Agile Development
 
Spatial filtering
Spatial filteringSpatial filtering
Spatial filtering
 
INCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP exam
INCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP examINCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP exam
INCOSE Systems Engineering Handbook and Changes to the CSEP/ASEP exam
 
Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...
Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...
Developing and optimizing a procedural game: The Elder Scrolls Blades- Unite ...
 
Watershed
WatershedWatershed
Watershed
 

Similar to Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top down features

Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors IIIYu Huang
 
Depth Fusion from RGB and Depth Sensors by Deep Learning
Depth Fusion from RGB and Depth Sensors by Deep LearningDepth Fusion from RGB and Depth Sensors by Deep Learning
Depth Fusion from RGB and Depth Sensors by Deep LearningYu Huang
 
Single Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learningSingle Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learningAhan M R
 
Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...
Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...
Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...Savvas Chatzichristofis
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IVYu Huang
 
A Robust Image Watermarking Technique using Luminance Based Area Selection an...
A Robust Image Watermarking Technique using Luminance Based Area Selection an...A Robust Image Watermarking Technique using Luminance Based Area Selection an...
A Robust Image Watermarking Technique using Luminance Based Area Selection an...IRJET Journal
 
Color and 3D Semantic Reconstruction of Indoor Scenes from RGB-D stream
Color and 3D Semantic Reconstruction of Indoor Scenes from RGB-D streamColor and 3D Semantic Reconstruction of Indoor Scenes from RGB-D stream
Color and 3D Semantic Reconstruction of Indoor Scenes from RGB-D streamNAVER Engineering
 
Introduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable RenderingIntroduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable RenderingPreferred Networks
 
Object Based Image Analysis
Object Based Image Analysis Object Based Image Analysis
Object Based Image Analysis Kabir Uddin
 
Efficient LDI Representation (TPCG 2008)
Efficient LDI Representation (TPCG 2008)Efficient LDI Representation (TPCG 2008)
Efficient LDI Representation (TPCG 2008)Matthias Trapp
 
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)Universitat Politècnica de Catalunya
 
Image pcm 09
Image pcm 09Image pcm 09
Image pcm 09anilbl
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving IIYu Huang
 
Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...
Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...
Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...NopphawanTamkuan
 
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis taeseon ryu
 

Similar to Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top down features (20)

Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors III
 
Depth Fusion from RGB and Depth Sensors by Deep Learning
Depth Fusion from RGB and Depth Sensors by Deep LearningDepth Fusion from RGB and Depth Sensors by Deep Learning
Depth Fusion from RGB and Depth Sensors by Deep Learning
 
Single Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learningSingle Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learning
 
Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...
Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...
Searching Images with MPEG-7 (& MPEG-7 Like) Powered Localized dEscriptors (S...
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IV
 
WT in IP.ppt
WT in IP.pptWT in IP.ppt
WT in IP.ppt
 
Depth estimation using deep learning
Depth estimation using deep learningDepth estimation using deep learning
Depth estimation using deep learning
 
A Robust Image Watermarking Technique using Luminance Based Area Selection an...
A Robust Image Watermarking Technique using Luminance Based Area Selection an...A Robust Image Watermarking Technique using Luminance Based Area Selection an...
A Robust Image Watermarking Technique using Luminance Based Area Selection an...
 
Color and 3D Semantic Reconstruction of Indoor Scenes from RGB-D stream
Color and 3D Semantic Reconstruction of Indoor Scenes from RGB-D streamColor and 3D Semantic Reconstruction of Indoor Scenes from RGB-D stream
Color and 3D Semantic Reconstruction of Indoor Scenes from RGB-D stream
 
Introduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable RenderingIntroduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable Rendering
 
Object Based Image Analysis
Object Based Image Analysis Object Based Image Analysis
Object Based Image Analysis
 
Efficient LDI Representation (TPCG 2008)
Efficient LDI Representation (TPCG 2008)Efficient LDI Representation (TPCG 2008)
Efficient LDI Representation (TPCG 2008)
 
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
 
Image pcm 09
Image pcm 09Image pcm 09
Image pcm 09
 
regions
regionsregions
regions
 
Charmi chokshi ppt
Charmi chokshi pptCharmi chokshi ppt
Charmi chokshi ppt
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II
 
Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...
Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...
Data Processing Using THEOS Satellite Imagery for Disaster Monitoring (Case S...
 
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
 
PhD_ppt_2012
PhD_ppt_2012PhD_ppt_2012
PhD_ppt_2012
 

Recently uploaded

Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxVishalSingh1417
 
Web & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfWeb & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfJayanti Pande
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Celine George
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajanpragatimahajan3
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Disha Kariya
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingTeacherCyreneCayanan
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...Sapna Thakur
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfchloefrazer622
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingTechSoup
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactPECB
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfAdmir Softic
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfchloefrazer622
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
Student login on Anyboli platform.helpin
Student login on Anyboli platform.helpinStudent login on Anyboli platform.helpin
Student login on Anyboli platform.helpinRaunakKeshri1
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 

Recently uploaded (20)

Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
Web & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfWeb & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdf
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajan
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writing
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdf
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdf
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Student login on Anyboli platform.helpin
Student login on Anyboli platform.helpinStudent login on Anyboli platform.helpin
Student login on Anyboli platform.helpin
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 

Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top down features

  • 1. Presenter: Benyamin Moadab Learning RGB-D Salient Object Detection using background enclosure, depth contrast, and top-down features Supervisor: Dr. Vahid Rostami Islamic Azad University of QAZVIN IN THE NAME OF GOD May - June 2021 Department of Electrical, Computer and Information Technology
  • 2. 1. Prerequisites 2. Introduction 3. Proposed architecture 4. RGB-D saliency detection system 5. Experimental Evaluation 6. Conclusion CONTENTS
  • 3. Prerequisites Basic concepts and requirements in the field of image processing
  • 4. Prerequisites 4 Top-down and Bottom-up information Visual attention Low-level Depth Features High-level Depth Features ▪ Colors, textures, edges , bubbles, …. ▪ objects and events
  • 6. Introduction • In human visual saliency, top-down and bottom-up information are combined as a basis of visual attention. • The use of top-down information and the exploration of CNNs for RGB-D salient object detection remains limited. • Saliency can be used in the context of many computer vision problems such as compression, object detection, visual tracking and etc.
  • 8. In this section, we introduce our approach to RGB-D salient object detection. We combine the strengths of previous approaches to high-level and low-level feature based deep CNN RGB salient object detection top-down and bottom-up methods Mid level depth feature low level features Proposed architecture 8 This paper makes three major contributions:
  • 10. RGB-D saliency detection system 10 2 3 5 Concatenation of Color and Depth Features RGB low and high level saliency from ELD-Net BED Feature • As in the previous section, features to identify the protrusion of the object introduced in this section, each of them are introduced in detail, which are: 1 4 Low level Depth Features Nonlinear combination of depth features
  • 11. . 1 BED Feature 11 ✓ BED is inspired by LBE ✓ LBE ✓ BED ✓ BED relies on learning ✓ BED feature captures the enclosure distribution properties of a patch. ❖ How the BED feature works 1) Descriptive 2) Math 𝑓 𝑃,𝑡 = 𝜃𝑎1+𝜃𝑎2 2𝜋 ‫و‬ g P,t = 𝜃𝑎3 2𝜋
  • 12. 2. Low-level Depth Features 12 ✓ Depth contrast ✓ Depth histogram distance ✓ SLIC Algorithm ✓ focused superpixel ✓ Procedures: 1) Calculate the average depth value 2) subdivide the image grid cells 3) subtracting the average depth value of each grid cell from the average depth value for each super-pixel 4) calculate the difference between the depth histogram of the focused super-pixel 5) divide the entire range of depth values into 8 intervals
  • 13. 3. RGB low and high level saliency from ELDNet 13 ✓ To represent high-level and low-level features for RGB, we make use of the extended version of ELD-Net ▪ VGG-Net ▪ GoogleNet VGG Net GoogleNet
  • 14. 4. Nonlinear combination of depth features 14 ✓ As described in the previous sections on the low-level feature and the BED feature, objects must be combined to detect protrusions. ✓ In order to better combine these features, we use three layers of convolution to create depth feature outputs.
  • 15. 5. Concatenation of Color and Depth Features 15 1) we make use of the pretrained caffe-model of ELD-Net. 2) The calculated 1 × 20 × 20 color feature layer is concatenated with the depth feature outputs. 3) We then connect the 1×20×20+1×20×20 concatenated output features with two fully connected layers and calculate the saliency score for the focused super-pixel. 4) We calculate the cross entropy loss for a softmax classifier to evaluate the outputs as:
  • 16. 5.1. Preprocessing on depth and color images 16 ✓ synchronize the scale of depth values with color values. ✓ Normalize the depth value to the same scale 0 to 255. • Depth values of RGBD1000 are represented with greater bit depth and so require normalization. • On NJUDS2000 the scale of depth values are already 0- 255, and so are not modified. ✓ After normalization, we resize the color and depth images to 324×324. ➢ The following figures show these processes:
  • 17. Superpixel Segmentation 17 ✓ We use gSLICr ,the GPU version of SLIC, to segment the images into super-pixels. ✓ We divide each image into approximately 18 × 18 super-pixels,
  • 19. Experimental Evaluation ❖ We evaluate our architecture’s performance on two datasets: 1) RGBD1000 (600 =training set, 200 =validation set, 200 = test) 2) NJUDS2000 (1200 =training set, 385= validation set, 400 =test) ❖The results are compared against other state-of-the-art RGB-D saliency detection methods: 1) local background enclosure (LBE) 2) multi-scale depth-contrast (LMH) 3) saliency based on region contrast and background, depth, and an orientation prior (GP) 4) anisotropic center surround depth based saliency (ACSD) 5) discriminative regional feature integration approach (DRFI) 6) Saliency detection via dense and sparse reconstruction (DSR ) 7) detection approaches: saliency from low and high level features (ELD) 8) the Deep hierarchical saliency network (DHS) 9) CNN-based RGB-D prominent object detection method (DF)
  • 21. 21 Discussion Table 1. Comparing average F-measure score with other state-of-the-art saliency methods on two datasets. Table 2. Replacing the super-pixel histogram with mean depth improves results for NJUDS2000 where depth data is noisy. Table 3. Comparing scores with different input features on RGBD1000 .Note that LD means Low-level Depth Features. Table 4. Comparing scores with different input features on NJUDS2000 .Note that LD means Low-level Depth Features.
  • 24. Conclusion 24 ✓ As can be seen in the results, the proposed method was based on CNN and using the combination of features: background space, contrast depth and top-down and bottom-up information achieved the best result in comparison with other available methods.
  • 25. Performance improvements (My suggestion) 25 ▪ Because the architectures used in high-level feature recognition, VGG and Google Net, have disadvantages (such as the following), it is recommended to use other architectures such as YOLO or RCNN to improve speed and better performance. Now Any Question??
  • 26. FINISH Email: Benyamin-moadab@qiau.ac.ir Benyamin Moadab Master of Computer Engineering (Artificial Intelligence and Robotics), Qazvin Branch, Islamic Azad University Thank you for your attention