SlideShare a Scribd company logo
Neural Radiance Fields
& Neural Rendering
Mildenhall, Ben, et al. "Nerf: Representing scenes as neural radiance fields
for view synthesis." Communications of the ACM 65.1 (2021): 99-106.
Navneet Paul
PlayerUnknown Productions
Rendering
● Process of generating an image from a 2D or 3D model using a computer program. The
resulting image is called render.
● A rendering application takes into account inputs such as the model (2D/3D), texture,
shading, lighting, viewpoints, etc, as features during the rendering process.
● It means we can say that each scene file contains multiple features that need to be
understood and processed by the rendering algorithm or application to generate a
processed image.
Rendering Equation
● The rendering algorithm or technique which tries to solve the problem of image
generation based on all given features is mostly trying to optimize the rendering equation.
● At a high-level the rendering equation computes for a radiance that is illumination
(reflection, refraction, and emittance of light) on an object from a source to an observer in
a given space.
● NeRF essentially computes for the Volume Rendering.
Volume Rendering
● Volume rendering (as per Wikipedia) is a set of a technique used to display a 2D projection of a
3D discretely sampled dataset.
● To render a 2D projection (output) of a 3D dataset we first need to define the camera position in
space relative to the volume then we need to define the RGBα (Red, Green. Blue, Alpha → it
stands for opacity channel) for every voxel.
● The primary objective in volume rendering is to get a transfer function which defines RGBα for
every value for every possible voxel value in a given space.
View Synthesis
● Click photos of an object from multiple camera angles and superimpose the images to
have a look at the same object from different known camera angles and positions.
● For NeRF, we are trying to predict the third missing axis (the first two being length &
breadth) which is the depth.
● Core application of NeRF: to predict a function for depth determination at various points
in the plane against the object itself.
Neural Radiance Fields
● Generate novel views of complex scenes by optimizing an underlying continuous
volumetric scene function using a sparse set of input views.
● The input can be provided as a blender model or a static set of images.
● The input is provided as a continuous 5D function that outputs the radiance emitted in
each direction (θ; Φ) at each point (x; y; z) in space, and a density at each point which
acts like a differential opacity controlling how much radiance is accumulated by a ray
passing through (x; y; z)
● A continuous scene can be described as a 5D vector-valued function whose input is a 3D
location x = (x; y; z) and 2D viewing direction (θ; Φ), and whose output is an emitted color
c = (r; g; b) and volume density (𝜎)
Volume
Rendering
Final rendering
MLP network F𝚯
: (x,y,z,d) →(RGB,𝜎)
Process overview
To generate a Neural Radiance Field from a particular viewpoint following steps were done:
● March camera rays through the scene to generate a sampled set of 3D points (Use either
COLMAP* or SfM for generating camera poses and viewing directions).
● Use those points and their corresponding 2D viewing directions as input to the neural network to
produce an output set of colors (RGB) and densities (𝜎)
● Use classical volume rendering approach to accumulate those colors and densities into a 2D
image
* a general-purpose Structure-from-Motion (SfM) and Multi-View Stereo (MVS) pipeline with a graphical and command-line interface
Network Architecture
● NeRF is an implicit Multi Layer Perceptron (MLP) based model that maps 5D vectors (3D
coordinates plus 2D viewing directions) to output RGB feature vector (c) & volume density (𝜎)
at that spatial location, using fully connected deep networks.
→ : layers with ReLU activation, 𝛾(x) : positional encoding, 𝛾(d) : directional encoding → : layer with no activation,
⇢ : layers with sigmoid activation, + : vector concatenation
Volume Rendering
● The authors used discrete data samples to estimate the expected color C(r) of camera ray r(t) with
the quadrature rule in classical volume rendering techniques.
Predicted colors
Volume density
Opacity
NeRF Optimization - Positional Encoding
● Previous studies show that optimizing the inputs to a higher dimensional space using high frequency functions
before passing them to the network enables better fitting of data that contains high frequency variation.
● Positional & Directional Encoding: A Fourier based feature mapping function that encodes features (pertaining
to position & direction) from lower dimensional space to a higher dimensional space.
Positional Encoding func.
Tancik, Srinivasan, Mildenhall et al., Fourier Features Let Networks Learn High Frequency
Functions in Low Dimensional Domains, NeurIPS 2020
No
Positional
Encoding
With
Positional
Encoding
NeRF Optimization - Hierarchical Sampling
● During the volume rendering phase, our model simultaneously optimizes two networks: coarse and fine
● We first sample a set of NC
locations with the RGB feature vector and density [σ (t)] outputs from the proposed
NeRF model, using stratified sampling, and evaluate the “coarse” network at these locations.
● The main function of coarse network is to compute the final rendered color of the ray for the coarse samples.
● a second set of Nf
locations are sampled from the [RGB + density] distribution using inverse transform sampling &
evaluate our “fine” network.
● All the samples are considered while computing the final rendered ray color, i.e, (NC
+ Nf
), at fine network stage.
This is done to ensure that more samples are allocated to regions we expect to contain visible content.
Final Rendering & Loss Function
● Optimize a separate neural continuous volume representation network, for each scene.
● At each optimization iteration, we randomly sample a batch of camera rays from the set of
all pixels in the dataset, and then follow the hierarchical sampling.
● NC
samples from the coarse network and NC
+ Nf
samples from the fine network.
● We then use the volume rendering procedure to render the color of each ray from both
sets of samples.
● Loss function is based on the the total squared error between the rendered and true pixel
colors for both the coarse and fine samples.
ℛ: set of rays in each batch; C(r): ground truth, ĈC
(r): coarse volume prediction and Ĉf
(r): fine volume prediction for RGB colors for ray “r”
Performance of NeRF
Comparison to other view synthesis techniques
● Neural Volumes, Local Light Field Fusion (LLFF)
& Scene Representation Networks (SRN
(Ours = NeRF)
Performance of NeRF
Ablation Studies
● To validate the model’s performance with respect to different parameters.
Summary
● Learn the radiance field of a scene based on a
collection of calibrated images
○ Use an MLP to learn continuous
geometry and view-dependent
appearance
● Use fully differentiable volume rendering with
reconstruction loss
● Combines hierarchical sampling and
Fourier-based encoding of 5D inputs to produce
high-fidelity novel view synthesis results
Some associated challenges
● Handling dynamic scenes when acquiring
calibrated views
● One network trained per scene - no
generalization
Related NeRF Research
● NeRF in Wild: a novel approach for 3D scene reconstruction of complex environments from unstructured
internet photo collections that incorporates transient and latent scene embedding upon conventional NeRF
model.
*Martin-Brualla, Ricardo, et al. "Nerf in the wild: Neural radiance fields for unconstrained photo collections." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021.
● The model captures lighting and photometric variations in a low-dimensional latent embedding space in
rendering appearance without affecting 3D geometry.
● Neural Radiance Fields for Dynamic Scenes : for synthesizing novel views, at an arbitrary point in
time, of dynamic scenes with complex non-rigid geometries.
● Optimize an underlying deformable volumetric function (using a deformation network) from a sparse set
of input monocular views without the need of ground-truth geometry nor multi-view images
Pumarola, Albert, et al. "D-nerf: Neural radiance fields for dynamic scenes." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021.
Related NeRF Research

More Related Content

What's hot

Computer Vision Structure from motion
Computer Vision Structure from motionComputer Vision Structure from motion
Computer Vision Structure from motion
Wael Badawy
 
Introduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable RenderingIntroduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable Rendering
Preferred Networks
 
Semantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network ApproachesSemantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network Approaches
Fellowship at Vodafone FutureLab
 
Depth estimation using deep learning
Depth estimation using deep learningDepth estimation using deep learning
Depth estimation using deep learning
University of Oklahoma
 
Image Segmentation Using Deep Learning : A survey
Image Segmentation Using Deep Learning : A surveyImage Segmentation Using Deep Learning : A survey
Image Segmentation Using Deep Learning : A survey
NUPUR YADAV
 
CNN Machine learning DeepLearning
CNN Machine learning DeepLearningCNN Machine learning DeepLearning
CNN Machine learning DeepLearning
Abhishek Sharma
 
[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
Deep Learning JP
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II
Yu Huang
 
Super resolution in deep learning era - Jaejun Yoo
Super resolution in deep learning era - Jaejun YooSuper resolution in deep learning era - Jaejun Yoo
Super resolution in deep learning era - Jaejun Yoo
JaeJun Yoo
 
Depth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors IIDepth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors II
Yu Huang
 
Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling
Yu Huang
 
Lec15 sfm
Lec15 sfmLec15 sfm
Lec15 sfm
BaliThorat1
 
Real-time lightmap baking
Real-time lightmap bakingReal-time lightmap baking
Real-time lightmap baking
Rosario Leonardi
 
Filmic Tonemapping for Real-time Rendering - Siggraph 2010 Color Course
Filmic Tonemapping for Real-time Rendering - Siggraph 2010 Color CourseFilmic Tonemapping for Real-time Rendering - Siggraph 2010 Color Course
Filmic Tonemapping for Real-time Rendering - Siggraph 2010 Color Course
hpduiker
 
Object Pose Estimation
Object Pose EstimationObject Pose Estimation
Object Pose Estimation
Arithmer Inc.
 
Compute shader
Compute shaderCompute shader
Compute shader
QooJuice
 
Super Resolution
Super ResolutionSuper Resolution
Super Resolution
alokahuti
 
【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】
【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】
【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】
YutaSuzuki27
 
SfMLearner++ Intro
 SfMLearner++ Intro SfMLearner++ Intro
SfMLearner++ Intro
Hirohito Okuda
 

What's hot (20)

Computer Vision Structure from motion
Computer Vision Structure from motionComputer Vision Structure from motion
Computer Vision Structure from motion
 
Introduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable RenderingIntroduction to 3D Computer Vision and Differentiable Rendering
Introduction to 3D Computer Vision and Differentiable Rendering
 
Semantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network ApproachesSemantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network Approaches
 
Depth estimation using deep learning
Depth estimation using deep learningDepth estimation using deep learning
Depth estimation using deep learning
 
Image Segmentation Using Deep Learning : A survey
Image Segmentation Using Deep Learning : A surveyImage Segmentation Using Deep Learning : A survey
Image Segmentation Using Deep Learning : A survey
 
CNN Machine learning DeepLearning
CNN Machine learning DeepLearningCNN Machine learning DeepLearning
CNN Machine learning DeepLearning
 
[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
[DL輪読会]NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II
 
Super resolution from a single image
Super resolution from a single imageSuper resolution from a single image
Super resolution from a single image
 
Super resolution in deep learning era - Jaejun Yoo
Super resolution in deep learning era - Jaejun YooSuper resolution in deep learning era - Jaejun Yoo
Super resolution in deep learning era - Jaejun Yoo
 
Depth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors IIDepth Fusion from RGB and Depth Sensors II
Depth Fusion from RGB and Depth Sensors II
 
Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling
 
Lec15 sfm
Lec15 sfmLec15 sfm
Lec15 sfm
 
Real-time lightmap baking
Real-time lightmap bakingReal-time lightmap baking
Real-time lightmap baking
 
Filmic Tonemapping for Real-time Rendering - Siggraph 2010 Color Course
Filmic Tonemapping for Real-time Rendering - Siggraph 2010 Color CourseFilmic Tonemapping for Real-time Rendering - Siggraph 2010 Color Course
Filmic Tonemapping for Real-time Rendering - Siggraph 2010 Color Course
 
Object Pose Estimation
Object Pose EstimationObject Pose Estimation
Object Pose Estimation
 
Compute shader
Compute shaderCompute shader
Compute shader
 
Super Resolution
Super ResolutionSuper Resolution
Super Resolution
 
【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】
【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】
【学会発表】U-Net++とSE-Netを統合した画像セグメンテーションのための転移学習モデル【IBIS2020】
 
SfMLearner++ Intro
 SfMLearner++ Intro SfMLearner++ Intro
SfMLearner++ Intro
 

Similar to Neural Radiance Fields & Neural Rendering.pdf

TransNeRF
TransNeRFTransNeRF
TransNeRF
NavneetPaul2
 
Presentation_ Stable View Synthesis.pptx
Presentation_ Stable View Synthesis.pptxPresentation_ Stable View Synthesis.pptx
Presentation_ Stable View Synthesis.pptx
daptivuse
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IV
Yu Huang
 
Shadow Techniques for Real-Time and Interactive Applications
Shadow Techniques for Real-Time and Interactive ApplicationsShadow Techniques for Real-Time and Interactive Applications
Shadow Techniques for Real-Time and Interactive Applications
stefan_b
 
Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3
Slide_N
 
Chapter 4 better.pptx
Chapter 4 better.pptxChapter 4 better.pptx
Chapter 4 better.pptx
AbanobZakaria1
 
Phong Shading over any Polygonal Surface
Phong Shading over any Polygonal Surface Phong Shading over any Polygonal Surface
Phong Shading over any Polygonal Surface
Bhuvnesh Pratap
 
The single image dehazing based on efficient transmission estimation
The single image dehazing based on efficient transmission estimationThe single image dehazing based on efficient transmission estimation
The single image dehazing based on efficient transmission estimation
AVVENIRE TECHNOLOGIES
 
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNN
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNNAutomatic Detection of Window Regions in Indoor Point Clouds Using R-CNN
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNNZihao(Gerald) Zhang
 
Advanced Lighting for Interactive Applications
Advanced Lighting for Interactive ApplicationsAdvanced Lighting for Interactive Applications
Advanced Lighting for Interactive Applications
stefan_b
 
regions
regionsregions
regions
mjbahmani
 
Single Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learningSingle Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learning
Ahan M R
 
[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...
[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...
[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...
Susang Kim
 
Scene understanding
Scene understandingScene understanding
Scene understanding
Mohammed Shoaib
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
wolf
 
Fisheye Omnidirectional View in Autonomous Driving
Fisheye Omnidirectional View in Autonomous DrivingFisheye Omnidirectional View in Autonomous Driving
Fisheye Omnidirectional View in Autonomous Driving
Yu Huang
 
Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)
Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)
Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)
Tatsunori Taniai
 
DIGITAL IMAGE PROCESSING - Day 5 Applications of DIP
DIGITAL IMAGE PROCESSING - Day 5 Applications of DIPDIGITAL IMAGE PROCESSING - Day 5 Applications of DIP
DIGITAL IMAGE PROCESSING - Day 5 Applications of DIP
vijayanand Kandaswamy
 
Conception_et_realisation_dun_site_Web_d.pdf
Conception_et_realisation_dun_site_Web_d.pdfConception_et_realisation_dun_site_Web_d.pdf
Conception_et_realisation_dun_site_Web_d.pdf
SofianeHassine2
 

Similar to Neural Radiance Fields & Neural Rendering.pdf (20)

TransNeRF
TransNeRFTransNeRF
TransNeRF
 
Presentation_ Stable View Synthesis.pptx
Presentation_ Stable View Synthesis.pptxPresentation_ Stable View Synthesis.pptx
Presentation_ Stable View Synthesis.pptx
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IV
 
Shadow Techniques for Real-Time and Interactive Applications
Shadow Techniques for Real-Time and Interactive ApplicationsShadow Techniques for Real-Time and Interactive Applications
Shadow Techniques for Real-Time and Interactive Applications
 
Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3
 
Chapter 4 better.pptx
Chapter 4 better.pptxChapter 4 better.pptx
Chapter 4 better.pptx
 
Phong Shading over any Polygonal Surface
Phong Shading over any Polygonal Surface Phong Shading over any Polygonal Surface
Phong Shading over any Polygonal Surface
 
The single image dehazing based on efficient transmission estimation
The single image dehazing based on efficient transmission estimationThe single image dehazing based on efficient transmission estimation
The single image dehazing based on efficient transmission estimation
 
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNN
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNNAutomatic Detection of Window Regions in Indoor Point Clouds Using R-CNN
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNN
 
Advanced Lighting for Interactive Applications
Advanced Lighting for Interactive ApplicationsAdvanced Lighting for Interactive Applications
Advanced Lighting for Interactive Applications
 
regions
regionsregions
regions
 
Single Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learningSingle Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learning
 
[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...
[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...
[Paper] GIRAFFE: Representing Scenes as Compositional Generative Neural Featu...
 
HS Demo
HS DemoHS Demo
HS Demo
 
Scene understanding
Scene understandingScene understanding
Scene understanding
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
 
Fisheye Omnidirectional View in Autonomous Driving
Fisheye Omnidirectional View in Autonomous DrivingFisheye Omnidirectional View in Autonomous Driving
Fisheye Omnidirectional View in Autonomous Driving
 
Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)
Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)
Neural Inverse Rendering for General Reflectance Photometric Stereo (ICML 2018)
 
DIGITAL IMAGE PROCESSING - Day 5 Applications of DIP
DIGITAL IMAGE PROCESSING - Day 5 Applications of DIPDIGITAL IMAGE PROCESSING - Day 5 Applications of DIP
DIGITAL IMAGE PROCESSING - Day 5 Applications of DIP
 
Conception_et_realisation_dun_site_Web_d.pdf
Conception_et_realisation_dun_site_Web_d.pdfConception_et_realisation_dun_site_Web_d.pdf
Conception_et_realisation_dun_site_Web_d.pdf
 

Recently uploaded

20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
Matthew Sinclair
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
Alan Dix
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
James Anderson
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
Ana-Maria Mihalceanu
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Albert Hoitingh
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
ControlCase
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Paige Cruz
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
Guy Korland
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
DianaGray10
 
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex Proofs
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofszkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex Proofs
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex Proofs
Alex Pruden
 
Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
Alpen-Adria-Universität
 
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdfFIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Aggregage
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
DianaGray10
 
Artificial Intelligence for XMLDevelopment
Artificial Intelligence for XMLDevelopmentArtificial Intelligence for XMLDevelopment
Artificial Intelligence for XMLDevelopment
Octavian Nadolu
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Nexer Digital
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
SOFTTECHHUB
 
Climate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing DaysClimate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing Days
Kari Kakkonen
 
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
名前 です男
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
DanBrown980551
 

Recently uploaded (20)

20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
 
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex Proofs
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofszkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex Proofs
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex Proofs
 
Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
 
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdfFIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
 
Artificial Intelligence for XMLDevelopment
Artificial Intelligence for XMLDevelopmentArtificial Intelligence for XMLDevelopment
Artificial Intelligence for XMLDevelopment
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
 
Climate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing DaysClimate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing Days
 
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
 

Neural Radiance Fields & Neural Rendering.pdf

  • 1. Neural Radiance Fields & Neural Rendering Mildenhall, Ben, et al. "Nerf: Representing scenes as neural radiance fields for view synthesis." Communications of the ACM 65.1 (2021): 99-106. Navneet Paul PlayerUnknown Productions
  • 2. Rendering ● Process of generating an image from a 2D or 3D model using a computer program. The resulting image is called render. ● A rendering application takes into account inputs such as the model (2D/3D), texture, shading, lighting, viewpoints, etc, as features during the rendering process. ● It means we can say that each scene file contains multiple features that need to be understood and processed by the rendering algorithm or application to generate a processed image.
  • 3. Rendering Equation ● The rendering algorithm or technique which tries to solve the problem of image generation based on all given features is mostly trying to optimize the rendering equation. ● At a high-level the rendering equation computes for a radiance that is illumination (reflection, refraction, and emittance of light) on an object from a source to an observer in a given space. ● NeRF essentially computes for the Volume Rendering.
  • 4. Volume Rendering ● Volume rendering (as per Wikipedia) is a set of a technique used to display a 2D projection of a 3D discretely sampled dataset. ● To render a 2D projection (output) of a 3D dataset we first need to define the camera position in space relative to the volume then we need to define the RGBα (Red, Green. Blue, Alpha → it stands for opacity channel) for every voxel. ● The primary objective in volume rendering is to get a transfer function which defines RGBα for every value for every possible voxel value in a given space.
  • 5. View Synthesis ● Click photos of an object from multiple camera angles and superimpose the images to have a look at the same object from different known camera angles and positions. ● For NeRF, we are trying to predict the third missing axis (the first two being length & breadth) which is the depth. ● Core application of NeRF: to predict a function for depth determination at various points in the plane against the object itself.
  • 6. Neural Radiance Fields ● Generate novel views of complex scenes by optimizing an underlying continuous volumetric scene function using a sparse set of input views. ● The input can be provided as a blender model or a static set of images. ● The input is provided as a continuous 5D function that outputs the radiance emitted in each direction (θ; Φ) at each point (x; y; z) in space, and a density at each point which acts like a differential opacity controlling how much radiance is accumulated by a ray passing through (x; y; z)
  • 7. ● A continuous scene can be described as a 5D vector-valued function whose input is a 3D location x = (x; y; z) and 2D viewing direction (θ; Φ), and whose output is an emitted color c = (r; g; b) and volume density (𝜎) Volume Rendering Final rendering MLP network F𝚯 : (x,y,z,d) →(RGB,𝜎)
  • 8. Process overview To generate a Neural Radiance Field from a particular viewpoint following steps were done: ● March camera rays through the scene to generate a sampled set of 3D points (Use either COLMAP* or SfM for generating camera poses and viewing directions). ● Use those points and their corresponding 2D viewing directions as input to the neural network to produce an output set of colors (RGB) and densities (𝜎) ● Use classical volume rendering approach to accumulate those colors and densities into a 2D image * a general-purpose Structure-from-Motion (SfM) and Multi-View Stereo (MVS) pipeline with a graphical and command-line interface
  • 9. Network Architecture ● NeRF is an implicit Multi Layer Perceptron (MLP) based model that maps 5D vectors (3D coordinates plus 2D viewing directions) to output RGB feature vector (c) & volume density (𝜎) at that spatial location, using fully connected deep networks. → : layers with ReLU activation, 𝛾(x) : positional encoding, 𝛾(d) : directional encoding → : layer with no activation, ⇢ : layers with sigmoid activation, + : vector concatenation
  • 10. Volume Rendering ● The authors used discrete data samples to estimate the expected color C(r) of camera ray r(t) with the quadrature rule in classical volume rendering techniques. Predicted colors Volume density Opacity
  • 11. NeRF Optimization - Positional Encoding ● Previous studies show that optimizing the inputs to a higher dimensional space using high frequency functions before passing them to the network enables better fitting of data that contains high frequency variation. ● Positional & Directional Encoding: A Fourier based feature mapping function that encodes features (pertaining to position & direction) from lower dimensional space to a higher dimensional space. Positional Encoding func. Tancik, Srinivasan, Mildenhall et al., Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains, NeurIPS 2020 No Positional Encoding With Positional Encoding
  • 12. NeRF Optimization - Hierarchical Sampling ● During the volume rendering phase, our model simultaneously optimizes two networks: coarse and fine ● We first sample a set of NC locations with the RGB feature vector and density [σ (t)] outputs from the proposed NeRF model, using stratified sampling, and evaluate the “coarse” network at these locations. ● The main function of coarse network is to compute the final rendered color of the ray for the coarse samples. ● a second set of Nf locations are sampled from the [RGB + density] distribution using inverse transform sampling & evaluate our “fine” network. ● All the samples are considered while computing the final rendered ray color, i.e, (NC + Nf ), at fine network stage. This is done to ensure that more samples are allocated to regions we expect to contain visible content.
  • 13. Final Rendering & Loss Function ● Optimize a separate neural continuous volume representation network, for each scene. ● At each optimization iteration, we randomly sample a batch of camera rays from the set of all pixels in the dataset, and then follow the hierarchical sampling. ● NC samples from the coarse network and NC + Nf samples from the fine network. ● We then use the volume rendering procedure to render the color of each ray from both sets of samples. ● Loss function is based on the the total squared error between the rendered and true pixel colors for both the coarse and fine samples. ℛ: set of rays in each batch; C(r): ground truth, ĈC (r): coarse volume prediction and Ĉf (r): fine volume prediction for RGB colors for ray “r”
  • 14. Performance of NeRF Comparison to other view synthesis techniques ● Neural Volumes, Local Light Field Fusion (LLFF) & Scene Representation Networks (SRN (Ours = NeRF)
  • 15. Performance of NeRF Ablation Studies ● To validate the model’s performance with respect to different parameters.
  • 16. Summary ● Learn the radiance field of a scene based on a collection of calibrated images ○ Use an MLP to learn continuous geometry and view-dependent appearance ● Use fully differentiable volume rendering with reconstruction loss ● Combines hierarchical sampling and Fourier-based encoding of 5D inputs to produce high-fidelity novel view synthesis results Some associated challenges ● Handling dynamic scenes when acquiring calibrated views ● One network trained per scene - no generalization
  • 17. Related NeRF Research ● NeRF in Wild: a novel approach for 3D scene reconstruction of complex environments from unstructured internet photo collections that incorporates transient and latent scene embedding upon conventional NeRF model. *Martin-Brualla, Ricardo, et al. "Nerf in the wild: Neural radiance fields for unconstrained photo collections." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021. ● The model captures lighting and photometric variations in a low-dimensional latent embedding space in rendering appearance without affecting 3D geometry.
  • 18. ● Neural Radiance Fields for Dynamic Scenes : for synthesizing novel views, at an arbitrary point in time, of dynamic scenes with complex non-rigid geometries. ● Optimize an underlying deformable volumetric function (using a deformation network) from a sparse set of input monocular views without the need of ground-truth geometry nor multi-view images Pumarola, Albert, et al. "D-nerf: Neural radiance fields for dynamic scenes." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021. Related NeRF Research