Slides for the "generalization" session of our CVPR 2022 tutorial on Neural Fields in Computer Vision.
Neural Fields are an emerging technique to parameterize signals that live in spatial coordinates plus time. They parameterize a signal as a continuous function that maps a space-time coordinate to whatever is at that spacetime coordinate - for instance, the geometry of a 3D scene could be encoded in a function that maps a 3D coordinate to whether that coordinate is occupied or not. A neural field parameterizes that function as a neural network.
In this session, I gave a high-level overview over how we may use neural fields as the output of a variety of inference algorithms, for instance to reconstruct a complete 3D shape from partial observations in the form of a pointcloud, or to reconstruct a 3D scene from only a single image.
You are free to use the slides for any purpose, as long as you keep a note on the slides that acknowledges their source.
Neural Fields database: https://neuralfields.cs.brown.edu/
Tutorial website: https://neuralfields.cs.brown.edu/cvpr22
Neural Scene Representation & Rendering: Introduction to Novel View SynthesisVincent Sitzmann
An overview over the neural scene representation and rendering framework and an introduction to novel view synthesis approaches. Slides made for the Eurographics, CVPR, and SIGGRAPH courses on neural rendering, connected to the state-of-the-art report on Neural Rendering at Eurographics 2020.
Feel free to re-use the slides! I just ask that you keep some form of attribution, either at the beginning of your presentation, or in the slide footer.
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis taeseon ryu
해당 논문은 3D Aware 모델입니다 StyleGAN 같은 경우에는 어떤 하나의 피처에 대해서 Editing 하고 싶을 때 입력에 해당하는 레이턴트 백터를 찾아서 레이턴트 백터를 수정함으로써 입에 해당하는 피쳐를 바꿀 수 있었는데 이런 컨셉을 그대로 착안해서
GAN 스페이스 논문에서는 인풋이 들어왔을 때 어떤 공간적인 정보까지도 에디팅하려고 시도했습니다 결과를 봤을 때 로테이션 정보가 어느 정도 잘 학습된 것 같지만 같은 사람이 아닌 것 같이 인식되기도 합니다 이러한 문제를 이제 disentangle 되지 않았다라고 하는 게 원하는 피처만 변화시켜야 되는 것과 달리 다른 피처까지도 모두 학습 모두 변했다는 것인데 이를 좀 더 효율적으로 3D를 더 잘 이해시키기 위해서 탄생한 논문입니다.
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View SynthesisHyeongmin Lee
드디어 PR12 Season 4가 시작되었습니다! 제가 이번 시즌에서 발표하게 된 첫 논문은 ""NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis"라는 논문입니다. View Synthesis라는 Task는 몇 개의 시점에서 대상을 찍은 영상이 주어지면 주어지지 않은 위치와 방향에서 바라본 대상의 영상을 합성해내는 기술입니다. 이를 위해서 본 논문에서는 대상의 3D 정보를 통째로 Neural Network가 외우게 하는 방법을 선택했는데요, 이 방식은 Implicit Neural Representation이라는 이름으로 유명해지고 있는 추세고, 2D 이미지에 대해서도 적용하려는 접근들이 늘고 있습니다.
영상 링크: https://youtu.be/zkeh7Tt9tYQ
논문 링크: https://arxiv.org/abs/2003.08934
Neural Scene Representation & Rendering: Introduction to Novel View SynthesisVincent Sitzmann
An overview over the neural scene representation and rendering framework and an introduction to novel view synthesis approaches. Slides made for the Eurographics, CVPR, and SIGGRAPH courses on neural rendering, connected to the state-of-the-art report on Neural Rendering at Eurographics 2020.
Feel free to re-use the slides! I just ask that you keep some form of attribution, either at the beginning of your presentation, or in the slide footer.
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis taeseon ryu
해당 논문은 3D Aware 모델입니다 StyleGAN 같은 경우에는 어떤 하나의 피처에 대해서 Editing 하고 싶을 때 입력에 해당하는 레이턴트 백터를 찾아서 레이턴트 백터를 수정함으로써 입에 해당하는 피쳐를 바꿀 수 있었는데 이런 컨셉을 그대로 착안해서
GAN 스페이스 논문에서는 인풋이 들어왔을 때 어떤 공간적인 정보까지도 에디팅하려고 시도했습니다 결과를 봤을 때 로테이션 정보가 어느 정도 잘 학습된 것 같지만 같은 사람이 아닌 것 같이 인식되기도 합니다 이러한 문제를 이제 disentangle 되지 않았다라고 하는 게 원하는 피처만 변화시켜야 되는 것과 달리 다른 피처까지도 모두 학습 모두 변했다는 것인데 이를 좀 더 효율적으로 3D를 더 잘 이해시키기 위해서 탄생한 논문입니다.
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View SynthesisHyeongmin Lee
드디어 PR12 Season 4가 시작되었습니다! 제가 이번 시즌에서 발표하게 된 첫 논문은 ""NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis"라는 논문입니다. View Synthesis라는 Task는 몇 개의 시점에서 대상을 찍은 영상이 주어지면 주어지지 않은 위치와 방향에서 바라본 대상의 영상을 합성해내는 기술입니다. 이를 위해서 본 논문에서는 대상의 3D 정보를 통째로 Neural Network가 외우게 하는 방법을 선택했는데요, 이 방식은 Implicit Neural Representation이라는 이름으로 유명해지고 있는 추세고, 2D 이미지에 대해서도 적용하려는 접근들이 늘고 있습니다.
영상 링크: https://youtu.be/zkeh7Tt9tYQ
논문 링크: https://arxiv.org/abs/2003.08934
Super resolution in deep learning era - Jaejun YooJaeJun Yoo
Abstract (Eng/Kor):
Image restoration (IR) is one of the fundamental problems, which includes denoising, deblurring, super-resolution, etc. Among those, in today's talk, I will more focus on the super-resolution task. There are two main streams in the super-resolution studies; a traditional model-based optimization and a discriminative learning method. I will present the pros and cons of both methods and their recent developments in the research field. Finally, I will provide a mathematical view that explains both methods in a single holistic framework, while achieving the best of both worlds. The last slide summarizes the remaining problems that are yet to be solved in the field.
영상 복원(Image restoration, IR)은 low-level vision에서 매우 중요하게 다루는 근본적인 문제 중 하나로서 denoising, deblurring, super-resolution 등의 다양한 영상 처리 문제를 포괄합니다. 오늘 발표에서는 영상 복원 분야 중에서도 super-resolution 문제에 대해 집중적으로 다루겠습니다. 전통적인 model-based optimization 방식과 deep learning을 적용하여 문제를 푸는 방식에 대해, 각각의 장단점과 최신 연구 발전 흐름을 소개하겠습니다. 마지막으로는 이 둘을 하나로 잇는 통일된 관점을 제시하고 관련 연구들 살펴본 후, super-resolution 분야에서 아직 남아있는 문제점들을 정리하겠습니다.
DeNA AIシステム部内の輪講で発表した資料です。Deep fakesの種類やその検出法の紹介です。
主に下記の論文の紹介
S. Agarwal, et al., "Protecting World Leaders Against Deep Fakes," in Proc. of CVPR Workshop on Media Forensics, 2019.
A. Rossler, et al., "FaceForensics++: Learning to Detect Manipulated Facial Images," in Proc. of ICCV, 2019.
본 논문은 single depth map으로부터의 정확한 3D hand pose estimation을 목표로 한다. 3D hand pose estimation은 HCI, AR등의 기술을 구현함에 있어서 매우 중요한 기술이다. 이를 위해 많은 연구자들이 정확도를 높이기 위해 여러 방법을 제시하였지만, 여전히 손가락들의 비슷한 생김새, 가려짐, 다양한 손가락의 움직임으로 인한 복잡성 때문에 정확도를 올리는데 한계가 있었다. 본 논문은 기존 방법들의 한계를 극복하기 위해 기존 방법들이 사용하는 입력 형태와 출력 형태를 바꾸었다. 2d depth image를 입력으로 받아 hand joint의 3D coordinate를 직접 regress하는 대부분의 기존 방법들과는 달리, 제안하는 모델은 3D voxelized depth map을 입력으로 받아 3D heatmap을 출력한다. 이를 위해 encoder-decoder 형식의 3D CNN을 사용하였고, 달라진 입력과 출력 형태로 인해 제안하는 모델은 널리 사용되는 3개의 3d hand pose estimation dataset, 1개의 3d human pose estimation dataset에서 가장 높은 성능을 내었다. 또한 ICCV 2017에서 주최된 HANDS 2017 challenge에서 우승 하였다.
文献紹介:TSM: Temporal Shift Module for Efficient Video UnderstandingToru Tamaki
Ji Lin, Chuang Gan, Song Han; TSM: Temporal Shift Module for Efficient Video Understanding, Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019, pp. 7083-7093
https://openaccess.thecvf.com/content_ICCV_2019/html/Lin_TSM_Temporal_Shift_Module_for_Efficient_Video_Understanding_ICCV_2019_paper.html
In this project, we propose methods for semantic segmentation with the deep learning state-of-the-art models. Moreover,
we want to filterize the segmentation to the specific object in specific application. Instead of concentrating on unnecessary objects we
can focus on special ones and make it more specialize and effecient for special purposes. Furtheromore, In this project, we leverage
models that are suitable for face segmentation. The models that are used in this project are Mask-RCNN and DeepLabv3. The
experimental results clearly indicate that how illustrated approach are efficient and robust in the segmentation task to the previous work
in the field of segmentation. These models are reached to 74.4 and 86.6 precision of Mean of Intersection over Union. The visual
Results of the models are shown in Appendix part.
발표자: 전석준(KAIST 박사과정)
발표일: 2018.8.
Super-resolution은 저해상도 이미지를 고해상도 이미지로 변환시키는 기술로 오랜기간 연구되어 온 주제입니다. 최근 딥러닝 기술이 적용됨에 따라 super-resolution 성능이 비약적으로 향상되었습니다. 저희는 스테레오 이미지를 이용하여 더 높은 해상도의 이미지를 얻는 기술을 개발하였습니다. 이에 관련 내용을 발표하고자 합니다.
1. Multi-Frame Super-Resolution
2. Learning-Based Super-Resolution
3. Stereo Imaging
4. Deep-Learning Based Stereo Super-Resolution
Camera-Based Road Lane Detection by Deep Learning IIYu Huang
lane detection, deep learning, autonomous driving, CNN, RNN, LSTM, GRU, lane localization, lane fitting, ego lane, end-to-end, vanishing point, segmentation, FCN, regression, classification
Light Field Networks: Neural Scene Representations with Single-Evaluation Ren...Vincent Sitzmann
Slides for the video and talk of the NeurIPS 2021 paper "Light Field Networks: Neural Scene Representations with Single-Evaluation Rendering".
You are free to use these slides for any purpose, so long as you keep a note that acknowledges their source.
LFNs project website: https://www.vincentsitzmann.com/lfns/
Website of my research group: https://www.scenerepresentations.org/
My personal website: https://www.vincentsitzmann.com/
-- Abstract --
Inferring representations of 3D scenes from 2D observations is a fundamental problem of computer graphics, computer vision, and artificial intelligence. Emerging 3D-structured neural scene representations are a promising approach to 3D scene understanding. In this work, we propose a novel neural scene representation, Light Field Networks or LFNs, which represent both geometry and appearance of the underlying 3D scene in a 360-degree, four-dimensional light field parameterized via a neural implicit representation. Rendering a ray from an LFN requires only a *single* network evaluation, as opposed to hundreds of evaluations per ray for ray-marching or volumetric based renderers in 3D-structured neural scene representations. In the setting of simple scenes, we leverage meta-learning to learn a prior over LFNs that enables multi-view consistent light field reconstruction from as little as a single image observation. This results in dramatic reductions in time and memory complexity, and enables real-time rendering. The cost of storing a 360-degree light field via an LFN is two orders of magnitude lower than conventional methods such as the Lumigraph. Utilizing the analytical differentiability of neural implicit representations and a novel parameterization of light space, we further demonstrate the extraction of sparse depth maps from LFNs.
Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Rep...Vincent Sitzmann
Slides for our 2019 NeurIPS paper and Honorable Mention for Promising New Directions in Research, "Scene Representation Networks: Continuous 3D-Structure Aware Neural Scene Representations".
You are free to use these slides for any purpose, so long as you keep an acknowledgement on the slide that denotes its source.
Project page: https://www.vincentsitzmann.com/srns/
The website of my research group at MIT: https://www.scenerepresentations.org/
My personal website: https://www.vincentsitzmann.com/
-- Abstract --
We propose Scene Representation Networks (SRNs), a continuous, 3D-structure-aware scene representation that encodes both geometry and appearance. SRNs represent scenes as continuous functions that map world coordinates to a feature representation of local scene properties. By formulating the image formation as a neural, 3D-aware rendering algorithm, SRNs can be trained end-to-end from only 2D observations, without access to depth or geometry. SRNs do not discretize space, smoothly parameterizing scene surfaces, and their memory complexity does not scale directly with scene resolution. This formulation naturally generalizes across scenes, learning powerful geometry and appearance priors in the process.
Super resolution in deep learning era - Jaejun YooJaeJun Yoo
Abstract (Eng/Kor):
Image restoration (IR) is one of the fundamental problems, which includes denoising, deblurring, super-resolution, etc. Among those, in today's talk, I will more focus on the super-resolution task. There are two main streams in the super-resolution studies; a traditional model-based optimization and a discriminative learning method. I will present the pros and cons of both methods and their recent developments in the research field. Finally, I will provide a mathematical view that explains both methods in a single holistic framework, while achieving the best of both worlds. The last slide summarizes the remaining problems that are yet to be solved in the field.
영상 복원(Image restoration, IR)은 low-level vision에서 매우 중요하게 다루는 근본적인 문제 중 하나로서 denoising, deblurring, super-resolution 등의 다양한 영상 처리 문제를 포괄합니다. 오늘 발표에서는 영상 복원 분야 중에서도 super-resolution 문제에 대해 집중적으로 다루겠습니다. 전통적인 model-based optimization 방식과 deep learning을 적용하여 문제를 푸는 방식에 대해, 각각의 장단점과 최신 연구 발전 흐름을 소개하겠습니다. 마지막으로는 이 둘을 하나로 잇는 통일된 관점을 제시하고 관련 연구들 살펴본 후, super-resolution 분야에서 아직 남아있는 문제점들을 정리하겠습니다.
DeNA AIシステム部内の輪講で発表した資料です。Deep fakesの種類やその検出法の紹介です。
主に下記の論文の紹介
S. Agarwal, et al., "Protecting World Leaders Against Deep Fakes," in Proc. of CVPR Workshop on Media Forensics, 2019.
A. Rossler, et al., "FaceForensics++: Learning to Detect Manipulated Facial Images," in Proc. of ICCV, 2019.
본 논문은 single depth map으로부터의 정확한 3D hand pose estimation을 목표로 한다. 3D hand pose estimation은 HCI, AR등의 기술을 구현함에 있어서 매우 중요한 기술이다. 이를 위해 많은 연구자들이 정확도를 높이기 위해 여러 방법을 제시하였지만, 여전히 손가락들의 비슷한 생김새, 가려짐, 다양한 손가락의 움직임으로 인한 복잡성 때문에 정확도를 올리는데 한계가 있었다. 본 논문은 기존 방법들의 한계를 극복하기 위해 기존 방법들이 사용하는 입력 형태와 출력 형태를 바꾸었다. 2d depth image를 입력으로 받아 hand joint의 3D coordinate를 직접 regress하는 대부분의 기존 방법들과는 달리, 제안하는 모델은 3D voxelized depth map을 입력으로 받아 3D heatmap을 출력한다. 이를 위해 encoder-decoder 형식의 3D CNN을 사용하였고, 달라진 입력과 출력 형태로 인해 제안하는 모델은 널리 사용되는 3개의 3d hand pose estimation dataset, 1개의 3d human pose estimation dataset에서 가장 높은 성능을 내었다. 또한 ICCV 2017에서 주최된 HANDS 2017 challenge에서 우승 하였다.
文献紹介:TSM: Temporal Shift Module for Efficient Video UnderstandingToru Tamaki
Ji Lin, Chuang Gan, Song Han; TSM: Temporal Shift Module for Efficient Video Understanding, Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019, pp. 7083-7093
https://openaccess.thecvf.com/content_ICCV_2019/html/Lin_TSM_Temporal_Shift_Module_for_Efficient_Video_Understanding_ICCV_2019_paper.html
In this project, we propose methods for semantic segmentation with the deep learning state-of-the-art models. Moreover,
we want to filterize the segmentation to the specific object in specific application. Instead of concentrating on unnecessary objects we
can focus on special ones and make it more specialize and effecient for special purposes. Furtheromore, In this project, we leverage
models that are suitable for face segmentation. The models that are used in this project are Mask-RCNN and DeepLabv3. The
experimental results clearly indicate that how illustrated approach are efficient and robust in the segmentation task to the previous work
in the field of segmentation. These models are reached to 74.4 and 86.6 precision of Mean of Intersection over Union. The visual
Results of the models are shown in Appendix part.
발표자: 전석준(KAIST 박사과정)
발표일: 2018.8.
Super-resolution은 저해상도 이미지를 고해상도 이미지로 변환시키는 기술로 오랜기간 연구되어 온 주제입니다. 최근 딥러닝 기술이 적용됨에 따라 super-resolution 성능이 비약적으로 향상되었습니다. 저희는 스테레오 이미지를 이용하여 더 높은 해상도의 이미지를 얻는 기술을 개발하였습니다. 이에 관련 내용을 발표하고자 합니다.
1. Multi-Frame Super-Resolution
2. Learning-Based Super-Resolution
3. Stereo Imaging
4. Deep-Learning Based Stereo Super-Resolution
Camera-Based Road Lane Detection by Deep Learning IIYu Huang
lane detection, deep learning, autonomous driving, CNN, RNN, LSTM, GRU, lane localization, lane fitting, ego lane, end-to-end, vanishing point, segmentation, FCN, regression, classification
Light Field Networks: Neural Scene Representations with Single-Evaluation Ren...Vincent Sitzmann
Slides for the video and talk of the NeurIPS 2021 paper "Light Field Networks: Neural Scene Representations with Single-Evaluation Rendering".
You are free to use these slides for any purpose, so long as you keep a note that acknowledges their source.
LFNs project website: https://www.vincentsitzmann.com/lfns/
Website of my research group: https://www.scenerepresentations.org/
My personal website: https://www.vincentsitzmann.com/
-- Abstract --
Inferring representations of 3D scenes from 2D observations is a fundamental problem of computer graphics, computer vision, and artificial intelligence. Emerging 3D-structured neural scene representations are a promising approach to 3D scene understanding. In this work, we propose a novel neural scene representation, Light Field Networks or LFNs, which represent both geometry and appearance of the underlying 3D scene in a 360-degree, four-dimensional light field parameterized via a neural implicit representation. Rendering a ray from an LFN requires only a *single* network evaluation, as opposed to hundreds of evaluations per ray for ray-marching or volumetric based renderers in 3D-structured neural scene representations. In the setting of simple scenes, we leverage meta-learning to learn a prior over LFNs that enables multi-view consistent light field reconstruction from as little as a single image observation. This results in dramatic reductions in time and memory complexity, and enables real-time rendering. The cost of storing a 360-degree light field via an LFN is two orders of magnitude lower than conventional methods such as the Lumigraph. Utilizing the analytical differentiability of neural implicit representations and a novel parameterization of light space, we further demonstrate the extraction of sparse depth maps from LFNs.
Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Rep...Vincent Sitzmann
Slides for our 2019 NeurIPS paper and Honorable Mention for Promising New Directions in Research, "Scene Representation Networks: Continuous 3D-Structure Aware Neural Scene Representations".
You are free to use these slides for any purpose, so long as you keep an acknowledgement on the slide that denotes its source.
Project page: https://www.vincentsitzmann.com/srns/
The website of my research group at MIT: https://www.scenerepresentations.org/
My personal website: https://www.vincentsitzmann.com/
-- Abstract --
We propose Scene Representation Networks (SRNs), a continuous, 3D-structure-aware scene representation that encodes both geometry and appearance. SRNs represent scenes as continuous functions that map world coordinates to a feature representation of local scene properties. By formulating the image formation as a neural, 3D-aware rendering algorithm, SRNs can be trained end-to-end from only 2D observations, without access to depth or geometry. SRNs do not discretize space, smoothly parameterizing scene surfaces, and their memory complexity does not scale directly with scene resolution. This formulation naturally generalizes across scenes, learning powerful geometry and appearance priors in the process.
Transformer Architectures in Vision
[2018 ICML] Image Transformer
[2019 CVPR] Video Action Transformer Network
[2020 ECCV] End-to-End Object Detection with Transformers
[2021 ICLR] An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
Point-GNN: Graph Neural Network for 3D Object Detection in a Point CloudNuwan Sriyantha Bandara
Presentation on the Point-GNN paper (presented at CVPR 2020) for the module: Advances in Machine Vision at the Department of Electronic and Telecommunication Engineering, University of Moratuwa, Sri Lanka.
Presentation slides are prepared by Nuwan Bandara.
HR3D: Content Adaptive Parallax Barriers, SIGGRAPH Asia 2010 Technical Paper presentation, presented by Douglas Lanman (http://web.media.mit.edu/~dlanman). Please see the project page for more details: http://web.media.mit.edu/~mhirsch/hr3d
This is a project in the Camera Culture group (http://cameraculture.media.mit.edu) at the MIT Media Lab, led by Professor Ramesh Raskar (http://web.media.mit.edu/~raskar).
2019年6月13日、SSII2019 Organized Session: Multimodal 4D sensing。エンドユーザー向け SLAM 技術の現在。登壇者:武笠 知幸(Research Scientist, Rakuten Institute of Technology)
https://confit.atlas.jp/guide/event/ssii2019/static/organized#OS2
https://telecombcn-dl.github.io/2018-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or image captioning.
https://telecombcn-dl.github.io/2017-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or image captioning.
SIGGRAPH 2014 Course on Computational Cameras and Displays (part 2)Matthew O'Toole
Recent advances in both computational photography and displays have given rise to a new generation of computational devices. Computational cameras and displays provide a visual experience that goes beyond the capabilities of traditional systems by adding computational power to optics, lights, and sensors. These devices are breaking new ground in the consumer market, including lightfield cameras that redefine our understanding of pictures (Lytro), displays for visualizing 3D/4D content without special eyewear (Nintendo 3DS), motion-sensing devices that use light coded in space or time to detect motion and position (Kinect, Leap Motion), and a movement toward ubiquitous computing with wearable cameras and displays (Google Glass).
This short (1.5 hour) course serves as an introduction to the key ideas and an overview of the latest work in computational cameras, displays, and light transport.
Tissue Segmentation Methods Using 2D Histogram Matching in a Sequence of MR B...Vladimir Kanchev
This presentation provides detailed description of the methodology of the segmentation method of brain tissues in MR image sequences using 2D histogram matching.
A brief information about the SCOP protein database used in bioinformatics.
The Structural Classification of Proteins (SCOP) database is a comprehensive and authoritative resource for the structural and evolutionary relationships of proteins. It provides a detailed and curated classification of protein structures, grouping them into families, superfamilies, and folds based on their structural and sequence similarities.
Multi-source connectivity as the driver of solar wind variability in the heli...Sérgio Sacani
The ambient solar wind that flls the heliosphere originates from multiple
sources in the solar corona and is highly structured. It is often described
as high-speed, relatively homogeneous, plasma streams from coronal
holes and slow-speed, highly variable, streams whose source regions are
under debate. A key goal of ESA/NASA’s Solar Orbiter mission is to identify
solar wind sources and understand what drives the complexity seen in the
heliosphere. By combining magnetic feld modelling and spectroscopic
techniques with high-resolution observations and measurements, we show
that the solar wind variability detected in situ by Solar Orbiter in March
2022 is driven by spatio-temporal changes in the magnetic connectivity to
multiple sources in the solar atmosphere. The magnetic feld footpoints
connected to the spacecraft moved from the boundaries of a coronal hole
to one active region (12961) and then across to another region (12957). This
is refected in the in situ measurements, which show the transition from fast
to highly Alfvénic then to slow solar wind that is disrupted by the arrival of
a coronal mass ejection. Our results describe solar wind variability at 0.5 au
but are applicable to near-Earth observatories.
Cancer cell metabolism: special Reference to Lactate PathwayAADYARAJPANDEY1
Normal Cell Metabolism:
Cellular respiration describes the series of steps that cells use to break down sugar and other chemicals to get the energy we need to function.
Energy is stored in the bonds of glucose and when glucose is broken down, much of that energy is released.
Cell utilize energy in the form of ATP.
The first step of respiration is called glycolysis. In a series of steps, glycolysis breaks glucose into two smaller molecules - a chemical called pyruvate. A small amount of ATP is formed during this process.
Most healthy cells continue the breakdown in a second process, called the Kreb's cycle. The Kreb's cycle allows cells to “burn” the pyruvates made in glycolysis to get more ATP.
The last step in the breakdown of glucose is called oxidative phosphorylation (Ox-Phos).
It takes place in specialized cell structures called mitochondria. This process produces a large amount of ATP. Importantly, cells need oxygen to complete oxidative phosphorylation.
If a cell completes only glycolysis, only 2 molecules of ATP are made per glucose. However, if the cell completes the entire respiration process (glycolysis - Kreb's - oxidative phosphorylation), about 36 molecules of ATP are created, giving it much more energy to use.
IN CANCER CELL:
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
introduction to WARBERG PHENOMENA:
WARBURG EFFECT Usually, cancer cells are highly glycolytic (glucose addiction) and take up more glucose than do normal cells from outside.
Otto Heinrich Warburg (; 8 October 1883 – 1 August 1970) In 1931 was awarded the Nobel Prize in Physiology for his "discovery of the nature and mode of action of the respiratory enzyme.
WARNBURG EFFECT : cancer cells under aerobic (well-oxygenated) conditions to metabolize glucose to lactate (aerobic glycolysis) is known as the Warburg effect. Warburg made the observation that tumor slices consume glucose and secrete lactate at a higher rate than normal tissues.
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
Tutorial on Generalization in Neural Fields, CVPR 2022 Tutorial on Neural Fields in Computer Vision
1. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
NEURAL FIELDS IN COMPUTER VISION
Full-Day Tutorial, June 20th, 2022
neuralfields.cs.brown.edu/cvpr22
Reality Labs Research
Yiheng Xie Towaki Takikawa Shunsuke Saito Or Litany James Tompkin Vincent Sitzmann Srinath Sridhar
2. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Prior-based Reconstruction of
Neural Fields
2
Vincent Sitzmann
Assistant Professor, Scene Representation Group
www.scenerepresentations.com
www.vincentsitzmann.com
3. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Motivation: Novel View Synthesis
+
+
Observations
Image + Pose & Intrinsics
{ ,
,
…
{ Model
Novel Views
4. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Motivation: Novel View Synthesis
4
Fitting /
Optimization
Neural Scene
Representatio
n
Neural
Renderer
5. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Motivation: Novel View Synthesis
5
Inference
Neural Scene
Representatio
n
Neural
Renderer
Inference maps a set of observations to the parameters of a Neural Scene Representation.
6. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Overfitting case: Inference = Fitting via Gradient Descent
6
,…
+ }
{
REN D ER 𝜽
SDF + Color MLPs
SR N 𝝓
Fitting
Rendering
Normal map RGB
Sitzmann et al: Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations, NeurIPS 2020.
min REN D ER 𝜽(SR N 𝝓, 𝜉𝑖) − ℐ𝑖
8. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
What if we have incomplete observations?
8
REN D ER 𝜽
SDF + Color MLPs
SR N 𝝓
Sitzmann et al: Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations, NeurIPS 2020.
min REN D ER 𝜽(SR N 𝝓, 𝜉𝑖) − ℐ𝑖
+
ℐ, 𝜉
No 3D inform.
Normal map RGB
9. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Inferring Neural Fields
9
Neural Scene
Representatio
n
Neural
Renderer
If only a single observation is available, or if only part of the scene has been observed,
Inference needs to be prior-based – i.e., we need to learn to reconstruct.
?
10. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
General Framework: Encoder-Decoder
10
Neural Scene
Representatio
n
Neural
Renderer
Decoder
Inference
Latent Variables {𝑧𝑖}𝑖=1
𝑁
Encoder
11. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
What are the latent variables?
11
Neural Scene
Representatio
n
Neural
Renderer
Inference
Encoder Latent Variables {𝑧𝑖}𝑖=1
𝑁
12. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
How to predict latent variables from observations?
12
Neural Scene
Representatio
n
Neural
Renderer
Inference
Encoder Latent Variables {𝑧𝑖}𝑖=1
𝑁
13. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
How do we decode latent variables into the Neural Field?
13
Neural Scene
Representatio
n
Neural
Renderer
Inference
Encoder Latent Variables {𝑧𝑖}𝑖=1
𝑁
14. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
What are the latent variables?
14
Neural Scene
Representatio
n
Neural
Renderer
Inference
Encoder Latent Variables {𝑧𝑖}𝑖=1
𝑁
15. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Key Consideration: Locality.
15
Neural Fields in Visual Computing and Beyond, Xie et al., EG STAR 2022
Global Conditioning Local Conditioning
16. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Global Latent Codes
16
Neural Fields in Visual Computing and Beyond, Xie et al., EG STAR 2022
Global Conditioning Local Conditioning
17. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Global conditioning
17
?
Latent code 𝑧
18. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Global conditioning
18
1[Schmidhuber et al. 1992, Schmidhuber et al. 1993, Stanley et al. 2009, Ha et al., 2016]
Hypernetwork1
Latent code 𝑧
19. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Global Latent Codes: Enables reconstruction from partial observations!
19
Scene Representation Networks: Continuous
3D-Structure-Aware Neural Scene Representations, NeurIPS 2019.
Differential Volumetric Rendering,
Niemeyer et al., CVPR 2020
DeepSDF, Occupancy Networks, IM-Net
20. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Global Latent Codes: Enables reconstruction from partial observations!
20
Scene Representation Networks: Continuous
3D-Structure-Aware Neural Scene Representations, NeurIPS 2019.
Differential Volumetric Rendering,
Niemeyer et al., CVPR 2020
DeepSDF, Occupancy Networks, IM-Net
Key limitation: Simple, non-compositional scenes.
But: Latent Space for full objects (interpolation etc)
21. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Local Latent Codes
21
Neural Fields in Visual Computing and Beyond, Xie et al., EG STAR 2022
Global Conditioning Local Conditioning
22. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
From point clouds: Conditioning on Feature Voxel grids
22
Convolutional Occupancy Networks [Peng et al. 2020]
Local Implicit Grid Representations for 3D Scenes [Jiang et al. 2020]
Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion [Chabra et al. 2020]
Deep Local Shapes: Learning Local SDF Priors for Detailed 3D Reconstruction [Chibane et al. 2020]
23. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
From point clouds: Conditioning on Feature Voxel grids
23
Convolutional Occupancy Networks [Peng et al. 2020]
Local Implicit Grid Representations for 3D Scenes [Jiang et al. 2020]
Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion [Chabra et al. 2020]
Deep Local Shapes: Learning Local SDF Priors for Detailed 3D Reconstruction [Chibane et al. 2020]
24. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
From point clouds: Conditioning on Feature Voxel grids
24
Generalizes to Compositional Scenes!
But: cubic memory complexity :/
25. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
From Point clouds: Ground-plan and Tri-plane factorizations
25
Convolutional Occupancy Networks [Peng et al. 2020]
26. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
From Point clouds: Ground-plan and Tri-plane factorizations
26
Convolutional Occupancy Networks [Peng et al. 2020]
27. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
From Point clouds: Ground-plan and Tri-plane factorizations
27
Convolutional Occupancy Networks [Peng et al. 2020]
28. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
From point clouds: Conditioning on Reconstructed Voxelgrids
28
5x less memory!
29. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
How to locally condition if sensor
domain different than field
domain?
29
30. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Local Conditioning: Pixel-Aligned Features.
30
PiFU, Saito et al., ICCV 2019.
PixelNeRF, Yu et al., CVPR 2021
Grf: Learning a general radiance field…, Trevithick et al.
31. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Local Conditioning: Pixel-Aligned Features.
31
PiFU, Saito et al., ICCV 2019.
PixelNeRF, Yu et al., CVPR 2021
Grf: Learning a general radiance field…, Trevithick et al.
Generalizes much better than global conditioning (like SRNs, DVR).
No persistent 3D representation.
All priors are learned in image space.
32. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Object-centric representations
32
CoLF: Unsupervised Learning of Compositional Object Light Fields, arXiv 2022.
33. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Object-centric representations
CoLF: Unsupervised Learning of Compositional
Object Light Fields, arXiv 2022.
uORF, ICLR 2022
Learns to disentangle objects self-supervised.
Inference of object-centric latent codes is hard problem.
Currently limited to relatively simple scenes, but progress is quick!
34. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Conditional Ground Plans for Single-Image 3D Reconstruction
34
Seeing 3D Objects in a Single Image via Self-Supervised Static-Dynamic Disentanglement, Sharma et al. 2022
35. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Conditional Ground Plans for Single-Image 3D Reconstruction
35
Seeing 3D Objects in a Single Image via Self-Supervised Static-Dynamic Disentanglement, Sharma et al. 2022
36. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Conditional Ground Plans for Single-Image 3D Reconstruction
36
Seeing 3D Objects in a Single Image via Self-Supervised Static-Dynamic Disentanglement, Sharma et al. 2022
37. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
How to infer latent codes?
37
Neural Scene
Representatio
n
Neural
Renderer
Inference
Encoder Latent Variables {𝑧𝑖}𝑖=1
𝑁
38. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Encoding vs. Auto-Decoding
38
Neural Fields in Visual Computing and Beyond, Xie et al., EG STAR 2022
Encoding Auto-Decoding
39. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Auto-Decoding for inverse graphics
39
REN D ER
Latent code 𝑧0
40. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Auto-Decoding for inverse graphics
40
REN D ER
Latent code 𝑧0
𝑧 = arg min
𝑧
REN D ER (Φ) − ℐ
3D-structured, resolution-invariant!
Samples need not lie on regular
grids!
41. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Out-of-distribution generalization
41
3D structure enables generalization
to out-of-distribution camera poses!
𝑧 = arg min
𝑧
REN D ER 𝜽(SR N 𝜙=𝐻𝑁𝜓(𝑧), 𝜉) − ℐ
Reconstruction
CNN encoder
Input
42. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Other forms of Generalization: Transformer Decoders
42
AIR-Nets, Giebenhain et al. 2022
Scene Representation Transformer
Sajjadi et al. 2022
43. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Other forms of Generalization: Gradient-based meta-learning
Representation
In-the loop
specialization via gradient
descent
Meta-Representation
43
MetaSDF: Meta-learning Signed Distance Functions, NeurIPS 2020
Backpropagate through gradient-
descent inference at training time.
Learn initialization that explains
held-out observations when fit to
context observation.
44. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Inferring Neural Scene Representations
44
Inference
Neural Scene
Representatio
n
Neural
Renderer
Generalization enables reconstruction from incomplete observations.
Any other benefits?
45. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Problem: Forward map might be expensive!
45
Inference
Neural Scene
Representatio
n
Neural
Renderer
46. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
3D-structured Neural Scene Representations
: ℝ3 → ℝn
Hundreds of samples per ray.
Time- and memory-intensive training.
47. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
: ℝ3 → ℝn
[Adelson et al. 1991, Levoy et al. 1996, Gortler et al. 1996]
Light Field
50. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Light Field Networks
Conditioning
Plücker Coords.
An Alternative Scene Representation
51. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Rendering is learned / representation is “already rendered”
51
Inference
Neural Scene
Representatio
n
Neural
Renderer
52. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Rendering is learned / representation is “already rendered”
52
Inference
“Rendered” Neural Scene
Representation
More difficult inference problem, but more general renderer.
53. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Light Field Networks Volumetric Rendering (pixelNeRF)
500 FPS
1 evaluation per ray
0.033 FPS
196 evaluations per ray
Real-time. No post-processing, no discrete data structures (octrees, voxelgrids, …).
>100x reduction in memory: Can be trained on small GPUs!
15,000x speed
1,000x speed
100x speed
10x speed
1x speed
54. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Light Field Networks
500 FPS
1 evaluation per ray
55. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Light Fields with Transformers:
Scene Representation Transformer (CVPR 2022)
No 3D Renderer: Directly parameterizes Light
Field!
56. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
56
Things I didn’t talk about
● Generalization in 2D, 1D, etc. neural fields: Images, audio…
see LIIF (Chen et al. 2021), …
● Neural field-to-neural field translation, see Spatially-Adaptive
Pixelwise Networks for Fast Image Translation (Shaham et al.
2020)
● Generalization for robotics applications (see Neural Descriptor
Fields (Simeonov et al.), 3D neural scene … (Li et al., CoRL 2022),
Learning Multi-Object Dynamics... (Driess et al. 2022), …
● Generalization for structured field with known a-priori structure
(humans, hands, faces, etc)
57. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
57
Outlook
● Generalization gaining traction: Single-scene optimization too
limited.
● Opens up completely new ways of thinking about problems:
Can amortize otherwise expensive forward maps (light fields).
● Making progress on the question of compositionality w/ object-
centric and locally conditioned neural fields. More to come.
● Processing & inferring regular grids is easy. Harder for point clouds
/ factorized representations, etc.
● Transformers seem to learn a type of local conditioning, but more
research necessary.
58. Vincent Sitzmann, CVPR 2022 Tutorial on Neural Fields
Prior-based Reconstruction of
Neural Fields
58
Vincent Sitzmann
Assistant Professor, Scene Representation Group
www.scenerepresentations.com
www.vincentsitzmann.com