With the increasing needs of intelligent and autonomous systems to sense, move and react with the surroundings, it is a clear necessity to train such systems with as much relevant data as can be obtained. However, there are many challenges in obtaining real world data, particularly in a 3D environment. In this talk, I will cover some of the recent advances in Graphics and Computing techniques in 3D processing and their possible application in dynamic settings for autonomous systems. A vision of how synthetic data could be relevant in the future of intelligent systems is presented, along with the challenges. Backup material covers latest papers on the subject
Introduction to mago3D, an Open Source Based Digital Twin PlatformSANGHEE SHIN
This talk was given at the Busan Eco Delta City(Korea National Pilot Smart City) technical workshop held on 18th July. I talked about introduction and history of mago3D, some core technologies, real cases, and lessons learnt in this workshop.
What we've done so far with mago3D, an open source based 'Digital Twin' platf...SANGHEE SHIN
mago3D = {Indoor, Outdoor} + {Overground, Underground} + {Objects, Phenomena} + {Static, Dynamic}
It would be awesome if you can have a virtual replica of real world that you can play with and do the simulation to see what would happen. That is 'Digital Twin', the ultimate goal of mago3D!
At the FOSS4G NA 2019, I talked about the recent achievements and improvements of mago3D project, an open source based 'Digital Twin' platform. mago3D(http://mago3d.com) is relatively new project that was first released in July 2017. The ultimate goal of mago3D project is developing an open source based digital twin platform that can replicate and simulate the real world objects, processes, and phenomena on web environment. mago3D is on its way to achieve this goal now. Currently mago3D more focuses on managing and visualization of various types of 3D data ranging from simple box style extrusion model, point clouds, realistic mesh, to complex BIM(Building Information Modeling), AEC(Architecture, Engineering, Construction) data. mago3D supports industry standards 3D formats such as IFC, CityGML, IndoorGML, 3DS, Collada DAE, OBJ, LAS, JT, and so on. mago3D has been used in various industry sectors including ship building, urban management, indoor data management, and national defense. In this talk I showcased several real projects that had employed the mago3D and talked about what I'd learned during this projects. I also talked more about the future plan of mago3D towards visualizing/simulating of {static and dynamic data}, {underground and overground features}, {indoor and outdoor spaces}, {objects and phenomena} at the same time on web browser.
As a tech-savvy country, there're lots of discussions and activities around digital twin in Korea. I also shared my real experiences on this in this talk.
Walking in the Cloud: A New Paradigm in Geospatial WorldICIMOD
Cloud computing allows scalable and efficient use of computing resources in the internet cloud. The use of cloud computing is increasing across all the application domains, including in the field of GIS and remote sensing.
The advent of Google Earth Engine (GEE) in particular has brought about a revolutionary change in the way we use geospatial technology. The GEE is a cloud based geospatial platform that stores petabyte of satellite imagery and geospatial data and enables carrying out of complex image processing tasks and spatial analyses without the need of any GIS and remote sensing software.
State of mago3D, An Open Source Based Digital Twin PlatformSANGHEE SHIN
I gave this talk at the FOSS4G Thailand 2019 which was held at Chulalongkorn University, Bangkok on 4th Nov 2019. I talked about the recent achievements and improvements of mago3D project, an open source based Digital Twin platform. mago3D(http mago3d.com) is relatively new project first released in July 2017. The ultimate goal of mago3D is developing an open source based digital twin platform that can replicate and simulate the real world objects, processes, and phenomena on web environment. mago3D has been used in various industry sectors including ship building, urban management, indoor data management, and national defense. In this talk I showcased several real projects that used the mago3D and shared what I learnt from these projects. Also I introduced new features and future plan of mago3D.
Current State of mago3D, an Open Source Based Digital Twin PlatformSANGHEE SHIN
I gave this talk at the FOSS4G 2019 Bucharest. I talked about the recent achievements and improvements of mago3D project, an open source based Digital Twin platform. mago3D(http mago3d.com) is relatively new project first released in July 2017. The ultimate goal of mago3D is developing an open source based digital twin platform that can replicate and simulate the real world objects, processes, and phenomena on web environment. mago3D is on its way to achieve this goal now. mago3D has been used in various industry sectors including ship building, urban management, indoor data management, and national defense. In this talk I showcased several real projects that employed the mago3D and shared what I learnt from these projects.
Introduction to mago3D: A Web Based Open Source GeoBIM PlatformSANGHEE SHIN
I gave this talk at the FOSS4G Asia 2018 held at University of Moratuwa, Sri Lanka. I've added some of recent improvements of mago3D features including CityGML, IndoorGML supporting. Also I've talked about the future plan of mago3D toward Digital Twin platform.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/qualcomm/embedded-vision-training/videos/pages/may-2016-embedded-vision-summit-mangen
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Michael Mangen, Product Manager for Camera and Computer Vision at Qualcomm, presents the "High-resolution 3D Reconstruction on a Mobile Processor" tutorial at the May 2016 Embedded Vision Summit.
Computer vision has come a long way. Use cases that were previously not possible in mass-market devices are now more accessible thanks to advances in depth sensors and mobile processors. In this presentation, Mangen provides an overview of how we are able to implement high-resolution 3D reconstruction – a capability typically requiring cloud/server processing – on a mobile processor. This is an exciting example of how new sensor technology and advanced mobile processors are bringing computer vision capabilities to broader markets.
Introduction to mago3D, an Open Source Based Digital Twin PlatformSANGHEE SHIN
This talk was given at the Busan Eco Delta City(Korea National Pilot Smart City) technical workshop held on 18th July. I talked about introduction and history of mago3D, some core technologies, real cases, and lessons learnt in this workshop.
What we've done so far with mago3D, an open source based 'Digital Twin' platf...SANGHEE SHIN
mago3D = {Indoor, Outdoor} + {Overground, Underground} + {Objects, Phenomena} + {Static, Dynamic}
It would be awesome if you can have a virtual replica of real world that you can play with and do the simulation to see what would happen. That is 'Digital Twin', the ultimate goal of mago3D!
At the FOSS4G NA 2019, I talked about the recent achievements and improvements of mago3D project, an open source based 'Digital Twin' platform. mago3D(http://mago3d.com) is relatively new project that was first released in July 2017. The ultimate goal of mago3D project is developing an open source based digital twin platform that can replicate and simulate the real world objects, processes, and phenomena on web environment. mago3D is on its way to achieve this goal now. Currently mago3D more focuses on managing and visualization of various types of 3D data ranging from simple box style extrusion model, point clouds, realistic mesh, to complex BIM(Building Information Modeling), AEC(Architecture, Engineering, Construction) data. mago3D supports industry standards 3D formats such as IFC, CityGML, IndoorGML, 3DS, Collada DAE, OBJ, LAS, JT, and so on. mago3D has been used in various industry sectors including ship building, urban management, indoor data management, and national defense. In this talk I showcased several real projects that had employed the mago3D and talked about what I'd learned during this projects. I also talked more about the future plan of mago3D towards visualizing/simulating of {static and dynamic data}, {underground and overground features}, {indoor and outdoor spaces}, {objects and phenomena} at the same time on web browser.
As a tech-savvy country, there're lots of discussions and activities around digital twin in Korea. I also shared my real experiences on this in this talk.
Walking in the Cloud: A New Paradigm in Geospatial WorldICIMOD
Cloud computing allows scalable and efficient use of computing resources in the internet cloud. The use of cloud computing is increasing across all the application domains, including in the field of GIS and remote sensing.
The advent of Google Earth Engine (GEE) in particular has brought about a revolutionary change in the way we use geospatial technology. The GEE is a cloud based geospatial platform that stores petabyte of satellite imagery and geospatial data and enables carrying out of complex image processing tasks and spatial analyses without the need of any GIS and remote sensing software.
State of mago3D, An Open Source Based Digital Twin PlatformSANGHEE SHIN
I gave this talk at the FOSS4G Thailand 2019 which was held at Chulalongkorn University, Bangkok on 4th Nov 2019. I talked about the recent achievements and improvements of mago3D project, an open source based Digital Twin platform. mago3D(http mago3d.com) is relatively new project first released in July 2017. The ultimate goal of mago3D is developing an open source based digital twin platform that can replicate and simulate the real world objects, processes, and phenomena on web environment. mago3D has been used in various industry sectors including ship building, urban management, indoor data management, and national defense. In this talk I showcased several real projects that used the mago3D and shared what I learnt from these projects. Also I introduced new features and future plan of mago3D.
Current State of mago3D, an Open Source Based Digital Twin PlatformSANGHEE SHIN
I gave this talk at the FOSS4G 2019 Bucharest. I talked about the recent achievements and improvements of mago3D project, an open source based Digital Twin platform. mago3D(http mago3d.com) is relatively new project first released in July 2017. The ultimate goal of mago3D is developing an open source based digital twin platform that can replicate and simulate the real world objects, processes, and phenomena on web environment. mago3D is on its way to achieve this goal now. mago3D has been used in various industry sectors including ship building, urban management, indoor data management, and national defense. In this talk I showcased several real projects that employed the mago3D and shared what I learnt from these projects.
Introduction to mago3D: A Web Based Open Source GeoBIM PlatformSANGHEE SHIN
I gave this talk at the FOSS4G Asia 2018 held at University of Moratuwa, Sri Lanka. I've added some of recent improvements of mago3D features including CityGML, IndoorGML supporting. Also I've talked about the future plan of mago3D toward Digital Twin platform.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/qualcomm/embedded-vision-training/videos/pages/may-2016-embedded-vision-summit-mangen
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Michael Mangen, Product Manager for Camera and Computer Vision at Qualcomm, presents the "High-resolution 3D Reconstruction on a Mobile Processor" tutorial at the May 2016 Embedded Vision Summit.
Computer vision has come a long way. Use cases that were previously not possible in mass-market devices are now more accessible thanks to advances in depth sensors and mobile processors. In this presentation, Mangen provides an overview of how we are able to implement high-resolution 3D reconstruction – a capability typically requiring cloud/server processing – on a mobile processor. This is an exciting example of how new sensor technology and advanced mobile processors are bringing computer vision capabilities to broader markets.
Let's integrate CAD/BIM/GIS on the same platform: A practical approach in rea...SANGHEE SHIN
I gave this keynote talk at 3D GeoInfo conference held at Singapore on 25th September 2019. I shared my experiences in integrating CAD/BIM/GIS on the web platform and introduced mago3D from the technical point of view.
mago3D: Let's integrate BIM and 3D GIS on top of FOSS4GSANGHEE SHIN
This presentation was given at FOSS4G Europe 2017 which was held at ENSG, France.
***
Let's integrate BIM and 3D GIS on top of FOSS4G!
Sanghee Shin, Sungdo Son, BJ Jang, JeongDae Cheon, Hakjoon Kim
Although there has been numerous attempts to integrate indoor and outdoor space on a single geospatial platform, the outcome of these attempts are not so satisfactory till to date. Difference of data model, massive number of data to be rendered, big volume of file size are among those major technical barriers that hindered seamless integration of indoor and outdoor space. This talk will introduce a brand-new FOSS4G project called Mago3D that could seamlessly integrate BIM(Building Information Model) and popular 3D GIS model in web browser using Cesium or Web World Wind. Mago3D project aims at developing a JavaScript plug-in for existing WebGL Globe to expand WebGL Globe's functionality and usability even to indoor space and architectural(BIM) areas. To do this, Mago3D.js has been designed and developed as a WebGL independent JavaScript. Mago3D.js is composed of 6 main components like follows:
Mago3D Connector that interacts with WebGL Globe such as Cesium
Mago3D Renderer that shades and renders 3D data
Mago3D Accelerator that carries out performance enhancing algorithms such frustum & occlusion culling, indexing, LOD(Level Of Detail) handing
Mago3D Data Container that contains and manages 3D data
Mago3D Process Manager that manages whole process from data receiving to rendering
Mago3D REST API that provides API for 3D data sending and receiving
By plug in Mago3D.js to Cesium, Web World Wind, users can expand those WebGL functionality and usability to indoor space. One of big hurdle to integrate indoor and outdoor space simultaneously is handling and visualisation of massive 3D data. To overcome this hurdle, new format called F4D has been devised adopting block reference concept. Also a format converter that converts popular 3D format to F4D has been developed. Currently industry standard IFC(Industry Foundation Classes), JT(Jupiter Tessellation), and popular 3D formats such as OBJ, 3DS, COLLADA DAE can be converted to F4D format. F4D format coupled with Mago3D.js has proven that it can increase memory management efficiency and rendering speed drastically. MAGO3D can now visualise massive 3D data including indoor objects, at least 100k objects, in a single scene seamlessly with traditional outdoor 3D GIS objects. Users can now manage and handle almost every geospatial object from bolt & nut level to whole globe level with MAGO3D. This project will evolve to manage and service more dynamic data such as IoT (Internet of Things), point clouds, climate & weather data, and transportation..
mago3D: A brand new Geo-BIM platform on top of Cesium & World Wind SANGHEE SHIN
This talk is about how to integrate very large size BIM data and 3D GIS in a web browser. mago3D makes use of existing features of Cesium and Web World WInd. And mago3D proposes F4D as new service format for 3D service over the internet.
What is #PointCloudAnnotation? It is creating better data models for #MachineLearning algorithms!!
Read more at: https://annotationsupport.com/blog/what-is-point-cloud-annotation/
Interactive Editing of Signed Distance FieldsMatthias Trapp
Presentation of paper "Interactive Editing of Voxel-Based Signed Distance Fields" presented at the 30th International Conference in Central Europe on Computer Graphics, Visualization and Computer Vision (WSCG 2022).
LIDAR Magizine 2015: The Birth of 3D Mapping Artificial IntelligenceJason Creadore 🌐
Artificial intelligence (AI) has the potential to take the LiDAR
mapping market into hypergrowth. Following Moore’s law, with computation capacity doubling every 2 years, it is now possible for point cloud feature extraction to outpace the speed of data generation from laser scanning systems using artificial intelligence.
Let's integrate CAD/BIM/GIS on the same platform: A practical approach in rea...SANGHEE SHIN
I gave this keynote talk at 3D GeoInfo conference held at Singapore on 25th September 2019. I shared my experiences in integrating CAD/BIM/GIS on the web platform and introduced mago3D from the technical point of view.
mago3D: Let's integrate BIM and 3D GIS on top of FOSS4GSANGHEE SHIN
This presentation was given at FOSS4G Europe 2017 which was held at ENSG, France.
***
Let's integrate BIM and 3D GIS on top of FOSS4G!
Sanghee Shin, Sungdo Son, BJ Jang, JeongDae Cheon, Hakjoon Kim
Although there has been numerous attempts to integrate indoor and outdoor space on a single geospatial platform, the outcome of these attempts are not so satisfactory till to date. Difference of data model, massive number of data to be rendered, big volume of file size are among those major technical barriers that hindered seamless integration of indoor and outdoor space. This talk will introduce a brand-new FOSS4G project called Mago3D that could seamlessly integrate BIM(Building Information Model) and popular 3D GIS model in web browser using Cesium or Web World Wind. Mago3D project aims at developing a JavaScript plug-in for existing WebGL Globe to expand WebGL Globe's functionality and usability even to indoor space and architectural(BIM) areas. To do this, Mago3D.js has been designed and developed as a WebGL independent JavaScript. Mago3D.js is composed of 6 main components like follows:
Mago3D Connector that interacts with WebGL Globe such as Cesium
Mago3D Renderer that shades and renders 3D data
Mago3D Accelerator that carries out performance enhancing algorithms such frustum & occlusion culling, indexing, LOD(Level Of Detail) handing
Mago3D Data Container that contains and manages 3D data
Mago3D Process Manager that manages whole process from data receiving to rendering
Mago3D REST API that provides API for 3D data sending and receiving
By plug in Mago3D.js to Cesium, Web World Wind, users can expand those WebGL functionality and usability to indoor space. One of big hurdle to integrate indoor and outdoor space simultaneously is handling and visualisation of massive 3D data. To overcome this hurdle, new format called F4D has been devised adopting block reference concept. Also a format converter that converts popular 3D format to F4D has been developed. Currently industry standard IFC(Industry Foundation Classes), JT(Jupiter Tessellation), and popular 3D formats such as OBJ, 3DS, COLLADA DAE can be converted to F4D format. F4D format coupled with Mago3D.js has proven that it can increase memory management efficiency and rendering speed drastically. MAGO3D can now visualise massive 3D data including indoor objects, at least 100k objects, in a single scene seamlessly with traditional outdoor 3D GIS objects. Users can now manage and handle almost every geospatial object from bolt & nut level to whole globe level with MAGO3D. This project will evolve to manage and service more dynamic data such as IoT (Internet of Things), point clouds, climate & weather data, and transportation..
mago3D: A brand new Geo-BIM platform on top of Cesium & World Wind SANGHEE SHIN
This talk is about how to integrate very large size BIM data and 3D GIS in a web browser. mago3D makes use of existing features of Cesium and Web World WInd. And mago3D proposes F4D as new service format for 3D service over the internet.
What is #PointCloudAnnotation? It is creating better data models for #MachineLearning algorithms!!
Read more at: https://annotationsupport.com/blog/what-is-point-cloud-annotation/
Interactive Editing of Signed Distance FieldsMatthias Trapp
Presentation of paper "Interactive Editing of Voxel-Based Signed Distance Fields" presented at the 30th International Conference in Central Europe on Computer Graphics, Visualization and Computer Vision (WSCG 2022).
LIDAR Magizine 2015: The Birth of 3D Mapping Artificial IntelligenceJason Creadore 🌐
Artificial intelligence (AI) has the potential to take the LiDAR
mapping market into hypergrowth. Following Moore’s law, with computation capacity doubling every 2 years, it is now possible for point cloud feature extraction to outpace the speed of data generation from laser scanning systems using artificial intelligence.
Machine learning in the Indian Context - IEEE talk at SRM InstitutePrabindh Sundareson
An introduction to the key concepts of machine learning, and 4 case studies in the Indian context - Medicine, Language, Privacy, and Quality of life. Presented at SRM Institute, Chennai, Apr 2018
IEEE Communications Disability Special Interest Group and Visually Impaired Focus Group formed under the IEEE Standards Association is organizing its 2nd Hackathon on Enabling Ability in Disability through Standards & Technology hosted by R V College of Engineering, Mysore Rd, Bengaluru with the theme of Consumer Level Assistive Devices/ Applications. Winners would get to demo at ICCE-Asia 2017.
Updated Program Outline for ICCE Asia 2017 Bengaluru. Consists of keynotes from Cyient, Qualcomm, Revxx, IIT Madras, WiE Track, Panels, and Startup competitions on the floor.
Open shading language provides a direct way to represent properties of materials, and effects, rather than shaders that directly calculate a color or position. This presentation covers briefly the background of OSL.
Growth of CE and its critical role in various advances including sensors, displays, machine intelligence, automotive and other areas. Invited Lecture at Oxford.
GFX Part 1 - Introduction to GPU HW and OpenGL ES specificationsPrabindh Sundareson
Introduction to OpenGL ES and GPU Programming portion of the 7 part session on GFX workshops. Introduces the OpenGL ES specifications from Khronos and provides a perspective of current GPU architectures.
Water scarcity is the lack of fresh water resources to meet the standard water demand. There are two type of water scarcity. One is physical. The other is economic water scarcity.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Synthetic Data and Graphics Techniques in Robotics
1. Synthetic Data and
Graphics Techniques in Robotics
Prabindh Sundareson, 2022
National Symposium on Trends and Developments in
Intelligent Systems & Robotics (TD-ISR’22)
PES University Bangalore
Disclaimer: Views expressed are my own and does not represent the views of my employer
2. Agenda
• Role/Definition of Data
• Types of Data – Random, Time, 2D, 3D
• Graphics techniques – Applications in Intelligent systems design
• Part 1 – Synthetic data and 3D processing in Navigation
• Part 2 – Synthetic data and 2D objects in Vision processing
• Challenges
• Conclusion
3. Role of Data in Intelligent systems
• Function (task) Learning
• Fit a given set of data points to a function (generator)
• Feature Understanding
• How relevant is a feature, correlations across features
• Forecasting
• Probability of an event occurring, given some data
• And so on …
In Machine Learning, training with more data has led to better results
4. What is Data ?
• Simplest – “Random number”
• Also, time series, etc
• import sklearn.datasets as dt
RANDOM_SEED = np.random.randint(2**10)
• dt.make_multilabel_classification
5. What is Data (continued)
• Onto 2D data – What processing is involved ?
• Annotated images for image classification
• Labelled (bounding-box) images for object detection
• Segmented (pixel-wise) images for object segmentation
• These activities require effort !
• Currently being solved with human effort
6. What is Data (Continued)
• Onto 3D data – What type of data ?
• LiDAR point-clouds
• Orientation data
• 3D Meshes ..
• As dimensions increase, the data, annotations /ground-truth
generation gets more complex
7. Why synthetic data – In summary
• Cost and speed of acquisition and annotation (ground-truth –
boxing/labelling, segmentation ..…)
• Privacy issues
• Consequences are virtual
• Ex Robots in Medical systems
• Ability to configure – coarse to fine features
• Ex, In object detection
• ball vs non-ball (coarse), ball vs what-type-of-ball (fine features)
Synthetic Data Vault
https://sdv.dev
Topics
8. Data in 2D and 3D – Topics in Robotics and Graphics
TopicField Robotics field Graphics field
Efficient 3D
processing
3D points - Localisation
and mapping in Robotics
Synthetic data/
conversions/generation in
rendering 3D meshes
Kinematics Movement control in
Robotics
Animation in Graphics
Rendering
Pose Pose estimation for
orientation detection in
Robotics
Animation transfer to
arbitrary 3D meshes
9. Part 1 - Robotics Navigation and 3D
processing
• How do intelligent systems navigate an unknown area ?
• On-board sensors, on-board analysis
• Satellite connectivity, external inputs
• Sensors on UAV/drones, combined inputs
• Onboard sensors – Radar, LiDAR, ToF, Stereo cam, depth, IMU ..
• Obtain 3D map of the environment
• Flow
• Sensor → Data preprocessing (Ground plane, align)
• What can we do with a Point-cloud ?
• SLAM (Simultaneous Localisation and Mapping)
• Registration, Loop closure etc ..
• What else ?
https://www.ncos.co.jp/en/products/3dpoint.html
10. What “else” can be done with a Point-Cloud ?
• Learn “Geometry” via Signed Distance Field
“A Volumetric Method for Building Complex Models from Range Images”, Curless and Levoy, 1996
“DemoScene” in popular art sub-culture
https://www.youtube.com/watch?v=XuSnLbB1j6E
11. Implicit Representations in SDF
Steps
1. Raymarching step by step
2. Surface hit ? → Angle of ray, normals
SDF of circle (r=1)
f(x,y,z)= sqrt(x2+y2+z2)−1
Why
12. Why SDF ? Some commonalities
• Concepts of
• “distance” - Similar in 3D rendering and Robotics navigation
• “operations” – union, intersection etc (ex the Snail)
• Robotics – Distance provided by a sensor, ex LiDAR, or learnt from RGB
• Application to Robotics - “Signed Distance Fields: A Natural Representation
for Both Mapping and Planning” – Oleynikova et al, 2016
• How do we generate the SDF !!
*https://www.youtube.com/watch?v=bMhR-UYlKqU - 3D Surface Reconstruction Using a Generalized Distance Function, 2016, Microsoft Research
* Differentiable Volumetric Rendering: Learning Implicit 3D Representations without 3D Supervision – Niemeyer et al, CVPR 2020
* Neural Geometric Level of Detail: Real-time Rendering with Implicit 3D Shapes – Towaki et al 2021
* DeepSDF - Learning Continuous Signed Distance Functions for Shape Representation, Jeong Park et al, 2019
* On the Effectiveness of Weight-Encoded Neural Implicit 3D Shapes – Thomas Davies et al, ICML 2021 (https://github.com/u2ni/ICML2021 )
* Neural-Pull, Ma et al, ICML 2021 - https://github.com/mabaorui/NeuralPull-Pytorch , learns directly from 2D-images, does not require ground truth SDF
LiDAR
13. (a) SDFs from LiDAR (using neural networks)
• Learn SDFs directly from LiDAR point clouds
• “Learning Deep SDF Maps Online for Robot Navigation and Exploration” –
Camps et al, 2022
• Neural network to regress SDF (per-frame) from LiDAR data
• Hausdorff distance between successive SDFs used to determine mapping changes
• Optimization steps applied for generating navigation stages
RGB
14. (b) SDFs from RGB images/Silhoutte (using neural networks)
“SDF-SRN: Learning Signed Distance 3D Object Reconstruction from Static Images”, Lin et al, 2020
Kaolin
15. Tool - Differentiable Rendering for training
• “Differentiable rendering can be used to optimize the underlying 3D
properties, like geometry and lighting, by backpropagating gradients
from the loss in the image space”
• “Learning to Estimate 3D Object Pose from Synthetic Data”, Zakharov, 2020
• Tools:
• Kaolin - A Pytorch Library for Accelerating 3D Deep Learning Research
• Ex – Fit a 3D bounding box around a 2D image
• Example Kaolin link
• Ex – Classification using point-cloud directly
• https://colab.research.google.com/drive/1DoBlEt0GyOF5ZGrZ0kf30Gp5wk-dV2X9
Part 2 - GTA
16. Part 2 - Synthetic Datasets from 3D Games
• Precise Synthetic Image and LiDAR (PreSIL) Dataset for
Autonomous Vehicle Perception – Hurl et al, 2019
• From Grand Theft Auto V (game)
• By scanning the “depth” buffer
• 50k+ HD images with
• full resolution depth information,
• semantic segmentation (images),
• point-wise segmentation (point clouds), ground point labels (point
clouds), and detailed annotations for all vehicles and people
• Improvement of up to 5% average precision on the KITTI 3D
Object Detection
• Others - https://paralleldomain.com/
2D Aug
Don.riccobene
17. Synthetic 2D Datasets by Augmentation
• Augmentation = Add data by applying operations on existing images
• Augmentation can improve accuracy / robustness on Vision tasks
Robertas Damaševičius et al, 2021
19. Optimal 2D augmentation for object detection
1. Small original set of n-class representative images = input
2. Perform augmentation on original n-class images at runtime
3. 2D rectangle packing into larger canvas at runtime (dynamically)
4. Multiple augmentation-packing steps results in combined canvas images
5. Multiple-classes in one image reduces storage requirements significantly
Augmented + multi-class combined-canvas image
21. (b) Synthetic Data for Recognition – from 3D
• “An Annotation Saved is an Annotation Earned: Using Fully Synthetic
Training for Object Instance Detection.” –
• Hinterstoisser et al, Google AI, 2019 (Generate 2D images from 3D)
For more realistic data – Start from BIM (Building Information Models)
22. Synthetic Data – for
Drone Flight Controls
• AirSim works as a flight simulator
for drones
• Built as Unreal Engine plugin
• Allows drone producers to train
their machines to work in risky
and dangerous places.
• Roadmap to Autonomous driving
• “Aerial Informatics and Robotics
Platform”, Microsoft Research,
https://www.microsoft.com/en-
us/research/project/aerial-
informatics-robotics-platform/
Challenges
23. Challenges in Synthetic Data
(Object detection)
• Appearance gap –
• Pixel differences, lighting, shadows, textures, …
• Content gap –
• Missing/ new objects
• Domain Randomization – from Synthetic data to
real world data
• https://www.youtube.com/watch?v=O6KEKI3abl0&t=11s
• “Evaluation of Domain Randomization Techniques for
Transfer Learning” – Grun et al, 2019
• Brings results close to real-world datasets
Future
24. Future of Synthetic Data
• Definite improvements in performance in current use-cases
• Much of the data today is encoded for the “man-in-the-middle”
• Ex, Traffic signs
• Possible to generate and train on synthetic data
• That may be more efficient for Robotics
• Ex, glyphs that encode much more data than human-readable-signs
• More efficient for storage (or even no-storage, if metadata is stored)
• Synthetic Data could become the “real” data for
Intelligent systems of the future
25. Conclusion
• Synthetic Data is valuable, not just for augmentation purpose, but
also increases the overall robustness of recognition systems
• New techniques like online-SDF learning (using differentiable
rendering) can increase the capabilities of robotics systems to learn
from surroundings
• Possibility of machine learning without intermediate human
involvement or “interpretation” step
27. Neural Radiance Fields
• Radiance - a differential opacity controlling how much radiance is
accumulated by a ray passing through (x, y, z)
• Utilises Differentiable volume rendering, ray marching
• “NeRF: Representing Scenes as Neural Radiance Fields for View
Synthesis”, MildenHall et al, ECCV 2020 (https://github.com/bmild/nerf)
• Represent a continuous scene as a 5D vector-valued function whose input is a
3D location x = (x, y, z) and 2D viewing direction (θ, φ), and whose output is
an emitted color c = (r, g, b) and volume density σ
28. Instant NeRF - network
• Network configuration ?
• Uses Huber Loss
• Requires camera position
• NeRF networks with joint optimization for camera position ie
operating only on image frames also emerging
30. COLMAP steps
1.Feature detection and extraction
2.Feature matching and geometric verification
3.Structure and motion reconstruction
• up vector was [-0.90487011 0.10419316 0.41273947]
• computing center of attention...
• [-12.27771833 1.27148092 -0.09314177]
• avg camera distance from origin 12.701290776903985
• 55 frames
• writing transforms.json
31. Adoption of feature-encoding layers
• As applicable to “Coordinate networks”
• “Fourier Features Let Networks Learn High Frequency Functions in
Low Dimensional Domains” – Tancik et al, 2020
• Concept of Neural Tangent Kernels (NTK)
• Faster convergence
• Examples
• https://colab.research.google.com/github/ndahlquist/pytorch-
fourier-feature-networks/blob/master/demo.ipynb
32. Differentiable Volume Rendering
• Makes a mesh “learnable” from the sensor data
• LiDAR data
• Eikonal Equation
||∇f||=1
Learns a SDF
37. Intelligence for Autonomous Agents
• Ex Minecraft
• Survival
• Harvest
• Combat
• Training with 750,000 Minecraft YouTube videos, scraping from
Minecraft Wiki, ...
• Tasks
• Programmatic tasks
• Creative tasks.
• https://github.com/MineDojo/MineDojo
38. 2D-3D
• EG3D: Efficient Geometry-aware 3D Generative Adversarial Networks
- Chan et al, 2022
• Unsupervised generation of high-quality multi-view-consistent images
and 3D shapes using only collections of single-view 2D photographs
39. Topic 2 - Pose
• Important for gripping in Robotics
• PVN3D - 3D keypoint estimator from 2D images
• Trained via Synthetic data, won 2nd in OCTOC challenge
40. Pose in Graphics Animation
• 2D video to 3D avatars
• Translate pose in 2D to any 3D object
• Steps
• Body-pose estimation (2D)
• Similar techniques as in Robotics, applied to human body landmarks
• 2D to 3D mapping
• Map to synthetic object’s keypoints
• Mapping can be innovative and depends on the use-case
41. SDF from point clouds (Traditional)
• (CSCI 621: Digital Geometry Processing), 2019
• Construct SDF from point samples - Distance to points is not enough
• • Need inside/outside information
• • Requires normal vectors
• • Examine local neighborhood for each point
• • Compute best approximating tangent plane , covariance, MST tree construction,..
• Implicit Reconstruction - Estimate signed distance function (SDF) •
Extract Zero isosurface by Marching Cubes • Approximation of input
points • Result is closed two-manifold surface
42. Does data change ?
• Early stage research vs productization challenges
• Conditions for data
• Sufficiency and relevancy
• Data needs to be relevant to the task
• Some data = static datasets
• Production goals demand new data, continuous training
43. Types of Synthetic Data
• Time-series (Movement, Distance, Sensory, ..)
• Visual data
• Augmented Natural Data
• Rotated, Resized, color ranges, Blurred, Lit, …
• Synthetic Data
• Generated by simulations (ex Robotic movements)
• Generated by Graphics API (ex Synthetic face/body, animation)
• Combined
• Generated by human in a virtual environment
• ex Flight simulators
https://varjo.com
44. Synthetic data generation
• “Act against Multiplication” – 1404 Law,
outlawing “synthetic creation” of gold and
silver, Britain
• Today, we can realistically generate any data
(almost)
• Sensor Data (Gazebo, Webots.. simulators)
• Visual Data (Unity, UE, Omniverse, ..)
• Columnar Data (numpy, SDV, Gretel ..)
• Synthetic != Fake, has to match the use-case
DALL-E-2
Why
45. LiDAR vs Depth
• LiDAR more accurate, but object segregation difficult
• A fusion of depth, RGB required for good results in both navigation
and detection use-cases