The acquisition of labeled, unbiased, high quality remote sensing information for training AI systems is expensive, error prone, and sometimes impossible or dangerous. The efficacy of Remote Sensing and Imagery Analysis tools that use AI depends directly on the data used for training and validation, meaning that the cost and availability of data limits the application of AI for imagery exploitation. Synthetic Computer Vision (CV) data has become a strategy to reduce the cost and limitations of using real-world data in detection problems in data sparse domains. Focusing on remote sensing data including visible and invisible electromagnetic spectra, attendees will learn about the expanding options for generating synthetic data that are being used in commercial and academic domains, the technology options available for users who want to create CV content of a variety of types, and patterns of creating synthetic data to support
Learning Objectives
- Describe synthetic data including different types such as Generative AI and physics-based data
- Identify the opportunities for applying synthetic data in place of real sensor data
Will be able to describe the steps required to generate synthetic data for computer vision workflows from concept to production for training and validating AI.
- The intent of this class is to introduce the concepts and mechanisms behind the creation of synthetic data and to expose students to approaches for generating synthetic data using tools currently on the market.
- Familiarity with concepts around AI training and validation using remotely sensed data will be helpful for attendees.
10 Key Considerations for AI/ML Model GovernanceQuantUniversity
As the financial industry continues to embrace AI and Machine Learning models, model risk management (MRM) departments are grappling with challenges on how to update model governance frameworks to adapt to the changing landscape of model management. While most MRM departments are structured and processes defined to address traditional statistical and quant models, data-driven models like Machine Learning models require modifications in the way models are defined, tested, validated, and governed.
In this webinar, we will discuss ten key aspects to factor when developing your model risk management framework when integrating Machine Learning models. We will discuss key drivers of model risk in today’s environment and how the scope of model governance is changing. We will introduce key concepts and discuss key aspects to be considered when developing a model governance framework when incorporating data science techniques and AI methodologies. Through this Decalogue, we aim to bring clarity on some of the model governance challenges when adopting data science, AI and machine learning methods in the enterprise.
Processing 3D images has many use cases. For example, to improve autonomous car driving, to enable digital conversions of old factory buildings, to enable augmented reality solutions for medical surgeries, etc. Also 3D images help in 3D modeling and safety evaluation of products.
3D image processing brings enormous benefits but also amplifies computing cost. The size of the point cloud, the number of points, sparse and irregular point cloud, and the adverse impact of the light reflections, (partial) occlusions, etc., make it difficult for engineers to process point clouds.
Moving from using hand crafted features to using deep learning techniques to semantically segment the images, to classify objects, to detect objects, to detect actions in 3D videos, etc., we have come a long way in 3D image processing.
3D Point Cloud image processing is increasingly used to solve Industry 4.0 use cases to help architects, builders and product managers. I will share some of the innovations that are helping the progress of 3D point cloud processing. I will share the practical implementation issues we faced while developing deep learning models to make sense of 3D Point Clouds.
Attendees: Beginners and Intermediate skilled in Image Processing and 3D Point Clouds
Profile of the speaker:
SK Reddy is the Chief Product Officer AI in Hexagon (www.hexagon.com). He is an AI and ML expert and a successful twice startup entrepreneur. He is an AI startup advisor too. Also he is a frequent speaker in conferences and is an AI blogger.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/fotonation/embedded-vision-training/videos/pages/may-2018-embedded-vision-summit-corcoran-tuesday
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Peter Corcoran, co-founder of FotoNation (now a core business unit of Xperi) and lead principle investigator and director of C3Imaging (a research partnership between Xperi and the National University of Ireland, Galway), presents the "Getting More from Your Datasets: Data Augmentation, Annotation and Generative Techniques" tutorial at the May 2018 Embedded Vision Summit.
Deep learning for embedded vision requires large datasets. Indeed, the more varied the training data is, the more accurate the resultant trained network tends to also be. But, acquiring and accurately annotating datasets costs time and money. This talk shows how to get more out of existing datasets.
First, state-of-art data augmentation techniques are reviewed, and a new approach, smart augmentation, is explained. Next, GANs (generative adversarial networks) that learn the structure of an existing dataset are explained; several example use cases (such as creating a very large dataset of facial training data) show how GANs can generate new data corresponding to the original dataset.
But building a dataset does not by itself represent the entirety of the challenge; data must also be annotated in a way that is meaningful for the training process. The presentation then gives an example of training a GAN from a dataset that incorporates annotations. This technique enables the generation of pre-annotated data" providing an exciting way to create large datasets at significantly reduced costs.
This talk was jointly organized by BSPIN and ASQ Bengaluru LMC.
It covered the following details-
- Industry 4.0
- Agile , CI/CD, DevOps
- DevOps and MLOps
- Evolution of MLOPS
- MLOps Capabilities
- AI Platform Pipelines
- Training and Tuning AI Platform
- Case Study
Accelerate AI w/ Synthetic Data using GANsRenee Yao
Strata Data Conference in Sep 2018 Presentation
Description:
Synthetic data will drive the next wave of deployment and application of deep learning in the real world across a variety of problems involving speech recognition, image classification, object recognition and language. All industries and companies will benefit, as synthetic data can create conditions through simulation, instead of authentic situations (virtual worlds enable you to avoid the cost of damages, spare human injuries, and other factors that come into play); unparalleled ability to test products, and interactions with them in any environment.
Join us for this introductory session to learn more about how Generative Adversarial Networks (GAN) are successfully used to improve data generation. We will cover specific real-world examples where customers have deployed GAN to solve challenges in healthcare, space, transportation, and retail industries.
Renee Yao explains how generative adversarial networks (GAN) are successfully used to improve data generation and explores specific real-world examples where customers have deployed GANs to solve challenges in healthcare, space, transportation, and retail industries.
10 Key Considerations for AI/ML Model GovernanceQuantUniversity
As the financial industry continues to embrace AI and Machine Learning models, model risk management (MRM) departments are grappling with challenges on how to update model governance frameworks to adapt to the changing landscape of model management. While most MRM departments are structured and processes defined to address traditional statistical and quant models, data-driven models like Machine Learning models require modifications in the way models are defined, tested, validated, and governed.
In this webinar, we will discuss ten key aspects to factor when developing your model risk management framework when integrating Machine Learning models. We will discuss key drivers of model risk in today’s environment and how the scope of model governance is changing. We will introduce key concepts and discuss key aspects to be considered when developing a model governance framework when incorporating data science techniques and AI methodologies. Through this Decalogue, we aim to bring clarity on some of the model governance challenges when adopting data science, AI and machine learning methods in the enterprise.
Processing 3D images has many use cases. For example, to improve autonomous car driving, to enable digital conversions of old factory buildings, to enable augmented reality solutions for medical surgeries, etc. Also 3D images help in 3D modeling and safety evaluation of products.
3D image processing brings enormous benefits but also amplifies computing cost. The size of the point cloud, the number of points, sparse and irregular point cloud, and the adverse impact of the light reflections, (partial) occlusions, etc., make it difficult for engineers to process point clouds.
Moving from using hand crafted features to using deep learning techniques to semantically segment the images, to classify objects, to detect objects, to detect actions in 3D videos, etc., we have come a long way in 3D image processing.
3D Point Cloud image processing is increasingly used to solve Industry 4.0 use cases to help architects, builders and product managers. I will share some of the innovations that are helping the progress of 3D point cloud processing. I will share the practical implementation issues we faced while developing deep learning models to make sense of 3D Point Clouds.
Attendees: Beginners and Intermediate skilled in Image Processing and 3D Point Clouds
Profile of the speaker:
SK Reddy is the Chief Product Officer AI in Hexagon (www.hexagon.com). He is an AI and ML expert and a successful twice startup entrepreneur. He is an AI startup advisor too. Also he is a frequent speaker in conferences and is an AI blogger.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/fotonation/embedded-vision-training/videos/pages/may-2018-embedded-vision-summit-corcoran-tuesday
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Peter Corcoran, co-founder of FotoNation (now a core business unit of Xperi) and lead principle investigator and director of C3Imaging (a research partnership between Xperi and the National University of Ireland, Galway), presents the "Getting More from Your Datasets: Data Augmentation, Annotation and Generative Techniques" tutorial at the May 2018 Embedded Vision Summit.
Deep learning for embedded vision requires large datasets. Indeed, the more varied the training data is, the more accurate the resultant trained network tends to also be. But, acquiring and accurately annotating datasets costs time and money. This talk shows how to get more out of existing datasets.
First, state-of-art data augmentation techniques are reviewed, and a new approach, smart augmentation, is explained. Next, GANs (generative adversarial networks) that learn the structure of an existing dataset are explained; several example use cases (such as creating a very large dataset of facial training data) show how GANs can generate new data corresponding to the original dataset.
But building a dataset does not by itself represent the entirety of the challenge; data must also be annotated in a way that is meaningful for the training process. The presentation then gives an example of training a GAN from a dataset that incorporates annotations. This technique enables the generation of pre-annotated data" providing an exciting way to create large datasets at significantly reduced costs.
This talk was jointly organized by BSPIN and ASQ Bengaluru LMC.
It covered the following details-
- Industry 4.0
- Agile , CI/CD, DevOps
- DevOps and MLOps
- Evolution of MLOPS
- MLOps Capabilities
- AI Platform Pipelines
- Training and Tuning AI Platform
- Case Study
Accelerate AI w/ Synthetic Data using GANsRenee Yao
Strata Data Conference in Sep 2018 Presentation
Description:
Synthetic data will drive the next wave of deployment and application of deep learning in the real world across a variety of problems involving speech recognition, image classification, object recognition and language. All industries and companies will benefit, as synthetic data can create conditions through simulation, instead of authentic situations (virtual worlds enable you to avoid the cost of damages, spare human injuries, and other factors that come into play); unparalleled ability to test products, and interactions with them in any environment.
Join us for this introductory session to learn more about how Generative Adversarial Networks (GAN) are successfully used to improve data generation. We will cover specific real-world examples where customers have deployed GAN to solve challenges in healthcare, space, transportation, and retail industries.
Renee Yao explains how generative adversarial networks (GAN) are successfully used to improve data generation and explores specific real-world examples where customers have deployed GANs to solve challenges in healthcare, space, transportation, and retail industries.
[DSC DACH 23] ChatGPT and Beyond: How generative AI is Changing the way peopl...DataScienceConferenc1
In recent years, generative AI has made significant advancements in language understanding and generation, leading to the development of chatbots like ChatGPT. These models have the potential to change the way people interact with technology. In this session, we will explore the advancements in generative AI. I will show how these models have evolved, their strengths and limitations, and their potential for improving various applications. Additionally, I will show some of the ethical considerations that arise from the use of these models and their impact on society.
GenerativeAI and Automation - IEEE ACSOS 2023.pptxAllen Chan
Generative AI has been rapidly evolving, enabling different and more sophisticated interactions with Large Language Models (LLMs) like those available in IBM watsonx.ai or Meta Llama2. In this session, we will take a use case based approach to look at how we can leverage LLMs together with existing automation technologies like Workflow, Content Management, and Decisions to enable new solutions.
Команда Data Phoenix Events приглашает всех, 17 августа в 19:00, на первый вебинар из серии "The A-Z of Data", который будет посвящен MLOps. В рамках вводного вебинара, мы рассмотрим, что такое MLOps, основные принципы и практики, лучшие инструменты и возможные архитектуры. Мы начнем с простого жизненного цикла разработки ML решений и закончим сложным, максимально автоматизированным, циклом, который нам позволяет реализовать MLOps.
https://dataphoenix.info/the-a-z-of-data/
https://dataphoenix.info/the-a-z-of-data-introduction-to-mlops/
PR-297: Training data-efficient image transformers & distillation through att...Jinwon Lee
안녕하세요 TensorFlow Korea 논문 읽기 모임 PR-12의 297번째 리뷰입니다
어느덧 PR-12 시즌 3의 끝까지 논문 3편밖에 남지 않았네요.
시즌 3가 끝나면 바로 시즌 4의 새 멤버 모집이 시작될 예정입니다. 많은 관심과 지원 부탁드립니다~~
(멤버 모집 공지는 Facebook TensorFlow Korea 그룹에 올라올 예정입니다)
오늘 제가 리뷰한 논문은 Facebook의 Training data-efficient image transformers & distillation through attention 입니다.
Google에서 나왔던 ViT논문 이후에 convolution을 전혀 사용하지 않고 오직 attention만을 이용한 computer vision algorithm에 어느때보다 관심이 높아지고 있는데요
이 논문에서 제안한 DeiT 모델은 ViT와 같은 architecture를 사용하면서 ViT가 ImageNet data만으로는 성능이 잘 안나왔던 것에 비해서
Training 방법 개선과 새로운 Knowledge Distillation 방법을 사용하여 mageNet data 만으로 EfficientNet보다 뛰어난 성능을 보여주는 결과를 얻었습니다.
정말 CNN은 이제 서서히 사라지게 되는 것일까요? Attention이 computer vision도 정복하게 될 것인지....
개인적으로는 당분간은 attention 기반의 CV 논문이 쏟아질 거라고 확신하고, 또 여기에서 놀라운 일들이 일어날 수 있을 거라고 생각하고 있습니다
CNN은 10년간 많은 연구를 통해서 발전해왔지만, transformer는 이제 CV에 적용된 지 얼마 안된 시점이라서 더 기대가 크구요,
attention이 inductive bias가 가장 적은 형태의 모델이기 때문에 더 놀라운 이들을 만들 수 있을거라고 생각합니다
얼마 전에 나온 open AI의 DALL-E도 그 대표적인 예라고 할 수 있을 것 같습니다. Transformer의 또하나의 transformation이 궁금하신 분들은 아래 영상을 참고해주세요
영상링크: https://youtu.be/DjEvzeiWBTo
논문링크: https://arxiv.org/abs/2012.12877
Unlocking the Power of Generative AI An Executive's Guide.pdfPremNaraindas1
Generative AI is here, and it can revolutionize your business. With its powerful capabilities, this technology can help companies create more efficient processes, unlock new insights from data, and drive innovation. But how do you make the most of these opportunities?
This guide will provide you with the information and resources needed to understand the ins and outs of Generative AI, so you can make informed decisions and capitalize on the potential. It covers important topics such as strategies for leveraging large language models, optimizing MLOps processes, and best practices for building with Generative AI.
The Transformation of Talent Management presentation takes an in depth look at the challenges facing the HR community as professionals attempt to navigate the changing Talent Management landscape. The piece is full of insights and thought leadership. The data was gathered as part of a survey conducted by the HR Exchange Network on the topic.
Presentation "AI Product Manager" at the Digital Product School (on 10/22/2020) from Datentreiber.
Content:
• Overview over the AI product innovation cycle
• AI Thinking: ideating and prioritizing the right use cases
• AI Prototyping: testing critical hypotheses with experiments
• AI Engineering: building scalable & user friendly AI applications
• AI Management: maintaining AI solutions with DataOps
• Outlook: how to become an AI product manager (links & more)
The Future of AI is Generative not Discriminative 5/26/2021Steve Omohundro
The deep learning AI revolution has been sweeping the world for a decade now. Deep neural nets are routinely used for tasks like translation, fraud detection, and image classification. PwC estimates that they will create $15.7 trillion/year of value by 2030. But most current networks are "discriminative" in that they directly map inputs to predictions. This type of model requires lots of training examples, doesn't generalize well outside of its training set, creates inscrutable representations, is subject to adversarial examples, and makes knowledge transfer difficult. People, in contrast, can learn from just a few examples, generalize far beyond their experience, and can easily transfer and reuse knowledge. In recent years, new kinds of "generative" AI models have begun to exhibit these desirable human characteristics. They represent the causal generative processes by which the data is created and can be compositional, compact, and directly interpretable. Generative AI systems that assist people can model their needs and desires and interact with empathy. Their adaptability to changing circumstances will likely be required by rapidly changing AI-driven business and social systems. Generative AI will be the engine of future AI innovation.
The catalyst for the success of automobiles came not through the invention of the car but rather through the establishment of an innovative assembly line. History shows us that the ability to mass produce and distribute a product is the key to driving adoption of any innovation, and machine learning is no different. MLOps is the assembly line of Machine Learning and in this presentation we will discuss the core capabilities your organization should be focused on to implement a successful MLOps system.
How Talent Analytics Can Help You Maximize Your HR StrategyGlassdoor
For most organizations, the promise of Big Data remains unfulfilled. The vast majority of organizations are stuck in a reporting cycle, churning out lots of metrics, but few insights or solutions. The ability to measure, analyze, and optimize talent practices is now critical to business success.
Many HR organizations have recognized this need and are starting to invest more strategically in measurement and analytics. With a plethora of data, recruiting is an area ripe to take advantage of analytics. With the right tools and capabilities, this data can be turned into competitive advantage.
Check out our webinar feat. Karen O'Leonard, VP of Benchmarking & Analytics Research of Bersin by Deloitte and Wiliam Blackstorm, Sr. Manager Sourcing & Market Intelligence & Director of Global Talent Analytics, Research Division of Cisco to learn:
-Where to start when analyzing recruitment data
-How to build an effective talent analytics capability
-How one organization, Cisco, is using analytics to develop a more effective recruitment strategy
Rendered.ai - Intro to Synthetic data for Computer Vision.pdfChris Andrews
Rendered.ai is a Platform as a Service that enables data scientists, data engineers, and developers to create and deploy unlimited, customized synthetic data generation for computer vision-related machine learning and artificial intelligence workflows, reducing expense, closing gaps, and overcoming bias, security, and privacy issues when compared with the use or acquisition of real data.
Visit https://www.rendered.ai
This guide was put together to help leaders who are working with teams who struggle to build computer vision algorithms using the data that they have available.
Find out how Rendered.ai helps overcome time to market and performance issues when training CV algorithms.
[DSC DACH 23] ChatGPT and Beyond: How generative AI is Changing the way peopl...DataScienceConferenc1
In recent years, generative AI has made significant advancements in language understanding and generation, leading to the development of chatbots like ChatGPT. These models have the potential to change the way people interact with technology. In this session, we will explore the advancements in generative AI. I will show how these models have evolved, their strengths and limitations, and their potential for improving various applications. Additionally, I will show some of the ethical considerations that arise from the use of these models and their impact on society.
GenerativeAI and Automation - IEEE ACSOS 2023.pptxAllen Chan
Generative AI has been rapidly evolving, enabling different and more sophisticated interactions with Large Language Models (LLMs) like those available in IBM watsonx.ai or Meta Llama2. In this session, we will take a use case based approach to look at how we can leverage LLMs together with existing automation technologies like Workflow, Content Management, and Decisions to enable new solutions.
Команда Data Phoenix Events приглашает всех, 17 августа в 19:00, на первый вебинар из серии "The A-Z of Data", который будет посвящен MLOps. В рамках вводного вебинара, мы рассмотрим, что такое MLOps, основные принципы и практики, лучшие инструменты и возможные архитектуры. Мы начнем с простого жизненного цикла разработки ML решений и закончим сложным, максимально автоматизированным, циклом, который нам позволяет реализовать MLOps.
https://dataphoenix.info/the-a-z-of-data/
https://dataphoenix.info/the-a-z-of-data-introduction-to-mlops/
PR-297: Training data-efficient image transformers & distillation through att...Jinwon Lee
안녕하세요 TensorFlow Korea 논문 읽기 모임 PR-12의 297번째 리뷰입니다
어느덧 PR-12 시즌 3의 끝까지 논문 3편밖에 남지 않았네요.
시즌 3가 끝나면 바로 시즌 4의 새 멤버 모집이 시작될 예정입니다. 많은 관심과 지원 부탁드립니다~~
(멤버 모집 공지는 Facebook TensorFlow Korea 그룹에 올라올 예정입니다)
오늘 제가 리뷰한 논문은 Facebook의 Training data-efficient image transformers & distillation through attention 입니다.
Google에서 나왔던 ViT논문 이후에 convolution을 전혀 사용하지 않고 오직 attention만을 이용한 computer vision algorithm에 어느때보다 관심이 높아지고 있는데요
이 논문에서 제안한 DeiT 모델은 ViT와 같은 architecture를 사용하면서 ViT가 ImageNet data만으로는 성능이 잘 안나왔던 것에 비해서
Training 방법 개선과 새로운 Knowledge Distillation 방법을 사용하여 mageNet data 만으로 EfficientNet보다 뛰어난 성능을 보여주는 결과를 얻었습니다.
정말 CNN은 이제 서서히 사라지게 되는 것일까요? Attention이 computer vision도 정복하게 될 것인지....
개인적으로는 당분간은 attention 기반의 CV 논문이 쏟아질 거라고 확신하고, 또 여기에서 놀라운 일들이 일어날 수 있을 거라고 생각하고 있습니다
CNN은 10년간 많은 연구를 통해서 발전해왔지만, transformer는 이제 CV에 적용된 지 얼마 안된 시점이라서 더 기대가 크구요,
attention이 inductive bias가 가장 적은 형태의 모델이기 때문에 더 놀라운 이들을 만들 수 있을거라고 생각합니다
얼마 전에 나온 open AI의 DALL-E도 그 대표적인 예라고 할 수 있을 것 같습니다. Transformer의 또하나의 transformation이 궁금하신 분들은 아래 영상을 참고해주세요
영상링크: https://youtu.be/DjEvzeiWBTo
논문링크: https://arxiv.org/abs/2012.12877
Unlocking the Power of Generative AI An Executive's Guide.pdfPremNaraindas1
Generative AI is here, and it can revolutionize your business. With its powerful capabilities, this technology can help companies create more efficient processes, unlock new insights from data, and drive innovation. But how do you make the most of these opportunities?
This guide will provide you with the information and resources needed to understand the ins and outs of Generative AI, so you can make informed decisions and capitalize on the potential. It covers important topics such as strategies for leveraging large language models, optimizing MLOps processes, and best practices for building with Generative AI.
The Transformation of Talent Management presentation takes an in depth look at the challenges facing the HR community as professionals attempt to navigate the changing Talent Management landscape. The piece is full of insights and thought leadership. The data was gathered as part of a survey conducted by the HR Exchange Network on the topic.
Presentation "AI Product Manager" at the Digital Product School (on 10/22/2020) from Datentreiber.
Content:
• Overview over the AI product innovation cycle
• AI Thinking: ideating and prioritizing the right use cases
• AI Prototyping: testing critical hypotheses with experiments
• AI Engineering: building scalable & user friendly AI applications
• AI Management: maintaining AI solutions with DataOps
• Outlook: how to become an AI product manager (links & more)
The Future of AI is Generative not Discriminative 5/26/2021Steve Omohundro
The deep learning AI revolution has been sweeping the world for a decade now. Deep neural nets are routinely used for tasks like translation, fraud detection, and image classification. PwC estimates that they will create $15.7 trillion/year of value by 2030. But most current networks are "discriminative" in that they directly map inputs to predictions. This type of model requires lots of training examples, doesn't generalize well outside of its training set, creates inscrutable representations, is subject to adversarial examples, and makes knowledge transfer difficult. People, in contrast, can learn from just a few examples, generalize far beyond their experience, and can easily transfer and reuse knowledge. In recent years, new kinds of "generative" AI models have begun to exhibit these desirable human characteristics. They represent the causal generative processes by which the data is created and can be compositional, compact, and directly interpretable. Generative AI systems that assist people can model their needs and desires and interact with empathy. Their adaptability to changing circumstances will likely be required by rapidly changing AI-driven business and social systems. Generative AI will be the engine of future AI innovation.
The catalyst for the success of automobiles came not through the invention of the car but rather through the establishment of an innovative assembly line. History shows us that the ability to mass produce and distribute a product is the key to driving adoption of any innovation, and machine learning is no different. MLOps is the assembly line of Machine Learning and in this presentation we will discuss the core capabilities your organization should be focused on to implement a successful MLOps system.
How Talent Analytics Can Help You Maximize Your HR StrategyGlassdoor
For most organizations, the promise of Big Data remains unfulfilled. The vast majority of organizations are stuck in a reporting cycle, churning out lots of metrics, but few insights or solutions. The ability to measure, analyze, and optimize talent practices is now critical to business success.
Many HR organizations have recognized this need and are starting to invest more strategically in measurement and analytics. With a plethora of data, recruiting is an area ripe to take advantage of analytics. With the right tools and capabilities, this data can be turned into competitive advantage.
Check out our webinar feat. Karen O'Leonard, VP of Benchmarking & Analytics Research of Bersin by Deloitte and Wiliam Blackstorm, Sr. Manager Sourcing & Market Intelligence & Director of Global Talent Analytics, Research Division of Cisco to learn:
-Where to start when analyzing recruitment data
-How to build an effective talent analytics capability
-How one organization, Cisco, is using analytics to develop a more effective recruitment strategy
Rendered.ai - Intro to Synthetic data for Computer Vision.pdfChris Andrews
Rendered.ai is a Platform as a Service that enables data scientists, data engineers, and developers to create and deploy unlimited, customized synthetic data generation for computer vision-related machine learning and artificial intelligence workflows, reducing expense, closing gaps, and overcoming bias, security, and privacy issues when compared with the use or acquisition of real data.
Visit https://www.rendered.ai
This guide was put together to help leaders who are working with teams who struggle to build computer vision algorithms using the data that they have available.
Find out how Rendered.ai helps overcome time to market and performance issues when training CV algorithms.
Platform for Big Data Analytics and Visual Analytics: CSIRO use cases. Februa...Tomasz Bednarz
Presented at the ACEMS workshop at QUT in February 2015.
Credits: whole project team (names listed in the first slide).
Approved by CSIRO to be shared externally.
Watch full webinar here: https://bit.ly/3H4vrlD
Data as a strategic imperative for any company to compete, New common self-service data experience required for all things intelligent, Modern data platform focused on producing data products, Data platform, product, people, process key solution ingredients and Denodo is the future and time is now to get started.
Arocom is a consulting and solution engineering company with expertise in providing engineering services for AI & Machine Learning, Data Operations & Analytics, MLOps and Cloud Computing.
Our clients include companies within biotech, drug discovery, therapeutics, manufacturing, retail and startups. Our consultants are best in their skills and offer hands-on talent to our clients in achieving their goals.
Vertex perspectives ai optimized chipsets (part i)Yanai Oron
Businesses are increasingly adopting AI to create new applications to transform existing operations, driving big data with the growth of IoT and 5G networks and increasing future process complexities for human operators. In this new environment, AI will be needed to write algorithms dynamically to automate the entire programming process. Fortunately, algorithms associated with deep learning are able to achieve enhanced performance with increasing data, unlike the rest associated with machine learning.
Vertex Perspectives | AI-optimized Chipsets | Part IVertex Holdings
Businesses are increasingly adopting AI to create new applications to transform existing operations, driving big data with the growth of IoT and 5G networks and increasing future process complexities for human operators. In this new environment, AI will be needed to write algorithms dynamically to automate the entire programming process. Fortunately, algorithms associated with deep learning are able to achieve enhanced performance with increasing data, unlike the rest associated with machine learning. To date, deep learning technology has primarily been a software play. Existing processors were not originally designed for these new applications. Hence the need to develop AI-optimized hardware.
As the adoption of AI technologies increases and matures, the focus will shift from exploration to time to market, productivity and integration with existing workflows. Governing Enterprise data, scaling AI model development, selecting a complete, collaborative hybrid platform and tools for rapid solution deployments are key focus areas for growing data scientist teams tasked to respond to business challenges. This talk will cover the challenges and innovations for AI at scale for the Enterprise focusing on the modernization of data analytics, the AI ladder and AI life cycle and infrastructure architecture considerations. We will conclude by viewing the benefits and innovation of running your modern AI and Data Analytics applications such as SAS Viya and SAP HANA on IBM Power Systems and IBM Storage in hybrid cloud environments.
ICP for Data- Enterprise platform for AI, ML and Data ScienceKaran Sachdeva
IBM Cloud Private for Data, an ultimate platform for all AI, ML and Data Science workloads. Integrated analytics platform based on Containers and micro services. Works with Kubernetes and dockers, even with Redhat openshift. Delivers the variety of business use cases in all industries- FS, Telco, Retail, Manufacturing etc
AI for Manufacturing (Machine Vision, Edge AI, Federated Learning)byteLAKE
This is the extended presentation about byteLAKE's and Lenovo's Artificial Intelligence solutions for Manufacturing.
Topics covered: AI strategy for manufacturing, Edge AI, Federated Learning and Machine Vision.
It's the first publication in the upcoming series: AI for Manufacturing. Highlights: AI-assisted quality monitoring automation, AI-assisted production line monitoring and issues detection, AI-assisted measurements, Intelligent Cameras and many more. Reach out to us to learn more: welcome@byteLAKE.com.
Presented during the world's first Federated Learning conference (Jun'20). Recording: https://youtu.be/IMqRIi45dDA
Related articles:
- Revolution in factories: Industry 4.0.
https://medium.com/@marcrojek/revolution-in-factories-industry-4-0-conference-made-in-wroclaw-2020-translation-ae96e5e14d55
- Cognitive Automation helps where RPAs fall short.
https://medium.com/@marcrojek/cognitive-automation-helps-where-rpas-fall-short-a1c5a01a66f8
- Machine Vision, how AI brings value to industries.
https://medium.com/@marcrojek/machine-vision-how-ai-brings-value-to-industries-e6a4f8e56f42
Learn more:
- https://www.bytelake.com/en/cognitive-services/
- https://www.lenovo.com/ai
- https://federatedlearningconference.com/
Open source Apache Hadoop is a great framework for distributed processing of large data sets. But there’s a difference between “playing” with big data versus solving real problems. The reality is that Hadoop alone is not enough. In fact, almost every organization that plans to use Hadoop for production use quickly discovers that it lacks the required features for enterprise use. And, fewer still have the Hadoop specialists on hand to navigate through the complexity to build reliable, robust applications. As a result, many Hadoop projects never make it to production as executives say, “we just don’t have the skills.” In this session, we will discuss these enterprise capabilities and why they’re important: analytics, visualization, security, enterprise integration, developer/admin tools, and more. Additionally, we will share several real-world client examples who have found it necessary to use an enterprise-grade Hadoop platform to tackle some of the most interesting and challenging business problems.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Enhancing Performance with Globus and the Science DMZGlobus
ESnet has led the way in helping national facilities—and many other institutions in the research community—configure Science DMZs and troubleshoot network issues to maximize data transfer performance. In this talk we will present a summary of approaches and tips for getting the most out of your network infrastructure using Globus Connect Server.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
2023 GEOINT Tutorial - Synthetic Data Tools for Computer Vision-Based AI - Rendered.ai
1.
2. Synthetic Data Tools for
Computer Vision-Based AI
Chris Andrews
COO, Rendered.ai
chris@rendered.ai
Dan Hedges
Lead Solution Architect, Rendered.ai
dan@rendered.ai
3. Presenters
3
• COO & Head of Product, Rendered.ai
• 25 years experience in commercial and
government geospatial-related products
and technologies
• 3D, enterprise integration, BIM-GIS,
defense-related apps & solutions
• 15 years experience with product
development at companies including Esri,
IBM, and Autodesk
• Lead Solution Engineer, Rendered.ai
• 11+ years experience building geospatial
solutions for industry verticals including
urban planning, local government, federal
government
• Subject matter expertise in remote
sensing, 3D data, and feature extraction
Chris Andrews Dan Hedges
4. Rendered.ai’s cloud-hosted platform for synthetic
data enables customers to overcome the costs
and challenges of acquiring and using real data
for training and validating computer vision ML and
AI systems and algorithms
• Established in 2019 in Bellevue, WA
• Inclusive subscription encompasses 2D & 3D content
creation, simulation design, data generation
• Rapid setup and configuration for shortest path to
synthetic data generation for multiple applications
• Available on the AWS Marketplace
Synthetic Data experts with experience in:
• Remote sensing – Satellite, Aerial
• Ground-based imagery & video
• Non-visible EM spectra
• 2D and 3D modeling and simulation
• GAN training and dataset post processing
• Dataset comparison and validation
The Platform as a Service for Synthetic Data
Partnering with
Member
5. The AI Data Problem
BIAS
COST & TIME
INNOVATION PRIVACY/SECURITY
Real data is expensive and often
costly and time consuming to
acquire and label
Rare objects and scenarios
are hard to capture
Without data it is impossible to
explore new sensors and data types
Real data can have security or high-risk
information concerns that limit usage
7. Synthetic Data solves the AI data problem
Rendered.ai is a PaaS and developer
framework for synthetic data
Synthetic Data is engineered data that
AI interprets as real data
60% of data used for AI and data analytics
projects will be synthetic, and by 2030, synthetic
data will have completely overtaken real data in AI
models.
- Gartner, September 2021
Imagine if it were possible to produce infinite
amounts of the world’s most valuable resource,
cheaply and quickly…
This is a reality today. It is called synthetic data.
- Forbes, July 2022
8. What do we mean by Synthetic Data?
Synthetic data can be created for any type of data used to train or
validate AI/ML systems, even for sensors or systems that don’t exist
CV-based synthetic data simulates bitmap sensor data capture
whether from sensors, recorded spatial patterns, or other CV input
content
Physics-based synthetic data includes creation of 2D/3D/4D output
based on ‘digital twins’ of physical sensors, the sensor platform, and
the scene in which the sensor would operate
Rendered.ai can be used to generate any kind of synthetic data
Initial focus has been on physics-based synthetic data generation for
CV workflows
• RGB imagery and video, RGB microscopy, IR imagery, X-ray, SAR,…
Source: Wikipedia
9. Today’s AI workflow relies on finding or
acquiring data
Acquire or
find data
Train algorithm Test algorithm Accept/Reject
result
Expensive, unpredictable data acquisition costs
Difficulty training algorithms on inconsistent data
Testing requires reuse of real datasets
Results are limited to what can be achieved with real datasets
10. Tomorrow’s AI workflow incorporates synthetic data
Inexpensive, unlimited data generation
100% accurate labeling, consistent data
Real datasets used for comparison and post processing
Data can be designed for edge or impossible cases and for removing bias
Create data Train algorithm
Test algorithm
Compare
datasets
11. Simulator
Dataset and
metadata
Managed Compute
Improved and explainable
outcomes
World building and
procedural gen
Asset Acquisition /
Integration
AI Model
Real-world workflow
For more information:info@rendered.ai
Post processing /
Domain adaptation
Quality assessment
Synthetic Data Engineers
Data Scientist
Platform Automation
Simulator
Synthetic
Dataset
AI Model
Hypothetical workflow
12. Synthetic data generation steps
1. Scenario characterization - Data output, variability, problem(s) addressed or tested
2. World building - Asset and scene content composition and aggregation
3. Sensor modeling & simulation - Rendering, visual effects, environmental effects
4. Annotation & mask calculation
5. Job execution & dataset compilation
6. Annotation mapping
7. Domain Adaptation post-processing
8. Dataset characterization and comparison
12
13. New AI job: Synthetic Data Engineer
If most data used to train AI will be synthetic…
…who will be engineering the data?
Design & engineer datasets to achieve specific AI outcomes
Software development-oriented
• python, data science, 3D, game engines
Domain or industry expertise
Expert in specific data types & technologies
• Sensors, Renderers, Modeling, Simulation
14. What about Generative AI?
Physics-based synthetic data
• Starts from a 3D simulation
• Can add wide variation including absurd,
unnatural, or extremely rare phenomena
• Can generate multiple ‘maps’ for depth,
instances, surfaces, normals, motion
• Can generate fully pixel-labeled content
• Can incorporate accurate physics-based
models for imagery generation
Generative AI (2023)
• Starts with large, known datasets
• Can add variation, but must be driven by
addition of additional training data
• Cannot generate extra maps with
information in the scene
• Cannot label at the pixel level
• Does not incorporate physics-based
models
Generative AI is moving fast and we see it as another tool for both
content generation and post processing or consuming other synthetic data
15. New AI job: Prompt Engineer
In the world of Generative AI, someone needs to tell the AI what to
produce!
Design & engineer inputs to Generative AI systems to achieve specific
outcomes
Narrative-oriented
• Good at defining context, describing problems
Domain or industry expertise
Expert in specific data types & technologies
• Sensors, Renderers, Modeling, Simulation
16. Common gaps when introducing customers to synthetic data
• Hyper focus on the bounds of found or acquired data only
• Most data scientists aren’t sensor experts
• Concern about ‘good data’
• Concern about one-off datasets vs. investment in data
• Belief that human perception is good enough to judge data quality
• Confusion over Generative AI vs. simulation ntechniques
… Note that the biggest hurdle is that customers rarely stop to ask what the
ideal dataset would be that would address their business problem!
17. Synthetic data generation is an empirical process
17
Identify the
problem
Describe
the (ideal)
data
Generate
data
Can I
achieve
any
training?
Refine
data
generation
Can I
improve
training?
18. Supporting GEOINT workflows with continuously evolving AI
Model digital
sensor
Aggregate &
create scene
content
Create
Channel
configuration
Publish to
Rendered.ai
Add Channel
to
Workspace
Create &
configure
Graphs
Run Jobs
Channel development
(GIS Developer, Database Engineer,
Synthetic Data Engineer)
Train and
Evaluate AI
Datasets
Graph configuration and job execution
(GIS Analyst, Computer Vision Engineers,
Data Scientists & Automated Workflows)
Change graph
configuration
Add/update sensor configuration,
Scene content, scene configuration
Annotation
Images
Masks
Statistics
GIS tools
Data Science toolkits
Embedded AI tools
20. Don’t rebuild everything for every AI application
Remote Sensing
Supply Chain
Object detection
Automotive
Economic monitoring
Medical Imaging
Security
…
Sensors
Radar Imagery
RGB Camera
Panchromatic
Infrared
High-Definition Radar
Microscopy
X-Ray/CT Scan
MRI
…
Applications Reusable modular architecture
in the cloud
• Content pipelines
• Sensor models
• Analytics toolsets
• AI integrations
Enabling access to synthetic
data as an enterprise capability
21. Channel Development | Blender
Content Code: SATRDEMO
- Dependencies installed:
- Blender and Python (versions harmonized),
OpenCV, GPU drivers, Ana, Anatools SDK
- Can Edit and Deploy Channels with SDK
- Offered as AMI or from git with
.devcontainer for VS-code
- ArcGIS integration for 2D raster
backgrounds
Custom Code
Available now
22. Case study slide: EO scenarios
Searching for cranes, and crane trucks as an economic indicator in satellite imagery
Objects are rare relative to other features in overhead imagery.
Which means very large labelling campaigns are needed to collect
examples. Original dataset only had ~100 examples of each class.
Objects are difficult to label. Inconsistent sizing of crane bounding
boxes and similarities between crane trucks and cement pumps
were two notable challenges in the real datasets.
Synthetic and real datasets
2-3x improvement in AP scores over peak performance
without Synthetic data
23. Channel Development | DIRSIGTM
Content Code: DIRSIGDEMO
- DIRSIG accessed through
python and web interface
- Can Edit and Deploy Channels
with SDK
- No RIT DIRSIG training
required! Custom Code
Available now
24. Example Applications:
Hyper-spectral Imaging,
Multi-spectral Imaging
Unique relationship with RIT allows
Rendered.ai to package DIRSIG in synthetic
data channels for customers
MSI, HSI, other radiometrically complex
imagery output
Validation possible with calibration panels,
3rd party consulting
Pixel-level geospatial accuracy
Geospatially accurate, high resolution scene
content used in cloud-based generation for
very large datasets
RGB bands from MSI, HSI images created with
DIRSIG and Rendered.ai
25. Channel Development | Omniverse
Available on request
• Preinstalled dependencies:
• USD, Python, OpenCV, GPU drivers,
Ana, Anatools SDK
• Edit and Deploy Channels to
Rendered.ai with SDK
• Offered as AMI or from git with
.devcontainer for VS-code
Custom Code
26. Example Applications:
Omniverse Replicator
channel
Use industry-leading 3D toolkit in the cloud
Configurable in a web-based SaaS experience
Starting place for users who may already
have some experience or investment in
NVIDIA tools
Familiar architecture that extends to
multiple use cases
Synthetic imagery chips generated with Omniverse Replicator running inside Rendered.ai on AWS
27. Example Application:
Synthetic Aperture
Radar
Enterprise & Developer Subscription
Customers
Experimental, cutting-edge Synthetic
Aperture Radar simulation built by
Rendered.ai
SAR output is not human readable,
making human labeling impossible
Emerging commercial SAR industry
seeking better tools for exploitation,
value creation
Applications in defense, disaster
response, Earth observation &
monitoring, insuretech
Synthetic SAR images generated using Rendered.ai
Identical object shown with several image capture scenarios
28. Example
Application: Marine
Imagery
Enterprise tier customer
Vessel detection in open ocean
scenarios for defense and
contraband interdiction
Supporting edge-based, onboard
object detection systems
Variable weather, wave, obstruction
characteristics
Variable object placement
generators
Synthetic RGB images simulating marine UAV imagery capture
29. Satellite Visible
Synthetic IR (MWIR)
Synthetic SAR
Over 1.2TB of
synthetic
images
produced with
channel coverage
growing
Security Imaging FLIR Camera
Examples of synthetic CV content
X-Ray and CT scans
Urban & natural
environments
Industrial and
residential settings
30. And after you have your imagery… compare it!
Creating datasets is a starting point
Training and Validation are next
Compare datasets to explore similarity
• Real-synthetic, synthetic-synthetic
Use tools such as UMAP, FID
Use inference to change SDG
Try again!
UMAP analysis enables data scientists
To explore similarities and differences in
The parameter embedding space of multiple
datasets
32. Internal
Past experience with cost or failure of
one-off synthetic data experiments
Unprepared for experimentation
Effort to achieve acceptable level of
realism
Complexity/difficulty with physics-
based modeling
External
• Information about emerging tools
• TCO of yet-another-IT project
• Talent shortage
• Lack of benchmarks/standards
─ Need for analytic tools
─ Need for sensitivity analysis
• Lack of industry collaboration
Typical challenges adopting synthetic data
33. Opportunity of Synthetic Data
Supplement real data
Evaluate and remove bias
Reduce expensive dataset
labeling and reacquisition
Explore scenarios
Simulate sensor models
and collection techniques
Create novel data with
zero PII or
security concerns
34. Synthetic data as a Standard
Synthetic data is rapidly moving from uncertain value to required tool. Synthetic
data has the opportunity to be used as part of regulatory and ethical frameworks
around bias reduction, demonstrable sensitivity analysis, and reducing the need
for human curation of training data.
Regulatory & compliance
• Bias reduction and testing
• Sensitivity analysis
• Efficacy demonstration
• Removing human-in-the-loop from ethical/harmful scenarios
35. Synthetic data as an enabler for innovation
As synthetic data generation capabilities improve and become more
accessible, users will have expanded opportunity to experiment,
innovate, and build AI without expensive or impossible real sensor
dataset collection.
Innovation
• Complex sensor fusion
• New & hard-to-acquire sensors
• New dataset combinations
• Digital Twins
36. Synthetic data driving sustainability
Synthetic data is 100% reliably labeled, has been shown to reduce the size
of training datasets, and potentially reduces the need for real sensor-based
data collection.
Cost and impact
• Reducing labeling costs
• Reducing collection costs
• Reducing environmental footprint of real sensor data collection
• Enabling innovation without physical material consumption/investment
37. Wrap up
37
For slides and supporting content:
https://bit.ly/GEOINT2023
Try it at:
https://rendered.ai/getstarted.html