For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/a-highly-data-efficient-deep-learning-approach-a-presentation-from-samsung/
Patrick Bangert, Vice President of AI at Samsung, presents the “Highly Data-Efficient Deep Learning Approach” tutorial at the May 2021 Embedded Vision Summit.
Many applications, such as medical imaging, lack the large amounts of data required for training popular CNNs to achieve sufficient accuracy. Often, these same applications suffer from an imbalanced class distribution problem that negatively impacts model accuracy. In this talk, Bangert proposes a highly data-efficient methodology that can achieve the same level of accuracy using significantly fewer labeled images and is insensitive to class imbalance.
The approach is based on a training pipeline with two components: a CNN trained in an unsupervised setting for image feature representation generation, and a multiclass Gaussian process classifier, trained in active learning cycles, using the image representations with labels. Bangert demonstrates his company’s approach with a COVID-19 chest X-ray classifier solution where data is scarce and highly imbalanced. He shows that the approach is insensitive to class imbalance and achieves comparable accuracy to prior approaches while using only a fraction of the training data.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/09/a-practical-guide-to-implementing-ml-on-embedded-devices-a-presentation-from-the-chamberlain-group/
Nathan Kopp, Principal Software Architect for Video Systems at the Chamberlain Group, presents the “Practical Guide to Implementing ML on Embedded Devices” tutorial at the May 2021 Embedded Vision Summit.
Deploying machine learning onto edge devices requires many choices and trade-offs. Fortunately, processor designers are adding inference-enhancing instructions and architectures to even the lowest cost MCUs, tools developers are constantly discovering optimizations that extract a little more performance out of existing hardware, and ML researchers are refactoring the math to achieve better accuracy using faster operations and fewer parameters.
In this presentation, Kopp takes a high-level look at what is involved in running a DNN model on existing edge devices, exploring some of the evolving tools and methods that are finally making this dream a reality. He also takes a quick look at a practical example of running a CNN object detector on low-compute hardware.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/pathpartner/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Jayachandra Dakala, Technical Architect at PathPartner Technology, presents the "Approaches for Vision-based Driver Monitoring" tutorial at the May 2017 Embedded Vision Summit.
Since many road accidents are caused by driver inattention, assessing driver attention is important to preventing accidents. Distraction caused by other activities and sleepiness due to fatigue are the main causes of driver inattention. Vision-based assessment of driver distraction and fatigue must estimate face pose, sleepiness, expression, etc. Estimating these aspects under real driving conditions, including day-to-night transition, drivers wearing sunglasses etc., is a challenging task.
A solution using deep learning to handle tasks from searching for a driver’s face in a given image to estimating attention would potentially be difficult to realize in an embedded system. In this talk, Dakala looks at the pros and cons of various machine learning approaches like multi-task deep networks, boosted cascades, etc. for this application, and then describes a hybrid approach that provides the required insights while being realizable in an embedded system.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/09/explainability-in-computer-vision-a-machine-learning-engineers-overview-a-presentation-from-altaml/
Navaneeth Kamballur Kottayil, Lead Machine Learning Developer at AltaML, presents the “Explainability in Computer Vision: A Machine Learning Engineer’s Overview” tutorial at the May 2021 Embedded Vision Summit.
With the increasing use of deep neural networks in computer vision applications, it has become more difficult for developers to explain how their algorithms work. This can make it difficult to establish trust and confidence among customers and other stakeholders, such as regulators. Lack of explainability also makes it more difficult for developers to improve their solutions.
In this talk, Kottayil introduces methods for enabling explainability in deep-learning-based computer vision solutions. He also illustrates some of these techniques via real-world examples, and shows how they can be used to improve customer trust in computer vision models, to debug computer vision models, to obtain additional insights about data and to detect bias in models.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-leontiev
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Anton Leontiev, Embedded Software Architect at ELVEES, JSC, presents the "Designing a Stereo IP Camera From Scratch" tutorial at the May 2017 Embedded Vision Summit.
As the number of cameras in an intelligent video surveillance system increases, server processing of the video quickly becomes a bottleneck. On the other hand, when computer vision algorithms are moved to a resource-limited camera platform, their output quality is often unsatisfactory.
The effectiveness of vision algorithms for surveillance can be greatly improved by using a depth map in addition to the regular image. Thus, using a stereo camera is a way to enable offloading of advanced algorithms from servers to IP cameras. This talk covers the main problems arising during the design of an embedded stereo IP camera, including capturing video streams from two sensors, frame synchronization between sensors, stereo calibration algorithms, and, finally, disparity map calculation.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-chiu
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Matthew Chiu, Founder of Almond AI, presents the "Designing CNN Algorithms for Real-time Applications" tutorial at the May 2017 Embedded Vision Summit.
The real-time performance of CNN-based applications can be improved several-fold by making smart decisions at each step of the design process – from the selection of the machine learning framework and libraries used to the design of the neural network algorithm to the implementation of the algorithm on the target platform. This talk delves into how to evaluate the runtime performance of a CNN from a software architecture standpoint. It then explains in detail how to build a neural network from the ground up based on the requirements of the target hardware platform.
Chiu shares his ideas on how to improve performance without sacrificing accuracy, by applying recent research on training very deep networks. He also shows examples of how network optimization can be achieved at the algorithm design level by making a more efficient use of weights before the model is compressed via more traditional methods for deployment in a real-time application.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/building-an-autonomous-detect-and-avoid-system-for-commercial-drones-a-presentation-from-iris-automation/
Alejandro Galindo, Head of Research and Development at Iris Automation, presents the “Building an Autonomous Detect-and-Avoid System for Commercial Drones” tutorial at the May 2021 Embedded Vision Summit.
Commercial and industrial drones have the potential to completely disrupt industries and create new ones. Used in applications such as infrastructure inspection, search and rescue, package delivery, and many others, they can save time, money, and lives. Most of these applications require a real-time understanding of the environment and the risks of collision.
At the same time, commercial drones are limited in the size, weight, and power they can carry, narrowing the options for sensors and computing architectures. In this presentation, Galindo dives into what it takes to build an autonomous detect-and-avoid system for commercial drones and, in particular, focuses on computer vision issues such as predictability and reduction of false positives. Why are they important and what does it take to drive them in the right direction?
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/08/getting-started-with-vision-ai-model-training-a-presentation-from-nvidia/
Ekaterina Sirazitdinova, Data Scientist at NVIDIA, presents the “Getting Started with Vision AI Model Training” tutorial at the May 2021 Embedded Vision Summit.
In many modern vision and graphics applications, deep neural networks (DNNs) enable state-of-the-art performance for tasks like image classification, object detection and segmentation, quality enhancement and even new content generation. In this talk, Sirazitdinova demystifies basic concepts behind DNN training: from the definition of a deep neural network to critical parameters controlling deep learning model training.
Sirazitdinova examines problems typically encountered when training a model and shares best practices for their mitigation. Finally, she sheds some light on commonly used software frameworks.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/08/case-study-facial-detection-and-recognition-for-always-on-applications-a-presentation-from-synopsys/
Jamie Campbell, Product Marketing Manager for Embedded Vision IP at Synopsys, presents the “Case Study: Facial Detection and Recognition for Always-On Applications” tutorial at the May 2021 Embedded Vision Summit.
Although there are many applications for low-power facial recognition in edge devices, perhaps the most challenging to design are always-on, battery-powered systems that use facial recognition for access control. Laptop, tablet and cellphone users expect hands-free and instantaneous facial recognition. This means the electronics must be always on, constantly looking to detect a face, and then ready to pull from a data set to recognize the face.
This presentation describes the challenges of moving traditional facial detection neural networks to the edge. It explores a case study of a face recognition access control application requiring continuous operation and extreme energy efficiency. Finally, it describes how the combination of Synopsys DesignWare ARC EM and EV processors provides low-power, efficient DSP and CNN acceleration for this application.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/09/a-practical-guide-to-implementing-ml-on-embedded-devices-a-presentation-from-the-chamberlain-group/
Nathan Kopp, Principal Software Architect for Video Systems at the Chamberlain Group, presents the “Practical Guide to Implementing ML on Embedded Devices” tutorial at the May 2021 Embedded Vision Summit.
Deploying machine learning onto edge devices requires many choices and trade-offs. Fortunately, processor designers are adding inference-enhancing instructions and architectures to even the lowest cost MCUs, tools developers are constantly discovering optimizations that extract a little more performance out of existing hardware, and ML researchers are refactoring the math to achieve better accuracy using faster operations and fewer parameters.
In this presentation, Kopp takes a high-level look at what is involved in running a DNN model on existing edge devices, exploring some of the evolving tools and methods that are finally making this dream a reality. He also takes a quick look at a practical example of running a CNN object detector on low-compute hardware.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/pathpartner/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Jayachandra Dakala, Technical Architect at PathPartner Technology, presents the "Approaches for Vision-based Driver Monitoring" tutorial at the May 2017 Embedded Vision Summit.
Since many road accidents are caused by driver inattention, assessing driver attention is important to preventing accidents. Distraction caused by other activities and sleepiness due to fatigue are the main causes of driver inattention. Vision-based assessment of driver distraction and fatigue must estimate face pose, sleepiness, expression, etc. Estimating these aspects under real driving conditions, including day-to-night transition, drivers wearing sunglasses etc., is a challenging task.
A solution using deep learning to handle tasks from searching for a driver’s face in a given image to estimating attention would potentially be difficult to realize in an embedded system. In this talk, Dakala looks at the pros and cons of various machine learning approaches like multi-task deep networks, boosted cascades, etc. for this application, and then describes a hybrid approach that provides the required insights while being realizable in an embedded system.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/09/explainability-in-computer-vision-a-machine-learning-engineers-overview-a-presentation-from-altaml/
Navaneeth Kamballur Kottayil, Lead Machine Learning Developer at AltaML, presents the “Explainability in Computer Vision: A Machine Learning Engineer’s Overview” tutorial at the May 2021 Embedded Vision Summit.
With the increasing use of deep neural networks in computer vision applications, it has become more difficult for developers to explain how their algorithms work. This can make it difficult to establish trust and confidence among customers and other stakeholders, such as regulators. Lack of explainability also makes it more difficult for developers to improve their solutions.
In this talk, Kottayil introduces methods for enabling explainability in deep-learning-based computer vision solutions. He also illustrates some of these techniques via real-world examples, and shows how they can be used to improve customer trust in computer vision models, to debug computer vision models, to obtain additional insights about data and to detect bias in models.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-leontiev
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Anton Leontiev, Embedded Software Architect at ELVEES, JSC, presents the "Designing a Stereo IP Camera From Scratch" tutorial at the May 2017 Embedded Vision Summit.
As the number of cameras in an intelligent video surveillance system increases, server processing of the video quickly becomes a bottleneck. On the other hand, when computer vision algorithms are moved to a resource-limited camera platform, their output quality is often unsatisfactory.
The effectiveness of vision algorithms for surveillance can be greatly improved by using a depth map in addition to the regular image. Thus, using a stereo camera is a way to enable offloading of advanced algorithms from servers to IP cameras. This talk covers the main problems arising during the design of an embedded stereo IP camera, including capturing video streams from two sensors, frame synchronization between sensors, stereo calibration algorithms, and, finally, disparity map calculation.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-chiu
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Matthew Chiu, Founder of Almond AI, presents the "Designing CNN Algorithms for Real-time Applications" tutorial at the May 2017 Embedded Vision Summit.
The real-time performance of CNN-based applications can be improved several-fold by making smart decisions at each step of the design process – from the selection of the machine learning framework and libraries used to the design of the neural network algorithm to the implementation of the algorithm on the target platform. This talk delves into how to evaluate the runtime performance of a CNN from a software architecture standpoint. It then explains in detail how to build a neural network from the ground up based on the requirements of the target hardware platform.
Chiu shares his ideas on how to improve performance without sacrificing accuracy, by applying recent research on training very deep networks. He also shows examples of how network optimization can be achieved at the algorithm design level by making a more efficient use of weights before the model is compressed via more traditional methods for deployment in a real-time application.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/building-an-autonomous-detect-and-avoid-system-for-commercial-drones-a-presentation-from-iris-automation/
Alejandro Galindo, Head of Research and Development at Iris Automation, presents the “Building an Autonomous Detect-and-Avoid System for Commercial Drones” tutorial at the May 2021 Embedded Vision Summit.
Commercial and industrial drones have the potential to completely disrupt industries and create new ones. Used in applications such as infrastructure inspection, search and rescue, package delivery, and many others, they can save time, money, and lives. Most of these applications require a real-time understanding of the environment and the risks of collision.
At the same time, commercial drones are limited in the size, weight, and power they can carry, narrowing the options for sensors and computing architectures. In this presentation, Galindo dives into what it takes to build an autonomous detect-and-avoid system for commercial drones and, in particular, focuses on computer vision issues such as predictability and reduction of false positives. Why are they important and what does it take to drive them in the right direction?
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/08/getting-started-with-vision-ai-model-training-a-presentation-from-nvidia/
Ekaterina Sirazitdinova, Data Scientist at NVIDIA, presents the “Getting Started with Vision AI Model Training” tutorial at the May 2021 Embedded Vision Summit.
In many modern vision and graphics applications, deep neural networks (DNNs) enable state-of-the-art performance for tasks like image classification, object detection and segmentation, quality enhancement and even new content generation. In this talk, Sirazitdinova demystifies basic concepts behind DNN training: from the definition of a deep neural network to critical parameters controlling deep learning model training.
Sirazitdinova examines problems typically encountered when training a model and shares best practices for their mitigation. Finally, she sheds some light on commonly used software frameworks.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/08/case-study-facial-detection-and-recognition-for-always-on-applications-a-presentation-from-synopsys/
Jamie Campbell, Product Marketing Manager for Embedded Vision IP at Synopsys, presents the “Case Study: Facial Detection and Recognition for Always-On Applications” tutorial at the May 2021 Embedded Vision Summit.
Although there are many applications for low-power facial recognition in edge devices, perhaps the most challenging to design are always-on, battery-powered systems that use facial recognition for access control. Laptop, tablet and cellphone users expect hands-free and instantaneous facial recognition. This means the electronics must be always on, constantly looking to detect a face, and then ready to pull from a data set to recognize the face.
This presentation describes the challenges of moving traditional facial detection neural networks to the edge. It explores a case study of a face recognition access control application requiring continuous operation and extreme energy efficiency. Finally, it describes how the combination of Synopsys DesignWare ARC EM and EV processors provides low-power, efficient DSP and CNN acceleration for this application.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/09/introduction-to-dnn-model-compression-techniques-a-presentation-from-xailient/
Sabina Pokhrel, Customer Success AI Engineer at Xailient, presents the “Introduction to DNN Model Compression Techniques” tutorial at the May 2021 Embedded Vision Summit.
Embedding real-time large-scale deep learning vision applications at the edge is challenging due to their huge computational, memory, and bandwidth requirements. System architects can mitigate these demands by modifying deep-neural networks to make them more energy efficient and less demanding of processing resources by applying various model compression approaches.
In this talk, Pokhrel provides an introduction to four established techniques for model compression. She discusses network pruning, quantization, knowledge distillation and low-rank factorization compression approaches.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/modern-machine-learning-from-basics-to-advanced-deep-learning-a-presentation-from-deep-netts/
Zoran Sevarac, Associate Professor at the University of Belgrade and Co-founder and CEO of Deep Netts, presents the “Modern Machine Vision from Basics to Advanced Deep Learning” tutorial at the May 2021 Embedded Vision Summit.
In this talk, Sevarac introduces the fundamentals of deep learning for image understanding. He begins by explaining the basics of convolutional neural networks (CNNs), and showing how CNNs are used to perform image classification and object detection. He provides an overview of the recent evolution of CNN topologies for object detection. He also illustrates typical use cases for CNN-based image classification and object detection, and provides a roadmap for getting started with deep learning for image understanding.
Artificial Intelligence (AI), specifically deep learning, is revolutionizing industries, products, and core capabilities by delivering dramatically enhanced experiences. However, the deep neural networks of today use too much memory, compute, and energy. Plus, to make AI truly ubiquitous, networks need to run on the end device within a tight power and thermal budget. One approach to help address these issues is quantization, which attempts to reduce the number of bits used for weight parameters and activation calculations without sacrificing model accuracy. This presentation covers: why quantization is important, existing quantization challenges, Qualcomm AI Research's existing quantization research, and how developers and researchers can take advantage of quantization on Qualcomm Snapdragon.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/applying-the-right-deep-learning-model-with-the-right-data-for-your-application-a-presentation-from-vision-elements/
Hila Blecher-Segev, Computer Vision and AI Research Associate at Vision Elements, presents the “Applying the Right Deep Learning Model with the Right Data for Your Application” tutorial at the May 2021 Embedded Vision Summit.
Deep learning has made a huge impact on a wide variety of computer vision applications. But while the capabilities of deep neural networks are impressive, understanding how to best apply them is not straightforward. In this talk, Blecher-Segev highlights key questions that must be answered when considering incorporating a deep neural network into a vision application.
What type of data will be most beneficial for the task? Should the DNN use other types of data in addition to images? How should the data be annotated? What classes should be defined? What is the minimum amount of data needed for the network to be generalized and robust? What algorithmic approach should we use for our task (classification, regression or segmentation)? What type of network should we choose (FCN, DCNN, RNN, GAN)? Blecher-Segev explains the options and trade-offs, and maps out a process for making good choices for a specific application.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2019-embedded-vision-summit-gormish
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Michael Gormish, Research Manager at Clarifai, presents the "Machine Learning- based Image Compression: Ready for Prime Time?" tutorial at the May 2019 Embedded Vision Summit.
Computer vision is undergoing dramatic changes because deep learning techniques are now able to solve complex non-linear problems. Computer vision pipelines used to consist of hand engineered stages mathematically optimized for some carefully chosen objective function. These pipelines are being replaced with machine- learned stages or end-to-end learning techniques where enough ground truth data is available.
Similarly, for decades image compression has relied on hand crafted algorithm pipelines, but recent efforts using deep learning are reporting higher image quality than that provided by conventional techniques. Is it time to replaced discrete cosine transforms with machine-learned compression techniques?
This talk examines practical aspects of deep learned image compression systems as compared with traditional approaches. Gormish considers memory, computation and other aspects, in addition to rate-distortion, to see when ML-based compression should be considered or avoided. He also discusses approaches using a combination of machine learned and traditional techniques.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/wavecomp/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-nicol
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Chris Nicol, CTO at Wave Computing, presents the "New Dataflow Architecture for Machine Learning" tutorial at the May 2017 Embedded Vision Summit.
Data scientists have made tremendous advances in the use of deep neural networks (DNNs) to enhance business models and service offerings. But training DNNs can take a week or more using traditional hardware solutions that rely on legacy architectures that are limited in performance and scalability. New innovations that can reduce training time for both image-centric and text-centric deep neural networks will lead to an explosion of new applications. Dr. Chris Nicol, Wave Computing’s Chief Technology Officer, examines the performance challenge faced by data scientists today. Nicol outlines the technical factors underlying this bottleneck for systems relying on CPUs, GPUs, FPGAs and ASICs, and introduces a new dataflow-centric approach to DNN training.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-zeller
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Sadie Zeller, Manager of Global Product Management and the Clinical Vertical Market at Microscan Systems, presents the "Another Set of Eyes: Machine Vision Automation Solutions for In Vitro Diagnostics" tutorial at the May 2017 Embedded Vision Summit.
In vitro diagnostics (IVD) are tests that can detect diseases, conditions, or infections. The use of automation, including machine vision inspection, in IVD has increased steadily, and is now a standard practice. Vision-based laboratory automation enables greater throughput efficiency and minimizes the risk of human error. But IVD is a challenging application: the healthcare industry requires systems that are, at a minimum, fail-safe, and ideally, error-proof.
Machine vision systems for IVD (and related life sciences) therefore require a robust development phase including an iterative design-validate process to ensure that the system is safe for use. This presentation addresses some of the key requirements and constraints of healthcare vision applications, and highlights approaches for application design and testing to meet tough industry demands.
For the full video of this presentation, please visit:
https://www.edge-ai-vision.com/2020/08/once-for-all-dnns-simplifying-design-of-efficient-models-for-diverse-hardware-a-presentation-from-mit/
For more information about edge AI and vision, please visit:
http://www.edge-ai-vision.com
Song Han, Associate Professor in the Department of Electrical Engineering and Computer Science at MIT, delivers the presentation “Once-for-All DNNs: Simplifying Design of Efficient Models for Diverse Hardware” at the Edge AI and Vision Alliance’s July 2020 Edge AI and Vision Innovation Forum. Han shares his group’s latest research on the automated generation of deep neural network architectures that execute efficiently across diverse hardware targets.
In this deck from the Stanford HPC Conference, Peter Dueben from the European Centre for Medium-Range Weather Forecasts (ECMWF) presents: Machine Learning for Weather Forecasts.
"I will present recent studies that use deep learning to learn the equations of motion of the atmosphere, to emulate model components of weather forecast models and to enhance usability of weather forecasts. I will than talk about the main challenges for the application of deep learning in cutting-edge weather forecasts and suggest approaches to improve usability in the future."
Peter is contributing to the development and optimization of weather and climate models for modern supercomputers. He is focusing on a better understanding of model error and model uncertainty, on the use of reduced numerical precision that is optimised for a given level of model error, on global cloud- resolving simulations with ECMWF's forecast model, and the use of machine learning, and in particular deep learning, to improve the workflow and predictions. Peter has graduated in Physics and wrote his PhD thesis at the Max Planck Institute for Meteorology in Germany. He worked as Postdoc with Tim Palmer at the University of Oxford and has taken up a position as University Research Fellow of the Royal Society at the European Centre for Medium-Range Weather Forecasts (ECMWF) in 2017.
Watch the video: https://youtu.be/ks3fkRj8Iqc
Learn more: https://www.ecmwf.int/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Accelerating algorithmic and hardware advancements for power efficient on-dev...Qualcomm Research
Artificial Intelligence (AI), specifically deep learning, is revolutionizing industries, products, and core capabilities by delivering dramatically enhanced experiences. However, the deep neural networks of today are growing quickly in size and use too much memory, compute, and energy. Plus, to make AI truly ubiquitous, it needs to run on the end device within a tight power and thermal budget. One approach to address these issues is Bayesian deep learning. This presentation covers:
• Why AI algorithms and hardware need to be energy efficient
• How Bayesian deep learning is making neural networks more power efficient through model compression and quantization
• How we are doing fundamental research on AI algorithms and hardware to maximize power efficiency
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/an-introduction-to-single-photon-avalanche-diodes-a-new-type-of-imager-for-computer-vision-a-presentation-from-the-university-of-wisconsin-madison/
Sebastian Bauer, Postdoctoral Student at the University of Wisconsin – Madison, presents the “Introduction to Single-Photon Avalanche Diodes—A New Type of Imager for Computer Vision” tutorial at the May 2021 Embedded Vision Summit.
The single-photon avalanche diode (SPAD) is an emerging image sensing technology with unique capabilities relevant to computer vision applications. Originally designed for imaging in low-light conditions, the ultra-high time resolution of SPADs also helps to achieve extremely high dynamic range, motion blur-free images and even seeing around corners. The use of SPADs in recent iPhone models has spurred increased interest in the use of SPADs in commercial products.
In this talk, Bauer introduces SPAD-based imagers, explains how they work, presents their fundamental capabilities, and identifies their key strengths and weaknesses relative to conventional image sensors. He also shows how they can be used in a variety of applications.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-bordoloi
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Unmesh Bordoloi, Senior Researcher at General Motors, presents the "Collaboratively Benchmarking and Optimizing Deep Learning Implementations" tutorial at the May 2017 Embedded Vision Summit.
For car manufacturers and other OEMs, selecting the right processors to run deep learning inference for embedded vision applications is a critical but daunting task. One challenge is the vast number of options in terms of neural network models, frameworks (such as Caffe, TensorFlow, Torch), and libraries such as CUDA and OpenCL. Another challenge is the large number of network parameters that can affect the computation requirements, such the choice of training data sets, precision, and batch size. These challenges also complicate efforts to optimize implementations of deep learning algorithms for deployment.
In this talk, Bordoloi presents a methodology and open-source software framework for collaborative and reproducible benchmarking and optimization of convolutional neural networks. General Motors' software framework, CK-Caffe, is based on the Collective Knowledge framework and the Caffe framework. GM invites the community to collaboratively evaluate, design and optimize convolutional neural networks to meet the performance, accuracy and cost requirements of a variety of applications – from sensors to self-driving cars.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/altera/embedded-vision-training/videos/pages/may-2016-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Bill Jenkins, Senior Product Specialist for High Level Design Tools at Intel, presents the "Accelerating Deep Learning Using Altera FPGAs" tutorial at the May 2016 Embedded Vision Summit.
While large strides have recently been made in the development of high-performance systems for neural networks based on multi-core technology, significant challenges in power, cost and, performance scaling remain. Field-programmable gate arrays (FPGAs) are a natural choice for implementing neural networks because they can combine computing, logic, and memory resources in a single device. Intel's Programmable Solutions Group has developed a scalable convolutional neural network reference design for deep learning systems using the OpenCL programming language built with our SDK for OpenCL. The design performance is being benchmarked using several popular CNN benchmarks: CIFAR-10, ImageNet and KITTI.
Building the CNN with OpenCL kernels allows true scaling of the design from smaller to larger devices and from one device generation to the next. New designs can be sized using different numbers of kernels at each layer. Performance scaling from one generation to the next also benefits from architectural advancements, such as floating-point engines and frequency scaling. Thus, you achieve greater than linear performance and performance per watt scaling with each new series of devices.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/09/introduction-to-dnn-model-compression-techniques-a-presentation-from-xailient/
Sabina Pokhrel, Customer Success AI Engineer at Xailient, presents the “Introduction to DNN Model Compression Techniques” tutorial at the May 2021 Embedded Vision Summit.
Embedding real-time large-scale deep learning vision applications at the edge is challenging due to their huge computational, memory, and bandwidth requirements. System architects can mitigate these demands by modifying deep-neural networks to make them more energy efficient and less demanding of processing resources by applying various model compression approaches.
In this talk, Pokhrel provides an introduction to four established techniques for model compression. She discusses network pruning, quantization, knowledge distillation and low-rank factorization compression approaches.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/modern-machine-learning-from-basics-to-advanced-deep-learning-a-presentation-from-deep-netts/
Zoran Sevarac, Associate Professor at the University of Belgrade and Co-founder and CEO of Deep Netts, presents the “Modern Machine Vision from Basics to Advanced Deep Learning” tutorial at the May 2021 Embedded Vision Summit.
In this talk, Sevarac introduces the fundamentals of deep learning for image understanding. He begins by explaining the basics of convolutional neural networks (CNNs), and showing how CNNs are used to perform image classification and object detection. He provides an overview of the recent evolution of CNN topologies for object detection. He also illustrates typical use cases for CNN-based image classification and object detection, and provides a roadmap for getting started with deep learning for image understanding.
Artificial Intelligence (AI), specifically deep learning, is revolutionizing industries, products, and core capabilities by delivering dramatically enhanced experiences. However, the deep neural networks of today use too much memory, compute, and energy. Plus, to make AI truly ubiquitous, networks need to run on the end device within a tight power and thermal budget. One approach to help address these issues is quantization, which attempts to reduce the number of bits used for weight parameters and activation calculations without sacrificing model accuracy. This presentation covers: why quantization is important, existing quantization challenges, Qualcomm AI Research's existing quantization research, and how developers and researchers can take advantage of quantization on Qualcomm Snapdragon.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/applying-the-right-deep-learning-model-with-the-right-data-for-your-application-a-presentation-from-vision-elements/
Hila Blecher-Segev, Computer Vision and AI Research Associate at Vision Elements, presents the “Applying the Right Deep Learning Model with the Right Data for Your Application” tutorial at the May 2021 Embedded Vision Summit.
Deep learning has made a huge impact on a wide variety of computer vision applications. But while the capabilities of deep neural networks are impressive, understanding how to best apply them is not straightforward. In this talk, Blecher-Segev highlights key questions that must be answered when considering incorporating a deep neural network into a vision application.
What type of data will be most beneficial for the task? Should the DNN use other types of data in addition to images? How should the data be annotated? What classes should be defined? What is the minimum amount of data needed for the network to be generalized and robust? What algorithmic approach should we use for our task (classification, regression or segmentation)? What type of network should we choose (FCN, DCNN, RNN, GAN)? Blecher-Segev explains the options and trade-offs, and maps out a process for making good choices for a specific application.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2019-embedded-vision-summit-gormish
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Michael Gormish, Research Manager at Clarifai, presents the "Machine Learning- based Image Compression: Ready for Prime Time?" tutorial at the May 2019 Embedded Vision Summit.
Computer vision is undergoing dramatic changes because deep learning techniques are now able to solve complex non-linear problems. Computer vision pipelines used to consist of hand engineered stages mathematically optimized for some carefully chosen objective function. These pipelines are being replaced with machine- learned stages or end-to-end learning techniques where enough ground truth data is available.
Similarly, for decades image compression has relied on hand crafted algorithm pipelines, but recent efforts using deep learning are reporting higher image quality than that provided by conventional techniques. Is it time to replaced discrete cosine transforms with machine-learned compression techniques?
This talk examines practical aspects of deep learned image compression systems as compared with traditional approaches. Gormish considers memory, computation and other aspects, in addition to rate-distortion, to see when ML-based compression should be considered or avoided. He also discusses approaches using a combination of machine learned and traditional techniques.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/wavecomp/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-nicol
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Chris Nicol, CTO at Wave Computing, presents the "New Dataflow Architecture for Machine Learning" tutorial at the May 2017 Embedded Vision Summit.
Data scientists have made tremendous advances in the use of deep neural networks (DNNs) to enhance business models and service offerings. But training DNNs can take a week or more using traditional hardware solutions that rely on legacy architectures that are limited in performance and scalability. New innovations that can reduce training time for both image-centric and text-centric deep neural networks will lead to an explosion of new applications. Dr. Chris Nicol, Wave Computing’s Chief Technology Officer, examines the performance challenge faced by data scientists today. Nicol outlines the technical factors underlying this bottleneck for systems relying on CPUs, GPUs, FPGAs and ASICs, and introduces a new dataflow-centric approach to DNN training.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-zeller
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Sadie Zeller, Manager of Global Product Management and the Clinical Vertical Market at Microscan Systems, presents the "Another Set of Eyes: Machine Vision Automation Solutions for In Vitro Diagnostics" tutorial at the May 2017 Embedded Vision Summit.
In vitro diagnostics (IVD) are tests that can detect diseases, conditions, or infections. The use of automation, including machine vision inspection, in IVD has increased steadily, and is now a standard practice. Vision-based laboratory automation enables greater throughput efficiency and minimizes the risk of human error. But IVD is a challenging application: the healthcare industry requires systems that are, at a minimum, fail-safe, and ideally, error-proof.
Machine vision systems for IVD (and related life sciences) therefore require a robust development phase including an iterative design-validate process to ensure that the system is safe for use. This presentation addresses some of the key requirements and constraints of healthcare vision applications, and highlights approaches for application design and testing to meet tough industry demands.
For the full video of this presentation, please visit:
https://www.edge-ai-vision.com/2020/08/once-for-all-dnns-simplifying-design-of-efficient-models-for-diverse-hardware-a-presentation-from-mit/
For more information about edge AI and vision, please visit:
http://www.edge-ai-vision.com
Song Han, Associate Professor in the Department of Electrical Engineering and Computer Science at MIT, delivers the presentation “Once-for-All DNNs: Simplifying Design of Efficient Models for Diverse Hardware” at the Edge AI and Vision Alliance’s July 2020 Edge AI and Vision Innovation Forum. Han shares his group’s latest research on the automated generation of deep neural network architectures that execute efficiently across diverse hardware targets.
In this deck from the Stanford HPC Conference, Peter Dueben from the European Centre for Medium-Range Weather Forecasts (ECMWF) presents: Machine Learning for Weather Forecasts.
"I will present recent studies that use deep learning to learn the equations of motion of the atmosphere, to emulate model components of weather forecast models and to enhance usability of weather forecasts. I will than talk about the main challenges for the application of deep learning in cutting-edge weather forecasts and suggest approaches to improve usability in the future."
Peter is contributing to the development and optimization of weather and climate models for modern supercomputers. He is focusing on a better understanding of model error and model uncertainty, on the use of reduced numerical precision that is optimised for a given level of model error, on global cloud- resolving simulations with ECMWF's forecast model, and the use of machine learning, and in particular deep learning, to improve the workflow and predictions. Peter has graduated in Physics and wrote his PhD thesis at the Max Planck Institute for Meteorology in Germany. He worked as Postdoc with Tim Palmer at the University of Oxford and has taken up a position as University Research Fellow of the Royal Society at the European Centre for Medium-Range Weather Forecasts (ECMWF) in 2017.
Watch the video: https://youtu.be/ks3fkRj8Iqc
Learn more: https://www.ecmwf.int/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Accelerating algorithmic and hardware advancements for power efficient on-dev...Qualcomm Research
Artificial Intelligence (AI), specifically deep learning, is revolutionizing industries, products, and core capabilities by delivering dramatically enhanced experiences. However, the deep neural networks of today are growing quickly in size and use too much memory, compute, and energy. Plus, to make AI truly ubiquitous, it needs to run on the end device within a tight power and thermal budget. One approach to address these issues is Bayesian deep learning. This presentation covers:
• Why AI algorithms and hardware need to be energy efficient
• How Bayesian deep learning is making neural networks more power efficient through model compression and quantization
• How we are doing fundamental research on AI algorithms and hardware to maximize power efficiency
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/an-introduction-to-single-photon-avalanche-diodes-a-new-type-of-imager-for-computer-vision-a-presentation-from-the-university-of-wisconsin-madison/
Sebastian Bauer, Postdoctoral Student at the University of Wisconsin – Madison, presents the “Introduction to Single-Photon Avalanche Diodes—A New Type of Imager for Computer Vision” tutorial at the May 2021 Embedded Vision Summit.
The single-photon avalanche diode (SPAD) is an emerging image sensing technology with unique capabilities relevant to computer vision applications. Originally designed for imaging in low-light conditions, the ultra-high time resolution of SPADs also helps to achieve extremely high dynamic range, motion blur-free images and even seeing around corners. The use of SPADs in recent iPhone models has spurred increased interest in the use of SPADs in commercial products.
In this talk, Bauer introduces SPAD-based imagers, explains how they work, presents their fundamental capabilities, and identifies their key strengths and weaknesses relative to conventional image sensors. He also shows how they can be used in a variety of applications.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-bordoloi
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Unmesh Bordoloi, Senior Researcher at General Motors, presents the "Collaboratively Benchmarking and Optimizing Deep Learning Implementations" tutorial at the May 2017 Embedded Vision Summit.
For car manufacturers and other OEMs, selecting the right processors to run deep learning inference for embedded vision applications is a critical but daunting task. One challenge is the vast number of options in terms of neural network models, frameworks (such as Caffe, TensorFlow, Torch), and libraries such as CUDA and OpenCL. Another challenge is the large number of network parameters that can affect the computation requirements, such the choice of training data sets, precision, and batch size. These challenges also complicate efforts to optimize implementations of deep learning algorithms for deployment.
In this talk, Bordoloi presents a methodology and open-source software framework for collaborative and reproducible benchmarking and optimization of convolutional neural networks. General Motors' software framework, CK-Caffe, is based on the Collective Knowledge framework and the Caffe framework. GM invites the community to collaboratively evaluate, design and optimize convolutional neural networks to meet the performance, accuracy and cost requirements of a variety of applications – from sensors to self-driving cars.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/altera/embedded-vision-training/videos/pages/may-2016-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Bill Jenkins, Senior Product Specialist for High Level Design Tools at Intel, presents the "Accelerating Deep Learning Using Altera FPGAs" tutorial at the May 2016 Embedded Vision Summit.
While large strides have recently been made in the development of high-performance systems for neural networks based on multi-core technology, significant challenges in power, cost and, performance scaling remain. Field-programmable gate arrays (FPGAs) are a natural choice for implementing neural networks because they can combine computing, logic, and memory resources in a single device. Intel's Programmable Solutions Group has developed a scalable convolutional neural network reference design for deep learning systems using the OpenCL programming language built with our SDK for OpenCL. The design performance is being benchmarked using several popular CNN benchmarks: CIFAR-10, ImageNet and KITTI.
Building the CNN with OpenCL kernels allows true scaling of the design from smaller to larger devices and from one device generation to the next. New designs can be sized using different numbers of kernels at each layer. Performance scaling from one generation to the next also benefits from architectural advancements, such as floating-point engines and frequency scaling. Thus, you achieve greater than linear performance and performance per watt scaling with each new series of devices.
Quadcopter for Monitoring and DetectionTejasDalvi15
Drone is designed to inspect whether the rule of wearing the facemask is practiced strictly or not in crowded place and to predict ripening stage of banana.
Micrography Image Analysis with Cell Broadband Engine ProjectAndrea Corona
Research project and development of a computer-aided diagnosis system prototype that provided anatomic pathologists with a valid support in the diagnosis of the carcinoma of the cervix.
Based on advanced machine learning and high performance computing algorithms on large images taken in the Papanicolaou (Pap) test.
The Indian Dental Academy is the Leader in continuing dental education , training dentists in all aspects of dentistry and
offering a wide range of dental certified courses in different formats.for more details please visit
www.indiandentalacademy.com
Use of Machine Intelligence in manufacturing industry poses a special challenge due to a wide range of use cases, inherent complexity in data collection, availability of information and disconnect between information islands in different manufacturing steps.
Within our talk we present several machine intelligence projects we did in the manufacturing industry, which helped our customers in product quality improvement, reduction of cost and better asset management. We will talk about the used methodologies, the results achieved and the lessons learned from these projects. We will specifically focus on the importance of process and business knowledge for successful implementation of any industrial project.
Predicting Medical Test Results using Driverless AISri Ambati
This talk was given at H2O World 2018 NYC and can be viewed here: https://youtu.be/n9g9GxIJoT4
Description:
The goal of the research was to develop an approach to predict individual medical test results based on longitudinal medical and pharma claims data without direct lab measures using data-driven techniques. Such discoveries may result in improved treatment strategies. In the presentation we demonstrate how Driverless AI was used both for estimating highly accurate model and results explanations.
Speaker's Bio:
Alexander is the Data Science leader at poder.IO. He is responsible for data flow architecture and insight mining, all powered by machine learning. Before joining poder.IO, Alexander made an academic career at Belarusian State University and Minsk Innovation University where he was Head of the Informatics and Mathematics Department.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/opencv-for-high-performance-low-power-vision-applications-on-snapdragon-a-presentation-from-qualcomm/
Xin Zhong, Computer Vision Product Manager at Qualcomm Technologies, presents the “OpenCV for High-performance, Low-power Vision Applications on Snapdragon” tutorial at the May 2024 Embedded Vision Summit.
For decades, the OpenCV software library has been popular for developing computer vision applications. However, developers have found it challenging to create efficient implementations of their OpenCV applications on processors optimized for edge applications, like the Qualcomm Snapdragon family. As part of its comprehensive support for computer vision application developers, Qualcomm provides a variety of tools to enable developers to take full advantage of the heterogeneous computing resources in the Snapdragon processors.
In this talk, Zhong introduces a new element of Qualcomm’s computer vision tools suite: a version of OpenCV optimized for Snapdragon platforms, which allows developers to leverage and port their existing OpenCV-based applications seamlessly to Snapdragon platforms. Supporting OpenCV v4.x and later releases, this implementation contains unique Qualcomm-specific accelerations of OpenCV and OpenCV extension APIs. Zhong explains how this library enables developers to leverage existing OpenCV code to achieve superior performance and power savings on Snapdragon platforms.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/deploying-large-models-on-the-edge-success-stories-and-challenges-a-presentation-from-qualcomm/
Vinesh Sukumar, Senior Director of Product Management at Qualcomm Technologies, presents the “Deploying Large Models on the Edge: Success Stories and Challenges” tutorial at the May 2024 Embedded Vision Summit.
In this talk, Dr. Sukumar explains and demonstrates how Qualcomm has been successful in deploying large generative AI and multimodal models on the edge for a variety of use cases in consumer and enterprise markets. He examines key challenges that must be overcome before large models at the edge can reach their full commercial potential. He also highlights how Qualcomm is addressing these challenges through upgraded processor hardware, improved developer tools and a comprehensive library of fully optimized AI models in the Qualcomm AI Hub.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/scaling-vision-based-edge-ai-solutions-from-prototype-to-global-deployment-a-presentation-from-network-optix/
Maurits Kaptein, Chief Data Scientist at Network Optix and Professor at the University of Eindhoven, presents the “Scaling Vision-based Edge AI Solutions: From Prototype to Global Deployment” tutorial at the May 2024 Embedded Vision Summit.
The Embedded Vision Summit brings together innovators in silicon, devices, software and applications and empowers them to bring computer vision and perceptual AI into reliable and scalable products. However, integrating recent hardware, software and algorithm innovations into prime-time-ready products is quite challenging. Scaling from a proof of concept—for example, a novel neural network architecture performing a valuable task efficiently on a new piece of silicon—to an AI vision system installed in hundreds of sites requires surmounting myriad hurdles.
First, building on Network Optix’s 14 years of experience, Professor Kaptein details how to overcome the networking, fleet management, visualization and monetization challenges that come with scaling a global vision solution. Second, Kaptein discusses the complexities of making vision AI solutions device-agnostic and remotely manageable, proposing an open standard for AI model deployment to edge devices. The proposed standard aims to simplify market entry for silicon manufacturers and enhance scalability for solution developers. Kaptein outlines the standard’s core components and invites collaborative contributions to drive market expansion.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/whats-next-in-on-device-generative-ai-a-presentation-from-qualcomm/
Jilei Hou, Vice President of Engineering and Head of AI Research at Qualcomm Technologies, presents the “What’s Next in On-device Generative AI” tutorial at the May 2024 Embedded Vision Summit.
The generative AI era has begun! Large multimodal models are bringing the power of language understanding to machine perception, and transformer models are expanding to allow machines to understand using multiple types of sensors. This new wave of approaches is poised to revolutionize user experiences, disrupt industries and enable powerful new capabilities. For generative AI to reach its full potential, however, we must deploy it on edge devices, providing improved latency, pervasive interaction and enhanced privacy.
In this talk, Hou shares Qualcomm’s vision of the compelling opportunities enabled by efficient generative AI at the edge. He also identifies the key challenges that the industry must overcome to realize the massive potential of these technologies. And he highlights research and product development work that Qualcomm is doing to lead the way via an end-to-end system approach—including techniques for efficient on-device execution of LLMs, LVMs and LMMs, methods for orchestration of large models at the edge and approaches for adaptation and personalization.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/11/learning-compact-dnn-models-for-embedded-vision-a-presentation-from-the-university-of-maryland-at-college-park/
Shuvra Bhattacharyya, Professor at the University of Maryland at College Park, presents the “Learning Compact DNN Models for Embedded Vision” tutorial at the May 2023 Embedded Vision Summit.
In this talk, Bhattacharyya explores methods to transform large deep neural network (DNN) models into effective compact models. The transformation process that he focuses on—from large to compact DNN form—is referred to as pruning. Pruning involves the removal of neurons or parameters from a neural network. When performed strategically, pruning can lead to significant reductions in computational complexity without significant degradation in accuracy. It is sometimes even possible to increase accuracy through pruning.
Pruning provides a general approach for facilitating real-time inference in resource-constrained embedded computer vision systems. Bhattacharyya provides an overview of important aspects to consider when applying or developing a DNN pruning method and presents details on a recently introduced pruning method called NeuroGRS. NeuroGRS considers structures and trained weights jointly throughout the pruning process and can result in significantly more compact models compared to other pruning methods.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/11/introduction-to-computer-vision-with-cnns-a-presentation-from-mohammad-haghighat/
Independent consultant Mohammad Haghighat presents the “Introduction to Computer Vision with Convolutional Neural Networks” tutorial at the May 2023 Embedded Vision Summit.
This presentation covers the basics of computer vision using convolutional neural networks. Haghighat begins by introducing some important conventional computer vision techniques and then transition to explaining the basics of machine learning and convolutional neural networks (CNNs) and showing how CNNs are used in visual perception.
Haghighat illustrates the building blocks and computational elements of neural networks through examples. This session provides an overview of how modern computer vision algorithms are designed, trained and used in real-world applications.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/11/selecting-tools-for-developing-monitoring-and-maintaining-ml-models-a-presentation-from-yummly/
Parshad Patel, Data Scientist at Yummly, presents the “Selecting Tools for Developing, Monitoring and Maintaining ML Models” tutorial at the May 2023 Embedded Vision Summit.
With the boom in tools for developing, monitoring and maintaining ML models, data science teams have many options to choose from. Proprietary tools provided by cloud service providers are enticing, but teams may fear being locked in—and may worry that these tools are too costly or missing important features when compared with alternatives from specialized providers.
Fortunately, most proprietary, fee-based tools have an open-source component that can be integrated into a home-grown solution at low cost. This can be a good starting point, enabling teams to get started with modern tools without making big investments and leaving the door open to evolve tool selection over time. In this talk, Patel presents a step-by-step process for creating an MLOps tool set that enables you to deliver maximum value as a data scientist. He shares how Yummly built pipelines for model development and put them into production using open-source projects.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/11/building-accelerated-gstreamer-applications-for-video-and-audio-ai-a-presentation-from-wave-spectrum/
Abdo Babukr, Accelerated Computing Consultant at Wave Spectrum, presents the “Building Accelerated GStreamer Applications for Video and Audio AI,” tutorial at the May 2023 Embedded Vision Summit.
GStreamer is a popular open-source framework for creating streaming media applications. Developers often use GStreamer to streamline the development of computer vision and audio perception applications. Since perceptual algorithms are often quite demanding in terms of processing performance, in many cases developers need to find ways to accelerate key GStreamer building blocks, taking advantage of specialized features of their target processor or co-processor.
In this talk, Babukr introduces GStreamer and shows how to use it to build computer vision and audio perception applications. He also shows how to create efficient, high-performance GStreamer applications that utilize specialized hardware features.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/11/understanding-selecting-and-optimizing-object-detectors-for-edge-applications-a-presentation-from-walmart-global-tech/
Md Nasir Uddin Laskar, Staff Machine Learning Engineer at Walmart Global Tech, presents the “Understanding, Selecting and Optimizing Object Detectors for Edge Applications” tutorial at the May 2023 Embedded Vision Summit.
Object detectors count objects in a scene and determine their precise locations, while also labeling them. Object detection plays a crucial role in many vision applications, from autonomous driving to smart appliances. In many of these applications, it’s necessary or desirable to implement object detection at the edge.
In this presentation, Laskar explores the evolution of object detection algorithms, from traditional approaches to deep learning-based methods and transformer-based architectures. He delves into widely used approaches for object detection, such as two-stage R-CNNs and one-stage YOLO algorithms, and examines their strengths and weaknesses. And he provides guidance on how to evaluate and select an object detector for an edge application.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/introduction-to-modern-lidar-for-machine-perception-a-presentation-from-the-university-of-ottawa/
Robert Laganière, Professor at the University of Ottawa and CEO of Sensor Cortek, presents the “Introduction to Modern LiDAR for Machine Perception” tutorial at the May 2023 Embedded Vision Summit.
In this presentation, Laganière provides an introduction to light detection and ranging (LiDAR) technology. He explains how LiDAR sensors work and their main advantages and disadvantages. He also introduces different approaches to LiDAR, including scanning and flash LiDAR.
Laganière explores the types of data produced by LiDAR sensors and explains how this data can be processed using deep neural networks. He also examines the synergy between LiDAR and cameras, and the concept of pseudo-LiDAR for detection.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/11/vision-language-representations-for-robotics-a-presentation-from-the-university-of-pennsylvania/
Dinesh Jayaraman, Assistant Professor at the University of Pennsylvania, presents the “Vision-language Representations for Robotics” tutorial at the May 2023 Embedded Vision Summit.
In what format can an AI system best present what it “sees” in a visual scene to help robots accomplish tasks? This question has been a long-standing challenge for computer scientists and robotics engineers. In this presentation, Jayaraman provides insights into cutting-edge techniques being used to help robots better understand their surroundings, learn new skills with minimal guidance and become more capable of performing complex tasks.
Jayaraman discusses recent advances in unsupervised representation learning and explains how these approaches can be used to build visual representations that are appropriate for a controller that decides how the robot should act. In particular, he presents insights from his research group’s recent work on how to represent the constituent objects and entities in a visual scene, and how to combine vision and language in a way that permits effectively translating language-based task descriptions into images depicting the robot’s goals.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/adas-and-av-sensors-whats-winning-and-why-a-presentation-from-techinsights/
Ian Riches, Vice President of the Global Automotive Practice at TechInsights, presents the “ADAS and AV Sensors: What’s Winning and Why?” tutorial at the May 2023 Embedded Vision Summit.
It’s clear that the number of sensors per vehicle—and the sophistication of these sensors—is growing rapidly, largely thanks to increased adoption of advanced safety and driver assistance features. In this presentation, Riches explores likely future demand for automotive radars, cameras and LiDARs.
Riches examines which vehicle features will drive demand out to 2030, how vehicle architecture change is impacting the market and what sorts of compute platforms these sensors will be connected to. Finally, he shares his firm’s vision of what the landscape could look like far beyond 2030, considering scenarios out to 2050 for automated driving and the resulting sensor demand.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/computer-vision-in-sports-scalable-solutions-for-downmarkets-a-presentation-from-sportlogiq/
Mehrsan Javan, Co-founder and CTO of Sportlogiq, presents the “Computer Vision in Sports: Scalable Solutions for Downmarket Leagues” tutorial at the May 2023 Embedded Vision Summit.
Sports analytics is about observing, understanding and describing the game in an intelligent manner. In practice, this requires a fully automated, robust end-to-end pipeline, spanning from visual input, to player and group activities, to player and team evaluation to planning. Despite major advancements in computer vision and machine learning, today sports analytics solutions are limited to top leagues and are not widely available for downmarket leagues and youth sports.
In this talk, Javan explains how his company has developed scalable and robust computer vision solutions to democratize sport analytics and offer pro-league-level insights to leagues with modest resources, including youth leagues. He highlights key challenges—such as the requirement for low-cost, low-latency processing and the need for robustness despite variations in venues. He discusses the approaches Sportlogiq tried and how it ultimately overcame these challenges, including the use of transformers and fusion of multiple type of data streams to maximize accuracy.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/detecting-data-drift-in-image-classification-neural-networks-a-presentation-from-southern-illinois-university/
Spyros Tragoudas, Professor and School Director at Southern Illinois University Carbondale, presents the “Detecting Data Drift in Image Classification Neural Networks” tutorial at the May 2023 Embedded Vision Summit.
An unforeseen change in the input data is called “drift,” and may impact the accuracy of machine learning models. In this talk, Tragoudas presents a novel scheme for diagnosing data drift in the input streams of image classification neural networks. His proposed method for drift detection and quantification uses a threshold dictionary for the prediction probabilities of each class in the neural network model.
The method is applicable to any drift type in images such as noise and weather effects, among others. Tragoudas shares experimental results on various data sets, drift types and neural network models to show that his proposed method estimates the drift magnitude with high accuracy, especially when the level of drift significantly impacts the model’s performance.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/deep-neural-network-training-diagnosing-problems-and-implementing-solutions-a-presentation-from-sensor-cortek/
Fahed Hassanat, Chief Operating Officer and Head of Engineering at Sensor Cortek, presents the “Deep Neural Network Training: Diagnosing Problems and Implementing Solutions” tutorial at the May 2023 Embedded Vision Summit.
In this presentation, Hassanat delves into some of the most common problems that arise when training deep neural networks. He provides a brief overview of essential training metrics, including accuracy, precision, false positives, false negatives and F1 score.
Hassanat then explores training challenges that arise from problems with hyperparameters, inappropriately sized models, inadequate models, poor-quality datasets, imbalances within training datasets and mismatches between training and testing datasets. To help detect and diagnose training problems, he also covers techniques such as understanding performance curves, recognizing overfitting and underfitting, analyzing confusion matrices and identifying class interaction issues.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/ai-start-ups-the-perils-of-fishing-for-whales-war-stories-from-the-entrepreneurial-front-lines-a-presentation-from-seechange-technologies/
Tim Hartley, Vice President of Product for SeeChange Technologies, presents the “AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepreneurial Front Lines)” tutorial at the May 2023 Embedded Vision Summit.
You have a killer idea that will change the world. You’ve thought through product-market fit and differentiation. You have seed funding and a world-beating team. Best of all, you’ve caught the attention of major players in your industry. You’ve reached peak “start-up”—that point of limitless possibility—when you go to bed with the same level of energy and enthusiasm you had when you woke. And then the first proof of concept starts…
In this talk, Hartley lays out some of the pitfalls that await those building the next big thing. Using real examples, he shares some of the dos and don’ts, particularly when dealing with that big potential first customer. Hartley discusses the importance of end-to-end design, ensuring your product solves real-world problems. He explores how far the big companies will tell you to jump—and then jump again—for free. And, most importantly, how to build long-term partnerships with major corporations without relying on over-promising sales pitches.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/a-computer-vision-system-for-autonomous-satellite-maneuvering-a-presentation-from-scout-space/
Andrew Harris, Spacecraft Systems Engineer at SCOUT Space, presents the “Developing a Computer Vision System for Autonomous Satellite Maneuvering” tutorial at the May 2023 Embedded Vision Summit.
Computer vision systems for mobile autonomous machines experience a wide variety of real-world conditions and inputs that can be challenging to capture accurately in training datasets. Few autonomous systems experience more challenging conditions than those in orbit. In this talk, Harris describes how SCOUT Space has designed and trained satellite vision systems using dynamic and physically informed synthetic image datasets.
Harris describes how his company generates synthetic data for this challenging environment and how it leverages new real-world data to improve our datasets. In particular, he explains how these synthetic datasets account for and can replicate real sources of noise and error in the orbital environment, and how his company supplements them with in-space data from the first SCOUT-Vision system, which has been in orbit since 2021.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/bias-in-computer-vision-its-bigger-than-facial-recognition-a-presentation-from-santa-clara-university/
Susan Kennedy, Assistant Professor of Philosophy at Santa Clara University, presents the “Bias in Computer Vision—It’s Bigger Than Facial Recognition!” tutorial at the May 2023 Embedded Vision Summit.
As AI is increasingly integrated into various industries, concerns about its potential to reproduce or exacerbate bias have become widespread. While the use of AI holds the promise of reducing bias, it can also have unintended consequences, particularly in high-stakes computer vision applications such as facial recognition. However, even seemingly low-stakes computer vision applications such as identifying potholes and damaged roads can also present ethical challenges related to bias.
This talk explores how bias in computer vision often poses an ethical challenge, regardless of the stakes involved. Kennedy discusses the limitations of technical solutions aimed at mitigating bias, and why “bias-free” AI may not be achievable. Instead, she focuses on the importance of adopting a “bias-aware” approach to responsible AI design and explores strategies that can be employed to achieve this.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/sensor-fusion-techniques-for-accurate-perception-of-objects-in-the-environment-a-presentation-from-sanborn-map-company/
Baharak Soltanian, Vice President of Research and Development for the Sanborn Map Company, presents the “Sensor Fusion Techniques for Accurate Perception of Objects in the Environment” tutorial at the May 2023 Embedded Vision Summit.
Increasingly, perceptual AI is being used to enable devices and systems to obtain accurate estimates of object locations, speeds and trajectories. In demanding applications, this is often best done using a heterogeneous combination of sensors (e.g., vision, radar, LiDAR). In this talk, Soltanian introduces techniques for combining data from multiple sensors to obtain accurate information about objects in the environment.
Soltanian briefly introduces the roles played by Kalman filters, particle filters, Bayesian networks and neural networks in this type of fusion. She then examines alternative fusion architectures, such as centralized and decentralized approaches, to better understand the trade-offs associated with different approaches to sensor fusion as used to enhance the ability of machines to understand their environment.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/10/updating-the-edge-ml-development-process-a-presentation-from-samsara/
Jim Steele, Vice President of Embedded Software at Samsara, presents the “Updating the Edge ML Development Process” tutorial at the May 2023 Embedded Vision Summit.
Samsara (NYSE:IOT) is focused on digitizing the world of operations. The company helps customers across many industries—including food and beverage, utilities and energy, field services and government—get information about their physical operations into the cloud, so they can operate more safely, efficiently and sustainably. Samsara’s sensors collect billions of data points per day and on-device processing is instrumental to its success. The company is constantly developing, improving and deploying ML models at the edge.
Samsara has found that the traditional development process—where ML scientists create models and hand them off to firmware engineers for embedded implementation—is slow and often produces difficult-to-resolve differences between the original model and the embedded implementation. In this talk, Steele presents an alternative development process that his company has adopted with good results. In this process, firmware engineers develop a general framework that ML scientists use to design, develop and deploy their models. This enables quick iterations and fewer confounding bugs.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.