How do you find the best solution when faced with many choices? Combinatorial optimization is a field of mathematics that seeks to find the most optimal solutions for complex problems involving multiple variables. There are numerous business verticals that can benefit from combinatorial optimization, whether transport, supply chain, or the mobile industry.
More recently, we’ve seen gains from AI for combinatorial optimization, leading to scalability of the method, as well as significant reductions in cost. This method replaces the manual tuning of traditional heuristic approaches with an AI agent that provides a fast metric estimation.
In this presentation you will find out:
Why AI is crucial in combinatorial optimization
How it can be applied to two use cases: improving chip design and hardware-specific compilers
The state-of-the-art results achieved by Qualcomm AI Research
Mavenir: Why and How Private LTE & 5G Networks Are Rapidly Evolving for Enter...Mavenir
Dean Bubley, Founder of Disruptive Analysis and well known industry analyst, and Aniruddho Basu, Mavenir SVP/GM of Global Emerging Business, showcase the future of Private LTE & 5G Networks. Presentation from the "Why and How Private LTE & 5G Networks Are Rapidly Evolving for Enterprises" webinar.
5G + AI: The Ingredients For Next Generation Wireless InnovationQualcomm Research
5G and AI are two of the most disruptive technologies the world has seen in decades. While each is individually revolutionizing industries and enabling new experiences, the combination of both 5G and AI is going to be truly transformative. Applying AI not only to the 5G network but also the device will lead to more efficient wireless communications, longer battery life and enhanced user experiences. The low latency and high capacity of 5G will also allow AI processing to be distributed amongst the device, edge cloud and central cloud, enabling flexible system solutions for a variety of use cases. At Qualcomm Technologies, we are not only working on cutting-edge research for 5G and AI, but we are also exploring their synergies to realize our vision of the future. View this presentation to learn how AI is making 5G better -- in the network and on the device, why on-device AI processing is essential, and how 5G is empowering distributed learning over wireless.
AI firsts: Leading from research to proof-of-conceptQualcomm Research
AI has made tremendous progress over the past decade, with many advancements coming from fundamental research from many decades ago. Accelerating the pipeline from research to commercialization has been daunting since scaling technologies in the real world faces many challenges beyond the theoretical work done in the lab. Qualcomm AI Research has taken on the task of not only generating novel AI research but also being first to demonstrate proof-of-concepts on commercial devices, enabling technology to scale in the real world. This presentation covers:
The challenges of deploying cutting-edge research on real-world mobile devices
How Qualcomm AI Research is solving system and feasibility challenges with full-stack optimizations to quickly move from research to commercialization
Examples where Qualcomm AI Research has had industrial or academic firsts
Transforming enterprise and industry with 5G private networksQualcomm Research
The 3GPP put the spotlight on industry expansion in July with 5G NR Release 16 and set the stage for enterprise and industry verticals to look at how to provide high-performance wireless connectivity with 5G private networks. With a variety of options for spectrum, different network architectures, a rich feature set to meet the demanding needs of the industrial Internet of Things (IIoT), and the privacy and security required for business assurance, 5G private networks are poised to transform enterprise and industry.
Watch the webinar at: https://pages.questexnetwork.com/Webinar-Qualcomm-Registration-101520.html?source=Qualcomm
3GPP Release 17: Completing the first phase of 5G evolutionQualcomm Research
This presentation summarizes 5G NR Release 17 projects that was completed in March 2022. It further enhances 5G foundation and expands into new devices, use cases, verticals.
Mavenir: Why and How Private LTE & 5G Networks Are Rapidly Evolving for Enter...Mavenir
Dean Bubley, Founder of Disruptive Analysis and well known industry analyst, and Aniruddho Basu, Mavenir SVP/GM of Global Emerging Business, showcase the future of Private LTE & 5G Networks. Presentation from the "Why and How Private LTE & 5G Networks Are Rapidly Evolving for Enterprises" webinar.
5G + AI: The Ingredients For Next Generation Wireless InnovationQualcomm Research
5G and AI are two of the most disruptive technologies the world has seen in decades. While each is individually revolutionizing industries and enabling new experiences, the combination of both 5G and AI is going to be truly transformative. Applying AI not only to the 5G network but also the device will lead to more efficient wireless communications, longer battery life and enhanced user experiences. The low latency and high capacity of 5G will also allow AI processing to be distributed amongst the device, edge cloud and central cloud, enabling flexible system solutions for a variety of use cases. At Qualcomm Technologies, we are not only working on cutting-edge research for 5G and AI, but we are also exploring their synergies to realize our vision of the future. View this presentation to learn how AI is making 5G better -- in the network and on the device, why on-device AI processing is essential, and how 5G is empowering distributed learning over wireless.
AI firsts: Leading from research to proof-of-conceptQualcomm Research
AI has made tremendous progress over the past decade, with many advancements coming from fundamental research from many decades ago. Accelerating the pipeline from research to commercialization has been daunting since scaling technologies in the real world faces many challenges beyond the theoretical work done in the lab. Qualcomm AI Research has taken on the task of not only generating novel AI research but also being first to demonstrate proof-of-concepts on commercial devices, enabling technology to scale in the real world. This presentation covers:
The challenges of deploying cutting-edge research on real-world mobile devices
How Qualcomm AI Research is solving system and feasibility challenges with full-stack optimizations to quickly move from research to commercialization
Examples where Qualcomm AI Research has had industrial or academic firsts
Transforming enterprise and industry with 5G private networksQualcomm Research
The 3GPP put the spotlight on industry expansion in July with 5G NR Release 16 and set the stage for enterprise and industry verticals to look at how to provide high-performance wireless connectivity with 5G private networks. With a variety of options for spectrum, different network architectures, a rich feature set to meet the demanding needs of the industrial Internet of Things (IIoT), and the privacy and security required for business assurance, 5G private networks are poised to transform enterprise and industry.
Watch the webinar at: https://pages.questexnetwork.com/Webinar-Qualcomm-Registration-101520.html?source=Qualcomm
3GPP Release 17: Completing the first phase of 5G evolutionQualcomm Research
This presentation summarizes 5G NR Release 17 projects that was completed in March 2022. It further enhances 5G foundation and expands into new devices, use cases, verticals.
- There is a rich roadmap of 5G technologies coming in the second half of the 5G decade with the 5G Advanced evolution
- 6G will be the future innovation platform for 2030 and beyond building on the 5G Advanced foundation
- 6G will be more than just a new radio design, expanding the role of AI, sensing and others in the connected intelligent edge
- Qualcomm is leading cutting-edge wireless research across six key technology vectors on the path to 6G
5G will transform the IoT, expanding the reach of 5G and mobile technologies beyond smartphones. This presentation talks about how 5G is opening doors to new use cases, what is in the 5G evolution that will address the expanding IoT needs, and what Qualcomm is doing to deliver end-to-end technologies and solutions.
Beginners: Energy Consumption in Mobile Networks - RAN Power Saving Schemes3G4G
This tutorial looks at energy consumption in the mobile networks, especially 4G and 5G and looks at various ways in which the vendors and standards are working on to reduce the power consumption.
At a high level, there are three layers of optimisation: Network level, Site level and Equipment level. This presentation looks at some of the ways the optimisation is achieved.
There is a long list of references available for anyone interested in researching this topic further.
All our #3G4G5G slides and videos are available at:
Videos: https://www.youtube.com/3G4G5G
Slides: https://www.slideshare.net/3G4GLtd
5G Page: https://www.3g4g.co.uk/5G/
Free Training Videos: https://www.3g4g.co.uk/Training/
Next-Generation Wireless Overview & Outlook Update 12/8/21Mark Goldstein
Mark Goldstein of International Research Center presented a Next-Gen Wireless Overview & Outlook to the IEEE Computer Society Phoenix (https://ewh.ieee.org/r6/phoenix/compsociety/) on Wednesday, 12/8/21. He explored the next-generation wireless landscape with its underlying emerging technologies, protocols & standards, market trends & opportunities in a deep dive presentation covering all of today's wireless essentials. New spectrum and technologies driven by a rapidly evolving application landscape will be served up in innovative ways through 5G/6G mobile, Wi-Fi 6E, CBRS, White Space, mmWave, satellite & varieties of LPWAN connecting billions of new IoT sensors & devices spread around smart spaces & enabling autonomous transportation. Explore emerging wireless advances, roadblocks & operational challenges bringing you the insight and strategies to leverage emerging wireless opportunities going forward.
6G: Potential Use Cases and Enabling Technologies3G4G
This white paper presents an overview of some of the promising applications and use case envisioned for 6G, with the objective to highlight the potential for new markets and to provide an indication of the expected technical requirements. The white paper then describes some of the enabling technologies for meeting the performance requirements of 6G.
Authors: Ritvik Gupta, Student (A-Levels), Sutton Grammar School, London, United Kingdom under the supervision of Dr Biplab Sidkar, Associate Professor , Dept of Electrical and Computer Engineering, National University of Singapore
This presentation outlines the synergistic nature of 5G and AI -- two disruptive areas of innovations that can change the world. It illustrates the benefits of adopting AI for the advancements of 5G, as well as showcases the latest progress made by Qualcomm Technologies, Inc.
While 6G is as yet 10 years away, organizations that keep themselves up to date with what this next networking architecture has to bring to the table will have a decisive advantage over their competitors.
While 5G commercialization is still in its initial stage, it's never too soon to begin planning for 6G on the grounds that it regularly takes around 10 years from the beginning of exploration to commercialization of new generation of communications technology.
Most tech insiders trust 6G would need to hit two or three benchmarks first beginning with a hyper-quick information rate that beats 5G, with download paces of at any rate 1,000Gbps (multiple times the speed of 5G) and record-breaking all-time low latency, or “air latency” of under 100 μs, start to finish (E2E) inertness under 1 ms, and amazingly low delay jitter in order of microseconds.
Technology doesn't rest.
Despite the fact that the 5G Technology is still in the beginning of its arrangement, the tech world is now bantering around thoughts on what the next generation – 6G – might resemble.
This isn't startling on the grounds that the technology that goes into 5G's replacement will set aside some effort to create.
Tonex offers 6G Introduction, IMT-2030, a one-day course covering the planning inspiration and basic technology of 6G engineering, just like the new 6G terminology. Members are familiar with the comparison between 5G and 6G, and understand how 6G achieves its goals by understanding the functions of 6G.
Introduction to 6G Course by Tonex
IMT-2030 is an introduction to 6G, a one-day overview of 6G technology consistent with ITU-T IMT-2030. Learn about 6G wireless systems, use cases, applications, trends, technologies and protocols.
6G or IMT-2030 is the future of mobile networks promised by ITU-T network 2030. Tonex now offers training courses to help develop the next generation of 6G skills.
Who Should Attend
This one-day training course covers the design motivation and basic technology of 6G architecture, as well as new 6G vocabulary. You will also understand the difference between 5G and 6G, and understand how 6G will achieve its goals by observing how 6G works.
An advanced 6G technical overview for anyone involved in 6G product development, product management, analysis, planning, design and engineering.
Learning Objectives
Describe the 6G vision and business case Explain the key technologies and basic components of 6G networks
Draw end-to-end 6G network architecture, including new radio types, core networks and applications
Gradually complete the evolution from 5G to 6G
Course outline
Overview of 6G Wireless Networks
6G Vision, Architecture, and Key Technologies
Hologram Type Communications
Learn More
Introduction to 6G, Prepare Now Training
https://www.tonex.com/introduction-to-6g-prepare-now-training/
Next IIoT wave: embedded digital twin for manufacturing IRS srl
Next IIoT wave will be a population of digital twin. A digital twin is a real-time digital replica of a physical device. Developing an embedded digital twin allows superior device diagnostic and failure anticipation. Discover how to to implement an embedded digital twin using real-time monitoring, physical models, and machine learning
3D perception is crucial for understanding the real world. It offers many benefits and new capabilities over 2D across diverse applications, from XR and autonomous driving to IOT, camera, and mobile. 3D perception with machine learning is creating the new state of the art (SOTA) in areas, such as depth estimation, object detection, and neural scene representation. Making these SOTA neural networks feasible for real-world deployment on mobile devices constrained by power, thermal, and performance has been a challenge. Qualcomm AI Research has developed not only novel AI techniques for 3D perception but also full-stack AI optimizations to enable real-world deployments and energy-efficient solutions. This presentation explores the latest research that is enabling efficient 3D perception while maintaining neural network model accuracy. You’ll learn about:
- The advantages of 3D perception over 2D and the need for 3D perception across applications
- Advancements in 3D perception research by Qualcomm AI Research
- Our future 3D perception research directions
Ericsson brings new updates to its 5G platform. Introducing 5G network services to support operators from preparation to 5G launch. Ericsson 5G services roadmap spans across three distinct phases, Prepare, Mobilize and Launch. Through our service offerings, Operators can now evolve their 4G network and smoothly start introducing 5G, reaching new heights on their journey to 5G.
This presentation and video looks at the concept of Open RAN, White Bix RAN and Virtualized RAN (vRAN). It looks at the motivation to move away from traditional architectures where the vendor supplies their own proprietary hardware and software to the new Open RAN architecture movement. Business case from an MNO / SP point of view is discussed and the results from joint Open RAN RFI by Telefonica and Vodafone is discussed.
AI model efficiency is crucial for making AI ubiquitous, leading to smarter devices and enhanced lives. Besides the performance benefit, quantized neural networks also increase power efficiency for two reasons: reduced memory access costs and increased compute efficiency.
The quantization work done by the Qualcomm AI Research team is crucial in implementing machine learning algorithms on low-power edge devices. In network quantization, we focus on both pushing the state-of-the-art (SOTA) in compression and making quantized inference as easy to access as possible. For example, our SOTA work on oscillations in quantization-aware training that push the boundaries of what is possible with INT4 quantization. Furthermore, for ease of deployment, the integer formats such as INT16 and INT8 give comparable performance to floating point, i.e., FP16 and FP8, but have significantly better performance-per-watt performance. Researchers and developers can make use of this quantization research to successfully optimize and deploy their models across devices with open-sourced tools like AI Model Efficiency Toolkit (AIMET).
Presenters: Tijmen Blankevoort and Chirag Patel
Setting off the 5G Advanced evolution with 3GPP Release 18Qualcomm Research
In December 2021, 3GPP has reached a consensus on the scope of 5G NR Release 18. This is a significant milestone marking the beginning of 5G Advanced — the second wave of wireless innovations that will fulfill the 5G vision. Release 18 will build on the solid foundation set by Releases 15, 16, and 17, and it sets the longer-term evolution direction of 5G and beyond. This release will encompass a wide range of new and enhancement projects, ranging from improved MIMO and application of AI/ML-enabled air interface to extended reality optimizations and broader IoT support.
5G has been projected as a catalyst in driving the adoption of AI, IoT and robotics. The security environment as we know in a pre-5G world is already relatively complex. Various discussions over its security concern are being raised, in anticipation of 5G especially in its use in IoT and autonomous systems.
The speaker intends to place the application of 5G in perspective of enterprise ecosystem. By breaking down the ecosystem into components, reviewing the concerns and requirements, the resolution and implementation of security controls could be mapped into respective roles, such as telco engineer, IoT developer, security professional and enterprise architect. Bringing us to a shared responsibility model, the key to creating a more secure 5G and IoT environment.
Artificial Intelligence (AI), specifically deep learning, is revolutionizing industries, products, and core capabilities by delivering dramatically enhanced experiences. However, the deep neural networks of today use too much memory, compute, and energy. To make AI truly ubiquitous, it needs to run on the end device within tight power and thermal budgets. Advancements in multiple areas are necessary to improve AI model efficiency, including quantization, compression, compilation, and neural architecture search (NAS). In this presentation, we’ll discuss:
- Qualcomm AI Research’s latest model efficiency research
- Our new NAS research to optimize neural networks more easily for on-device efficiency
- How the AI community can take advantage of this research though our open-source projects, such as the AI Model Efficiency Toolkit (AIMET) and AIMET Model Zoo
ISC Frankfurt 2015: Good, bad and ugly of accelerators and a complementary pathJohn Holden
Accelerators Vs Adjoint Algorithmic Differentation (AAD).... NONSENSE. It is not a choice. The two can be combined to provide the ultimate accelerator. Accelerators such as NVIDIA GPUs, Intel Xeon Phis CAN be combined with AD. NAG has the software tools and expertise to deliver AD solutions for traditional architectures and accelerarors
- There is a rich roadmap of 5G technologies coming in the second half of the 5G decade with the 5G Advanced evolution
- 6G will be the future innovation platform for 2030 and beyond building on the 5G Advanced foundation
- 6G will be more than just a new radio design, expanding the role of AI, sensing and others in the connected intelligent edge
- Qualcomm is leading cutting-edge wireless research across six key technology vectors on the path to 6G
5G will transform the IoT, expanding the reach of 5G and mobile technologies beyond smartphones. This presentation talks about how 5G is opening doors to new use cases, what is in the 5G evolution that will address the expanding IoT needs, and what Qualcomm is doing to deliver end-to-end technologies and solutions.
Beginners: Energy Consumption in Mobile Networks - RAN Power Saving Schemes3G4G
This tutorial looks at energy consumption in the mobile networks, especially 4G and 5G and looks at various ways in which the vendors and standards are working on to reduce the power consumption.
At a high level, there are three layers of optimisation: Network level, Site level and Equipment level. This presentation looks at some of the ways the optimisation is achieved.
There is a long list of references available for anyone interested in researching this topic further.
All our #3G4G5G slides and videos are available at:
Videos: https://www.youtube.com/3G4G5G
Slides: https://www.slideshare.net/3G4GLtd
5G Page: https://www.3g4g.co.uk/5G/
Free Training Videos: https://www.3g4g.co.uk/Training/
Next-Generation Wireless Overview & Outlook Update 12/8/21Mark Goldstein
Mark Goldstein of International Research Center presented a Next-Gen Wireless Overview & Outlook to the IEEE Computer Society Phoenix (https://ewh.ieee.org/r6/phoenix/compsociety/) on Wednesday, 12/8/21. He explored the next-generation wireless landscape with its underlying emerging technologies, protocols & standards, market trends & opportunities in a deep dive presentation covering all of today's wireless essentials. New spectrum and technologies driven by a rapidly evolving application landscape will be served up in innovative ways through 5G/6G mobile, Wi-Fi 6E, CBRS, White Space, mmWave, satellite & varieties of LPWAN connecting billions of new IoT sensors & devices spread around smart spaces & enabling autonomous transportation. Explore emerging wireless advances, roadblocks & operational challenges bringing you the insight and strategies to leverage emerging wireless opportunities going forward.
6G: Potential Use Cases and Enabling Technologies3G4G
This white paper presents an overview of some of the promising applications and use case envisioned for 6G, with the objective to highlight the potential for new markets and to provide an indication of the expected technical requirements. The white paper then describes some of the enabling technologies for meeting the performance requirements of 6G.
Authors: Ritvik Gupta, Student (A-Levels), Sutton Grammar School, London, United Kingdom under the supervision of Dr Biplab Sidkar, Associate Professor , Dept of Electrical and Computer Engineering, National University of Singapore
This presentation outlines the synergistic nature of 5G and AI -- two disruptive areas of innovations that can change the world. It illustrates the benefits of adopting AI for the advancements of 5G, as well as showcases the latest progress made by Qualcomm Technologies, Inc.
While 6G is as yet 10 years away, organizations that keep themselves up to date with what this next networking architecture has to bring to the table will have a decisive advantage over their competitors.
While 5G commercialization is still in its initial stage, it's never too soon to begin planning for 6G on the grounds that it regularly takes around 10 years from the beginning of exploration to commercialization of new generation of communications technology.
Most tech insiders trust 6G would need to hit two or three benchmarks first beginning with a hyper-quick information rate that beats 5G, with download paces of at any rate 1,000Gbps (multiple times the speed of 5G) and record-breaking all-time low latency, or “air latency” of under 100 μs, start to finish (E2E) inertness under 1 ms, and amazingly low delay jitter in order of microseconds.
Technology doesn't rest.
Despite the fact that the 5G Technology is still in the beginning of its arrangement, the tech world is now bantering around thoughts on what the next generation – 6G – might resemble.
This isn't startling on the grounds that the technology that goes into 5G's replacement will set aside some effort to create.
Tonex offers 6G Introduction, IMT-2030, a one-day course covering the planning inspiration and basic technology of 6G engineering, just like the new 6G terminology. Members are familiar with the comparison between 5G and 6G, and understand how 6G achieves its goals by understanding the functions of 6G.
Introduction to 6G Course by Tonex
IMT-2030 is an introduction to 6G, a one-day overview of 6G technology consistent with ITU-T IMT-2030. Learn about 6G wireless systems, use cases, applications, trends, technologies and protocols.
6G or IMT-2030 is the future of mobile networks promised by ITU-T network 2030. Tonex now offers training courses to help develop the next generation of 6G skills.
Who Should Attend
This one-day training course covers the design motivation and basic technology of 6G architecture, as well as new 6G vocabulary. You will also understand the difference between 5G and 6G, and understand how 6G will achieve its goals by observing how 6G works.
An advanced 6G technical overview for anyone involved in 6G product development, product management, analysis, planning, design and engineering.
Learning Objectives
Describe the 6G vision and business case Explain the key technologies and basic components of 6G networks
Draw end-to-end 6G network architecture, including new radio types, core networks and applications
Gradually complete the evolution from 5G to 6G
Course outline
Overview of 6G Wireless Networks
6G Vision, Architecture, and Key Technologies
Hologram Type Communications
Learn More
Introduction to 6G, Prepare Now Training
https://www.tonex.com/introduction-to-6g-prepare-now-training/
Next IIoT wave: embedded digital twin for manufacturing IRS srl
Next IIoT wave will be a population of digital twin. A digital twin is a real-time digital replica of a physical device. Developing an embedded digital twin allows superior device diagnostic and failure anticipation. Discover how to to implement an embedded digital twin using real-time monitoring, physical models, and machine learning
3D perception is crucial for understanding the real world. It offers many benefits and new capabilities over 2D across diverse applications, from XR and autonomous driving to IOT, camera, and mobile. 3D perception with machine learning is creating the new state of the art (SOTA) in areas, such as depth estimation, object detection, and neural scene representation. Making these SOTA neural networks feasible for real-world deployment on mobile devices constrained by power, thermal, and performance has been a challenge. Qualcomm AI Research has developed not only novel AI techniques for 3D perception but also full-stack AI optimizations to enable real-world deployments and energy-efficient solutions. This presentation explores the latest research that is enabling efficient 3D perception while maintaining neural network model accuracy. You’ll learn about:
- The advantages of 3D perception over 2D and the need for 3D perception across applications
- Advancements in 3D perception research by Qualcomm AI Research
- Our future 3D perception research directions
Ericsson brings new updates to its 5G platform. Introducing 5G network services to support operators from preparation to 5G launch. Ericsson 5G services roadmap spans across three distinct phases, Prepare, Mobilize and Launch. Through our service offerings, Operators can now evolve their 4G network and smoothly start introducing 5G, reaching new heights on their journey to 5G.
This presentation and video looks at the concept of Open RAN, White Bix RAN and Virtualized RAN (vRAN). It looks at the motivation to move away from traditional architectures where the vendor supplies their own proprietary hardware and software to the new Open RAN architecture movement. Business case from an MNO / SP point of view is discussed and the results from joint Open RAN RFI by Telefonica and Vodafone is discussed.
AI model efficiency is crucial for making AI ubiquitous, leading to smarter devices and enhanced lives. Besides the performance benefit, quantized neural networks also increase power efficiency for two reasons: reduced memory access costs and increased compute efficiency.
The quantization work done by the Qualcomm AI Research team is crucial in implementing machine learning algorithms on low-power edge devices. In network quantization, we focus on both pushing the state-of-the-art (SOTA) in compression and making quantized inference as easy to access as possible. For example, our SOTA work on oscillations in quantization-aware training that push the boundaries of what is possible with INT4 quantization. Furthermore, for ease of deployment, the integer formats such as INT16 and INT8 give comparable performance to floating point, i.e., FP16 and FP8, but have significantly better performance-per-watt performance. Researchers and developers can make use of this quantization research to successfully optimize and deploy their models across devices with open-sourced tools like AI Model Efficiency Toolkit (AIMET).
Presenters: Tijmen Blankevoort and Chirag Patel
Setting off the 5G Advanced evolution with 3GPP Release 18Qualcomm Research
In December 2021, 3GPP has reached a consensus on the scope of 5G NR Release 18. This is a significant milestone marking the beginning of 5G Advanced — the second wave of wireless innovations that will fulfill the 5G vision. Release 18 will build on the solid foundation set by Releases 15, 16, and 17, and it sets the longer-term evolution direction of 5G and beyond. This release will encompass a wide range of new and enhancement projects, ranging from improved MIMO and application of AI/ML-enabled air interface to extended reality optimizations and broader IoT support.
5G has been projected as a catalyst in driving the adoption of AI, IoT and robotics. The security environment as we know in a pre-5G world is already relatively complex. Various discussions over its security concern are being raised, in anticipation of 5G especially in its use in IoT and autonomous systems.
The speaker intends to place the application of 5G in perspective of enterprise ecosystem. By breaking down the ecosystem into components, reviewing the concerns and requirements, the resolution and implementation of security controls could be mapped into respective roles, such as telco engineer, IoT developer, security professional and enterprise architect. Bringing us to a shared responsibility model, the key to creating a more secure 5G and IoT environment.
Artificial Intelligence (AI), specifically deep learning, is revolutionizing industries, products, and core capabilities by delivering dramatically enhanced experiences. However, the deep neural networks of today use too much memory, compute, and energy. To make AI truly ubiquitous, it needs to run on the end device within tight power and thermal budgets. Advancements in multiple areas are necessary to improve AI model efficiency, including quantization, compression, compilation, and neural architecture search (NAS). In this presentation, we’ll discuss:
- Qualcomm AI Research’s latest model efficiency research
- Our new NAS research to optimize neural networks more easily for on-device efficiency
- How the AI community can take advantage of this research though our open-source projects, such as the AI Model Efficiency Toolkit (AIMET) and AIMET Model Zoo
ISC Frankfurt 2015: Good, bad and ugly of accelerators and a complementary pathJohn Holden
Accelerators Vs Adjoint Algorithmic Differentation (AAD).... NONSENSE. It is not a choice. The two can be combined to provide the ultimate accelerator. Accelerators such as NVIDIA GPUs, Intel Xeon Phis CAN be combined with AD. NAG has the software tools and expertise to deliver AD solutions for traditional architectures and accelerarors
AI Solutions for Industries | Quality Inspection | Data Insights | AI-accelerated CFD | Self-Checkout | byteLAKE.com
byteLAKE: Empowering Industries with AI Solutions. Embrace cutting-edge technology for advanced quality inspection, data insights, and more. Harness the potential of our CFD Suite, accelerating Computational Fluid Dynamics for heightened productivity. Unlock new possibilities with Cognitive Services: image analytics for precise visual inspection for Manufacturing, sound analytics enabling proactive maintenance for Automotive, and wet line analytics for the Paper Industry. Seamlessly convert data into actionable insights using Data Insights' AI module, enabling advanced predictive maintenance and risk detection. Simplify Restaurant and Retail operations with our efficient self-checkout solution, recognizing meals and groceries and elevating customer satisfaction. Custom AI Development services available for tailored solutions. Discover more at www.byteLAKE.com.
► byteLAKE's CFD Suite: Accelerate your Computational Fluid Dynamics (CFD) simulations by leveraging the speed and efficiency of artificial intelligence. Slash simulation times, minimize trial-and-error costs, and supercharge decision-making for heightened productivity. Learn more at www.byteLAKE.com/en/CFDSuite.
The increasing demand for computing power in fields such as biology, finance, machine learning is pushing the adoption of reconfigurable hardware in order to keep up with the required performance level at a sustainable power consumption. Within this context, FPGA devices represent an interesting solution as they combine the benefits of power efficiency, performance and flexibility. Nevertheless, the steep learning curve and experience needed to develop efficient FPGA-based systems represents one of the main limiting factor for a broad utilization of such devices.
In this talk, we present CAOS, a framework which helps the application designer in identifying acceleration opportunities and guides through the implementation of the final FPGA-based system. The CAOS platform targets the full stack of the application optimization process, starting from the identification of the kernel functions to accelerate, to the optimization of such kernels and to the generation of the runtime management and the configuration files needed to program the FPGA.
Den Datenschatz heben und Zeit- und Energieeffizienz steigern: Mathematik und...Joachim Schlosser
In einer Gesellschaft, in der das Sammeln von personenbezogenen Daten mittlerweile alltäglich geworden ist, ist es nicht weiter verwunderlich, dass auch der innovative Maschinenbauer Daten sammelt, wo er nur kann. Produktdaten, Maschinendaten, Statistikdaten – in einer durchschnittlichen Produktionsanlage fallen bereits heute jeden Tag Gigabytes an Daten an. „Big Data“ wurde eines der Schlagworte der Industrie 4.0.
Doch was verspricht man sich davon? Welche Information steckt in den aufgezeichneten Maschinen- und Produktdaten? Und wie erfolgt die Auswertung?
Im Rahmen des Vortrags wird aufgezeigt, wie Unternehmen auf Basis einer etablierten Plattform wie MATLAB® ihre Auswertealgorithmen entwickeln, testen und ausrollen können. Die kontinuierliche Auswertung selbst erfolgt dann wahlweise auf einem Anlagenserver oder aber auch in Echtzeit direkt an der Maschine. Veranschaulicht wird dies anhand von Beispielen aus der Praxis.
Doch neben der gesammelten Daten kommt auch den Steuerungseinheiten in der Produktion in der Industrie 4.0 eine größere Bedeutung zu.
Wenn Werkstücke demnächst selbst wissen, wo sie im Produktionsablauf hin möchten und welcher Verarbeitungsschritt ihnen angedeihen soll, dann bedeutet das auch für die einzelnen Komponenten und Module in Produktion und Logistik ein mehr an Funktionalität, da sie auf diese Eingaben ebenfalls reagieren sollen.
Wie stellen Sie sicher, dass diese zusätzliche Funktionalität nicht zu Lasten der Energiebilanz gehen? Wie fahren Sie die Motoren und anderen aktiven Komponenten Ihrer Fertigung so, dass sie flexibel auf veränderte Routen der Werkstücke reagieren und dennoch im optimalen Bereich fahren?
Mehr denn je brauchen Sie gesteuerte und geregelte Komponenten und Module. Das sollte schon seit Industrie 3.0 vorhanden sein, jedoch ist auch hier noch viel ganz konkretes Potential zur Steigerung von Produktivität und Einsparung von Energie und Produktionszeit vorhanden.
Sie sehen im Vortrag, wie Sie ihre Komponenten besser beschalten, dass die vernetzten dynamischen Anforderungen von Industrie 4.0 lokal effizient umgesetzt werden können.
Artificial Intelligence in practice - Gerbert Kaandorp - Codemotion Amsterdam...Codemotion
In this talk Gerbert will give an overview of Artificial Intelligence, outline the current state of the art in research and explain what it takes to actually do an AI project. Using practical cases and tools he will give you insight in the phases of an AI project and explain some of the problems you might encounter along the way and how you might be able to solve them.
- 소개
2018년 11월 2일, Tech Meets Startup 발표자료
http://tech-startup.kr/
- 발표 제목: 글로벌 격전지에서 발견한 기회 : 기술 스타트업을 위한 궁극의 엔지니어링
- 발표자: FuriosaAI 백준호 CEO
- 내용: AI 반도체를 개발하고 있는 FuriosaAI가 글로벌 기업들의 틈에서 어떻게 기회를 발견하고, 스스로의 존재감을 만들어내고 있는지를 소개합니다. 치열한 경쟁 속에서 고군분투하는 여러 기술 스타트업에 도움되는 세션이길 바랍니다.
Generative AI models, such as ChatGPT and Stable Diffusion, can create new and original content like text, images, video, audio, or other data from simple prompts, as well as handle complex dialogs and reason about problems with or without images. These models are disrupting traditional technologies, from search and content creation to automation and problem solving, and are fundamentally shaping the future user interface to computing devices. Generative AI can apply broadly across industries, providing significant enhancements for utility, productivity, and entertainment. As generative AI adoption grows at record-setting speeds and computing demands increase, on-device and hybrid processing are more important than ever. Just like traditional computing evolved from mainframes to today’s mix of cloud and edge devices, AI processing will be distributed between them for AI to scale and reach its full potential.
In this presentation you’ll learn about:
- Why on-device AI is key
- Full-stack AI optimizations to make on-device AI possible and efficient
- Advanced techniques like quantization, distillation, and speculative decoding
- How generative AI models can be run on device and examples of some running now
- Qualcomm Technologies’ role in scaling on-device generative AI
As generative AI adoption grows at record-setting speeds and computing demands increase, hybrid processing is more important than ever. But just like traditional computing evolved from mainframes and thin clients to today’s mix of cloud and edge devices, AI processing must be distributed between the cloud and devices for AI to scale and reach its full potential. In this talk you’ll learn:
• Why on-device AI is key
• Which generative AI models can run on device
• Why the future of AI is hybrid
• Qualcomm Technologies’ role in making hybrid AI a reality
5G is going mainstream across the globe, and this is an exciting time to harness the low latency and high capacity of 5G to enable the metaverse. A distributed-compute architecture across device and cloud can enable rich extended reality (XR) user experiences. Virtual reality (VR) and mixed reality (MR) are ready for deployment in private networks, while augmented reality (AR) for wide area networks can be enabled in the near term with Wi-Fi powered AR glasses paired with a 5G-enabled phone. Device APIs enabling application adaptation is critical for good user experience. 5G standards are evolving to support the deployment of AR glasses at a large scale and setting the stage for 6G-era with the merging of the physical, digital, and virtual worlds. Techniques like perception-enhanced wireless offer significant potential to improve user experience. Qualcomm Technologies is enabling the XR industry with platforms, developer SDKs, and reference designs.
Check out this webinar to learn:
• How 5G and distributed-compute architectures enable the metaverse
• The latest results from our boundless XR 5G/6G testbed, including device APIs and perception-enhanced wireless
• 5G standards evolution for enhancing XR applications and the road to 6G
• How Qualcomm Technologies is enabling the industry with platforms, SDKs, and reference designs
Bringing AI research to wireless communication and sensingQualcomm Research
AI for wireless is already here, with applications in areas such as mobility management, sensing and localization, smart signaling and interference management. Recently, Qualcomm Technologies has prototyped the AI-enabled air interface and launched the Qualcomm 5G AI Suite. These developments are possible thanks to expertise in both wireless and machine learning from over a decade of foundational research in these complementing fields.
Our approach brings together the modeling flexibility and computational efficiency of machine learning and the out-of-domain generalization and interpretability of wireless domain expertise.
In this webinar, Qualcomm AI Research presents an overview of state-of-the-art research at the intersection of the two fields and offers a glimpse into the future of the wireless industry.
Qualcomm AI Research is an initiative of Qualcomm Technologies, Inc.
Speakers:
Arash Behboodi, Machine Learning Research Scientist (Senior Staff Engineer/Manager), Qualcomm AI Research Daniel Dijkman, Machine Learning Research Scientist (Principal Engineer), Qualcomm AI Research
How will sidelink bring a new level of 5G versatility.pdfQualcomm Research
Today, the 5G system mainly operates on a network-to-device communication model, exemplified by enhanced mobile broadband use cases where all data transmissions are between the network (i.e., base station) and devices (e.g., smartphone). However, to fully deliver on the original 5G vision of supporting diverse devices, services, and deployment scenarios, we need to expand the 5G topology further to reach new levels of performance and efficiency.
That is why sidelink communication was introduced in 3GPP standards, designed to facilitate direct communication between devices, independent of connectivity via the cellular infrastructure. Beyond automotive communication, it also benefits many other 5G use cases such as IoT, mobile broadband, and public safety.
5G is designed to serve an unprecedented range of capabilities with a single global standard. With enhanced mobile broadband (eMBB), massive IoT (mIoT), and mission-critical IoT, the three pillars of 5G represent extremes in performance and associated complexity. For IoT services, NB-IoT and eMTC devices prioritize low power consumption and the lowest complexity for wide-area deployments (LPWA), while enhanced ultra-reliable, low-latency communication (eURLLC), along with time-sensitive networking (TSN), delivers the most stringent use case requirements. But there exists an opportunity to more efficiently address a broad range of mid-tier applications with capabilities ranging between these extremes.
In 5G NR Release 17, 3GPP introduced a new tier of reduced capability (RedCap) devices, also known as NR-Light. It is a new device platform that bridges the capability and complexity gap between the extremes in 5G today with an optimized design for mid-tier use cases. With the recent standards completion, NR-Light is set to efficiently expand the 5G universe to connect new frontiers.
Download this presentation to learn:
• What NR-Light is and why it can herald the next wave of 5G expansion
• How NR-Light is accelerating the growth of the connected intelligent edge
• Why NR-Light is a suitable 5G migration path for mid-tier LTE devices
Realizing mission-critical industrial automation with 5GQualcomm Research
Manufacturers seeking better operational efficiencies, with reduced downtime and higher yield, are at the leading edge of the Industry 4.0 transformation. With mobile system components and reliable wireless connectivity between them, flexible manufacturing systems can be reconfigured quickly for new tasks, to troubleshoot issues, or in response to shifts in supply and demand.
There is a long history of R&D collaboration between Bosch Rexroth and Qualcomm Technologies for the effective application of these 5G capabilities to industrial automation use cases. At the Robert Bosch Elektronik GmbH factory in Salzgitter, Germany, this collaboration has reached new heights.
Download this deck to learn how:
• Qualcomm Technologies and Bosch Rexroth are collaborating to accelerate the Industry 4.0 transformation
• 5G technologies deliver key capabilities for mission-critical industrial automation
• Distributed control solutions can work effectively across 5G TSN networks
• A single 5G technology platform solves connectivity and positioning needs for flexible manufacturing
Cellular networks have facilitated positioning in addition to voice or data communications from the beginning, since 2G, and we’ve since grown to rely on positioning technology to make our lives safer, simpler, more productive, and even fun. Cellular positioning complements other technologies to operate indoors and outdoors, including dense urban environments where tall buildings interfere with satellite positioning. It works whether we’re standing still, walking, or in a moving vehicle. With 5G, cellular positioning breaks new ground to bring robust precise positioning indoors and outdoors, to meet even the most demanding Industry 4.0 needs.
As we look to the future, the Connected Intelligent Edge will bring a new dimension of positional insight to a broad range of devices, improving wireless use cases still under development. We’re already charting the course to 5G Advanced and beyond by working on the evolution of cellular positioning technology to include RF sensing for situational awareness.
Download the deck to learn more.
The need for intelligent, personalized experiences powered by AI is ever-growing. Our devices are producing more and more data that could help improve our AI experiences. How do we learn and efficiently process all this data from edge devices while maintaining privacy? On-device learning rather than cloud training can address these challenges. In this presentation, we’ll discuss:
- Why on-device learning is crucial for providing intelligent, personalized experiences without sacrificing privacy
- Our latest research in on-device learning, including few-shot learning, continuous learning, and federated learning
- How we are solving system and feasibility challenges to move from research to commercialization
Data compression has increased by leaps and bounds over the years due to technical innovation, enabling the proliferation of streamed digital multimedia and voice over IP. For example, a regular cadence of technical advancement in video codecs has led to massive reduction in file size – in fact, up to a 1000x reduction in file size when comparing a raw video file to a VVC encoded file. However, with the rise of machine learning techniques and diverse data types to compress, AI may be a compelling tool for next-generation compression, offering a variety of benefits over traditional techniques. In this presentation we discuss:
- Why the demand for improved data compression is growing
- Why AI is a compelling tool for compression in general
- Qualcomm AI Research’s latest AI voice and video codec research
- Our future AI codec research work and challenges
How to build high performance 5G networks with vRAN and O-RANQualcomm Research
5G networks are poised to deliver an unprecedented amount of data from a richer set of use cases than we have ever seen. This makes efficient networking in terms of scalability, cost, and power critical for the sustainable growth of 5G. Cloud technologies such as virtualization, containerization and orchestration are now powering a surge of innovation in virtualized radio access network (vRAN) infrastructure with modular hardware and software components, and standardized interfaces. While commercial off-the-shelf (COTS) hardware platforms provide the compute capacity for running vRAN software, hardware accelerators will also play a major role in offloading real-time and complex signal processing functions. Together, COTS platforms and hardware accelerators provide the foundation for building the intelligent 5G network and facilitate innovative new use cases with the intelligent wireless edge.
This presentation takes a look at the technology roadmap for 5G NR millimeter wave (mmWave). Including features such as integrated access and backhaul (IAB), enhancements in beam management, mobility, coverage, and more. For more information, please visit www.qualcomm.com/mmwave
Video data is abundant and being generated at ever increasing rates. Analyzing video with AI can provide valuable insights and capabilities for many applications ranging from autonomous driving and smart cameras to smartphones and extended reality. However, as video resolution and frame rates increase while AI video perception models become more complex, running these workloads in real time is becoming more challenging. This presentation explores the latest research that is enabling efficient video perception while maintaining neural network model accuracy. You’ll learn about:
- How video perception is crucial for understanding the world and making devices smarter
- The challenges of on-device real-time video perception at high resolution through AI
- Qualcomm AI Research’s latest research and techniques for efficient video perception
Checkout: https://www.qualcomm.com/AI
Enabling the rise of the smartphone: Chronicling the developmental history at...Qualcomm Research
Today’s smartphones are a marvel of modern technology — handheld devices with vast computing power, incredible multimedia and AI capabilities, and blazing fast data rates that support mobile browsing, social media interaction, and more. From humble beginnings as a cellphone focused purely on voice communication, the capability and functionality of modern smartphones have advanced tremendously. This presentation chronicles Qualcomm’s role in the rise of the smartphone from its initial beginnings to becoming the largest computing platform in the world. It includes:
- Key technology developments that led to today’s smartphones
- The role of Moore’s Law in driving new innovations and additional integration into mobile processors
- Qualcomm’s critical role in advancing the smartphone’s capabilities through groundbreaking innovations and key technology developments
This presentation provides an overview of important 5G innovations around new and enhanced use of spectrum. It also captures the current 5G spectrum status across the globe.
Today, we take it for granted that our mobile devices and applications just work out of the box — smartphones can roam virtually anywhere in the world, laptops can seamlessly connect to any Wi-Fi access point & Bluetooth peripheral, and the videos recorded on one device can be played back perfectly on any other device.
The magic behind all this? Technology standards. Not only do they power a wide range of systems and devices but also bring many benefits to the broader technology ecosystem. At Qualcomm Technologies, we are leading the standardization of many key technologies that will move the world forward.
Download this presentation to learn:
- The value of technology standards, specifically in the areas of cellular, Wi-Fi, Bluetooth, and video codecs
- Why standardized technologies are essential for industry growth and ecosystem development
- How standard bodies operate in a complex, challenging, and ever changing environment
- How Qualcomm is driving innovation in different technology standards
Artificial Intelligence (AI) is revolutionizing industries, products, and core capabilities by delivering dramatically enhanced experiences. However, this is just the start of the AI revolution. The field of AI, especially deep learning, is still in its infancy with tremendous opportunity for exploration and improvement. For instance, deep neural networks of today are rapidly growing in size and use too much memory, compute, and energy. To make AI truly ubiquitous, it needs to run on the end device within a tight power and thermal budget. New approaches and fundamental research in AI, as well as applying that research, is required to advance machine learning further and speed up adoption. View this presentation to learn about select research topics that Qualcomm AI Research is investigating, including:
o AI model optimization research for power efficiency, including our latest quantization research
o Applied AI research, such as using deep learning for improved radar functionality
o Fundamental AI research, such as source compression and quantum AI
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Qualcomm Webinar: Solving Unsolvable Combinatorial Problems with AI
1. Chris Lott
Senior Director, Engineering
Qualcomm Technologies, Inc.
January 31, 2023
@QCOMResearch
Solving unsolvable
combinatorial
problems with AI
2. 3
Today’s agenda
• The need for combinatorial optimization
• Solving combinatorial optimization
problems with AI
• Improved chip design with AI
• Improved compilers with AI
• Future directions
• Questions?
3. 4
How do you find an
optimal solution
when faced with
many choices?
Some problems have
more possible solutions
than a game of Go:
~10170
4. 5
Supply chain optimization Hardware-specific compiler
Chip design
Airline network planning
Combinatorial
optimization
problems are
all around us
Finding solutions can
provide significant benefits:
• Reducing cost
• Reducing time
• Increasing performance
5. 6
Traveling Salesman
Problem (TSP)
Given a set of N cities with travel
distance between each pair, find
the shortest path that visits each
city exactly once
Example:
Using a brute force search method,
if a computer can check a solution
in a microsecond, then it would take
2 microseconds to solve 3 cities,
3.6 seconds for 11 cities, and
3857 years for 20 cities.
Exemplifies the combinatorial
optimization problem
6. 7
7
How to approach solving TSP with brute force
An instance of the
traveling salesman problem
Search space
a 100 b
75 100
125
75
50
125
c
125
50
e
100
d
300
d e c e
e
375
a
abcdea
375
a
abceda
425
d
250
d
150
c
b
100
a
125
c
100
d
75
e
e
Each tree leaf node represents
one full tour of the cities
There are (N-1)! paths (tours) in the search
tree. Full enumeration quickly becomes
infeasible as N grows
Search space
(represented as a tree)
• Start at any city
• Choose from N-1 next cities
• Choose from N-2 next cities
• ......
• Choose last city and connect
to start city
7. 8
a 100 b
75 100
125
75
50
125
c
125
50
e
100
d
300
d e c e
e
375
a
abcdea
375
a
abceda
425
d
250
d
150
c
b
100
a
125
c
100
d
75
e
e
N1
cost = 35
N2
cost = 53
N3
cost = 25
N4
cost = 31
N1
cost = 28
N2
cost = 50
N4
cost = 36
N2
cost = 52
N4
cost = 28
N2
cost = 28
N0
cost = 25
Brute force method
An instance of the
traveling salesman problem Search space
• Full path enumeration
• Naive
• Scales as (N-1)!
• infeasible for N>20
Heuristic methods
• The Nearest Neighbor method
• N-opt iterative path improvement
• Not guaranteed optimal
• Heuristics are problem-specific
(human-designed)
Exact solver methods
• Dynamic programming
• Formulate as Integer Linear Programming
(ILP), use Branch and Cut (B&C)
• Uses branch and bound to rule out
whole solution subspaces
• Combine with problem-specific cutting planes
• Scales up to 1000’s of nodes,
but at high computational cost
• Problem-specific
C 7 B
4 7
8
6
6
7
5
9
E
A
D 3
C 7 B
4 7
8
6
6
7
5
9
E
A
D 3
C 7 B
4
7
8
6
6
7
5
9
E
A
D 3
C 7 B
4
7
8
6
6
7
5
9
E
A
D 3
A
C 7 B
4 7
8
6
6
7
5
9
E
D 3
C 7 B
4 7
8
6
6
7
5
9
E
A
D 3
Existing TSP solutions face challenges
8
B&C
ruled out
subspaces
8. 9
Existing combinatorial optimization
techniques have limitations
How can we improve this situation with AI?
Scale
Search heuristics don’t scale to larger problems in
acceptable compute time and cost, and do not
guarantee satisfaction of all the constraints resulting in
expensive manual intervention
Learning
Techniques don’t incorporate knowledge learned
from solving many problems. They start each
new problem instance from scratch.
9. 10
Notable prior work: “Attention, learn to solve routing problems!”, ICLR 2019
AI addresses
challenges of
traditional combinatorial
optimization solutions
Leverages learned
problem structure
Scales to larger instances
Offers a general framework
Can achieve desired outcome
with resource, cost, and time
constraints
Optimization
metric
AI Solver
Develop an AI algorithm that can learn
correlation between design parameters
and optimization metrics from a limited
set of problem instances
𝒑𝒂𝒓𝒂𝒎𝒆𝒕𝒆𝒓𝒔
For new instances, the AI algorithm uses an
existing solver more efficiently by reducing
parameter search space
Heuristics
AI
Standard process
AI process
10. Exploring Bayesian optimization to reduce combinatorial search space
Optimizing chip design
with AI
“Bayesian Optimization for Macro Placement”, ICML 2022
11. 12
12
Competing combinatorial optimization objectives in chip design
Need to account for all business metrics
Chip
Area
Yield – more
complex fab process
Test
Time
Production
cost reduction
Power-performance
optimization
System-level power Chip-level power
Design
efficiency
# of tools iterations License cost
Capital
expense
(Capex)
# of compute servers Emulation platforms
12. 13
Semiconductor
scaling advantage
is approaching a cliff
PPA = Power, Performance, Area
Foundation
Chips
Design
automation
Computing
Key elements
Moore’s Law
PPA scaling
Combinatorial
optimization
Cloud
servers
Disruption
50%
less gain
100x
PVT corners
10x
expense
13. 14
0,01
0,1
1
2012 2014 2016 2018 2020 2022 2024 2026 2028 2030
Scaling
Year
Stdcell
SRAM bitcell
analog/IO
CO needs to compensate for diminishing technology gains and
PVT corner increase within acceptable compute expense 14
Theoretical Moore’s Law
0.5X / 2yr
Analog/IO
Memory
Digital logic
Area scaling over time
14. 15
Macros (memories)
0 (10 — 100)
AND
Standard cells (logic gates)
0 (107
— 109
)
AND
1000 x
Point-like
Challenges
in chip
placement:
How can we solve the chip optimization problem?
1. Placing mixed-size blocks –
standard cells and macros
(memories)
Minimize power and area while
satisfying timing constraints within
limited design resources (people
and compute servers)
2. Scale with increasing
complexity of design (# of
blocks) and constraints
(e.g., PVT corners)
15. 16
Each iteration can take up to several weeks for state-of-the-art designs and technologies
Chip design is
comprised of
iterative macro
and standard
cell placement
A complex and very
computationally
intensive part of
chip design
Outer loop
Hours-days
Macro placement Cell placement
Inner loop
No-overlap 2D:
𝑁! !
Designers manually select a macro
placement and use solvers to
optimize the standard cell
placement (inner loop) and then
manually iterate (outer loop)
16. 1.
Fit a
probabilistic
surrogate
model
3.
Evaluate
and
back to 1
2.
Minimize cheap
acquisition
function
Goal
Find minimum of
an expensive function
⋆: Data
Cost
True function
Surrogate
Surrogate
uncertainty
Search space
Exploration
of highly uncertain areas
Exploitation
of promising current
minimal areas
vs
(tradeoff)
Next
point
Bayesian optimization efficiently solves problems iteratively
How to apply to
macro placement?
Cost
Cost
Search space
Search space
17. 18
Bayesian optimization learns a surrogate function, which maps each macro placement to a
PPA quality metric, and uses it to narrow down the search over the large macro placement space
Bayesian
optimization
for macro
layout
Inner loop optimization
incorporated into
surrogate function
Outer loop
Hours-days
Macro placement Cell placement
Inner loop
(50!)2 ≈ 10128
PPA
Surrogate function
“Bayesian Optimization for Macro Placement”, ICML 2022
18. 19
Simulated annealing
Bayesian optimization
Number of evaluations Number of evaluations Number of evaluations
Wire
length
Results are for public MCNC benchmark for layouts (without standard cells)
Three different chip designs: hp, ami33, ami49
Optimization objective is to minimize HPWL (wire length)
This is a simpler objective than the Inner Loop PPA
Further research aimed at generalizing this technique for production designs, across
all PPA metrics and with the inclusion of design constraints
Bayesian
optimization can
converge faster
with better design metrics
compared to conventional
simulated annealing heuristics
for an unconstrained version
of the problem
19
“Bayesian Optimization for Macro Placement”, ICML 2022
19. 20
Slow, costly,
but accurate
for sign-off and
automation
Algorithmic optimization
based on analytic
solutions needs
human guidance
Fast, cheap,
and
accurate for
optimization
Data-driven AI optimization
can aid designers with fast
evaluations and guide
algorithmic optimization
with optimal inputs
20. AI can improve core components of compilers, including
sequencing, scheduling, tiling, and placement
AI compilers can be
optimized with AI
“Neural topological ordering for computation graphs”, NeurIPS 2022
21. 22
Qualcomm AI Stack, Qualcomm Neural Processing SDK ,and Qualcomm AI Engine Direct are products of Qualcomm Technologies, Inc. and/or its subsidiaries. AIMET Model Zoo is a product of Qualcomm Innovation Center, Inc.
Infrastructure:
Programming Languages
Virtual platforms
Core Libraries
Math Libraries
Profilers & Debuggers
Compilers
System Interface SoC, accelerator drivers
Qualcomm® AI Engine Direct
Emulation Support
AI Frameworks
Auto
XR Robotics
IoT
ACPC
Smartphones Cloud
Platforms
AI Runtimes
Qualcomm® Neural Processing SDK TF Lite
TF Lite Micro Direct ML
AIMET
AIMET
Model Zoo
NAS
Model
analyzers
Qualcomm AI
Model Studio
Tools:
22. 23
Deployment
Compiler
Tiling
Sequencing
Scheduling
…
Placement
Tiling and placement
Splits net blocks into efficient
code Ops and places them
on multiple compute devices
Sequencing
Determines the best compute
ordering of the nodes
Scheduling
Parallelizes across compute
engines and sets final timing
Deployment
Puts the resulting
generated code onto
the target hardware
Our example here will focus
on the sequencing problem
The AI compiler converts an input neural net into an
efficient program that runs on target hardware
23
Optimizing for latency, performance, and power
Neural net Sequencing
23. 24
The runtime
on the target
device is
expensive
to evaluate
• It can take minutes to
set up the compiler with
all your decisions and run
one computation graph
• Quite a large
decision space
• Goal: minimizing the
runtime of the pipeline
over the space of tiling,
sequencing, and
scheduling choices
• Metrics: Double Data Rate
(DDR) and Tightly Coupled
Memories (TCM)
24. 25
Minimizing
memory usage:
• Maximizes inferences/sec
by reducing waits for
external memory access
and by allowing for
compute parallelism
• Minimizes energy/inference
by reducing data movement
Sequencing has a big impact on memory usage
CONV_A_0 CONV_A_1
CONV_B_0 CONV_B_1
POOL_0 POOL_1
ACT_0 ACT_1
WT_B
WT_A
0
1
2
3
4
5
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Memory
Time
Option #1: Data pre-fetch
High memory utilization,
Peak > 4MB
Memory utilization
CONV_A_0 CONV_A_1
CONV_B_0 CONV_B_1
POOL_0 POOL_1
ACT_0 ACT_1
WT_B
WT_A
0
1
2
3
4
5
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Memory
Time
Memory utilization
Lowest memory utilization,
Peak = 3.25 MB
Option #2: Low memory utilization
25. 26
• # sequences is
# node permutations
that follow precedence
constraints
• Grows exponentially
in # nodes in
compute graph
“Neural Topological Ordering for Computation Graphs”, NeurIPS 2022
Input
A computation graph for a neural net
(up to 10k’s of nodes)
Path cost during execution
progression of improvements
Objective
Find a sequence that minimizes
the peak memory usage
Application
Reduce the inference time of AI models on chips
by minimizing access to external memory
External memory access can be 100x local memory
and compute times, and is energy-intensive
Computation graph (Ops)
Sequencing computation graphs to minimize memory usage
26
26. 27
End-to-end machine learning (ML) for sequencing
Formulation originally motivated by “Attention, learn to solve routing problems!”, ICLR 2019
Initial embeddings
Use node properties based
on graph structure as initial
node embeddings
27
Key components
Encoder
Use custom graph NN to capture
the graph topology in embeddings
while allowing for arbitrary graphs
Decoder
Generates a distribution
in the sequence space
𝑃! 𝑠𝑒𝑞|𝐺 and autoregressively
generates a sequence
Objective
min. 𝔼"#$~&!
𝐶𝑜𝑠𝑡 𝑠𝑒𝑞 ,
use RL to train encoder
decoder architecture
end-to-end
ML-based sequencer
Embedding Propagator Sequencer
Sequence
probabilities
Embed Encoder Decoder
Policy
Net
RL-trained
agent
Input
compute
graph
Search
sampling, beam
search, greedy
Ordering
27. 28
28
A combination of real and synthetic graphs used for training
We developed a novel approach to generating realistic synthetic graphs since real graph data is scarce
Real graphs
Synthetic graphs
We released the algorithm to make these
28. 29
BS: Beam Search; DP: Dynamic Programming; BFS: Breadth-First Sequence; DFS: Depth-First Sequence
Snapdragon is a product of Qualcomm Technologies, Inc. and/or its subsidiaries.
Memory
usage
gap
%
Real graphs test set (23 graphs)
Sequence generation time
Our model
generalizes well
and beats baselines
comprehensively
• Dataset of 115
representative graphs
• Size: few dozen
to 1k nodes
• Our model performs
better and is much faster
• Results on Snapdragon 8
29. 30
“Neural DAG Scheduling via One-Shot Priority Sampling”, NeurIPS 2022 Optimization Workshop, to appear at ICLR 2023
End-to-end ML
for scheduling
Input
A computation graph with
Op durations, set of compute devices
Output
Schedule: Define the start time
and device allocation for all nodes
Objective
Find a schedule that minimizes
the runtime (latency)
Applications
Reduce the inference time
of ML models on chips
Directed Acyclic Graph
(DAG)
duration
1.0
1.0
2.0
1.0
1
3
2
4
Final schedule
1 3
2
4
Maximize utilization of
parallel hardware while
maintaining sequence
30. 31
31
CP: Critical Path; SPT: Shortest Processing Time; MOPNR: Most Operations Remaining
Our end-to-end ML scheduler achieves state-of-the-art results
Can be optimized for different performance metrics, such as latency
Algorithm
Node graphs
200 – 500 500 – 700 700 – 1000
SpeedUp SpeedUp SpeedUp
Baseline
schedulers
CP 3.17 2.80 2.74
SPT 3.11 2.87 2.66
MOPNR 3.18 2.82 2.74
Our AI
scheduler
S(256) 3.28 3.20 2.86
End-to-end ML structure
similar to previous for
sequencing, train to
minimize latency
Results for a set of compute
graphs of different sizes
Achieves better speedup
(inversely proportional to
latency, higher is better)
SpeedUp – higher is better
31. 32
AI for full end-to-end compiler
Self-supervised learning for CO
Dynamical systems and AI
Combine efficiency of dynamic models with AI learning
Scale AI up to larger graphs Multi-core compiler
AI-augmentation of solvers
Combine best of solvers and AI methods
Robustify learning to distribution shift
Ensure good results for diverse inputs
Need to efficiently extract
global graph structure
Further develop AI approaches
targeting many distributed cores
Broad range
of research directions
for AI combinatorial
optimization
Learn efficient problem
representation to solve easier
Include all aspects including
compute primitives and tiling
32. 33
Improved combinatorial
optimization techniques
offer benefits for a variety of
use cases across industries
Qualcomm AI Research has
achieved state-of-the-art results
in combinatorial optimization for
chip design and AI compilers
We are enabling combinatorial
optimization technology at scale
to address challenging problems
Qualcomm AI Research is an initiative of Qualcomm Technologies, Inc.