Many emerging applications require methods tailored towards high-speed data acquisition and filtering of streaming data followed by offline event reconstruction and analysis. In this case, the main objective is to relieve the immense pressure on the storage and communication resources within the experimental infrastructure. In other applications, ultra low latency real time analysis is required for autonomous experimental systems and anomaly detection in acquired scientific data in the absence of any prior data model for unknown events. At these data rates, traditional computing approaches cannot carry out even cursory analyses in a time frame necessary to guide experimentation. In this talk, Prof. Ogrenci will present some examples of AI hardware architectures. She will discuss the concept of co-design, which makes the unique needs of an application domain transparent to the hardware design process and present examples from three applications: (1) An in-pixel AI chip built using the HLS methodology; (2) A radiation hardened ASIC chip for quantum systems; (3) An FPGA-based edge computing controller for real-time control of a High Energy Physics experiment.
"Computing systems for AI workloads have evolved towards data-center clusters of GPUs and TPUs, with architectures optimized for performing linear algebra and tunable for variable precision. As new AI paradigms emerge, more distinct divergence between hardware architectures for powering AI and other workloads are observed. GPU manufacturers are developing different architectures and chipsets for the HPC/supercomputing, cloud, edge computing, and robotics domains. FPGA vendors are also joining this ecosystem (e.g., Intel FPGAs deployed within Microsoft Azure). Moving forward, many industries and services ranging from cloud computing to consumer electronics are making hardware-accelerated AI a prominent component in their portfolio.
In this talk, some examples of AI hardware architectures and available silicon technologies will be presented. The concept of co-design will be discussed. This makes the unique needs of an application domain transparent to the hardware design process. Finally, an overview of design automation tool flows will be presented to gain an understanding of how to support a high productivity framework for domain experts to design and deploy AI hardware."
How to create innovative architecture using VisualSim?Deepak Shankar
In this presentation, we will get you started on using VisualSim Architect to conduct performance analysis, power measurement and functional validation. You will learn advanced concepts of system modeling and how to apply VisualSim Architect for a variety of applications.
Highlights include the application for both System-on-Chip and Large Systems including Designing memory interfaces using DDR3 and LPDDR3.
VisualSim Architect is used by systems and semiconductor companies to validate and analyze the specification of the product. The environment offers an easy-to-use methodology, huge library of technology components, extremely fast simulator and a huge reports list.
How to create innovative architecture using ViualSim?Deepak Shankar
In this presentation, we will get you started on using VisualSim Architect to conduct performance analysis, power measurement and functional validation. You will learn advanced concepts of system modeling and how to apply VisualSim Architect for a variety of applications.
Highlights include the application for both System-on-Chip and Large Systems including Designing memory interfaces using DDR3 and LPDDR3.
VisualSim Architect is used by systems and semiconductor companies to validate and analyze the specification of the product. The environment offers an easy-to-use methodology, huge library of technology components, extremely fast simulator and a huge reports list.
Please find our webinar video - How to create innovative architecture using ViualSim? at the last slide.
How to create innovative architecture using VisualSim?Deepak Shankar
In this presentation, we will get you started on using VisualSim Architect to conduct performance analysis, power measurement and functional validation. You will learn advanced concepts of system modeling and how to apply VisualSim Architect for a variety of applications.
Highlights include the application for both System-on-Chip and Large Systems including Designing memory interfaces using DDR3 and LPDDR3.
VisualSim Architect is used by systems and semiconductor companies to validate and analyze the specification of the product. The environment offers an easy-to-use methodology, huge library of technology components, extremely fast simulator and a huge reports list.
How HPC and large-scale data analytics are transforming experimental scienceinside-BigData.com
In this deck from DataTech19, Debbie Bard from NERSC presents: Supercomputing and the scientist: How HPC and large-scale data analytics are transforming experimental science.
"Debbie Bard leads the Data Science Engagement Group NERSC. NERSC is the mission supercomputing center for the USA Department of Energy, and supports over 7000 scientists and 700 projects with supercomputing needs. A native of the UK, her career spans research in particle physics, cosmology and computing on both sides of the Atlantic. She obtained her PhD at Edinburgh University, and has worked at Imperial College London as well as the Stanford Linear Accelerator Center (SLAC) in the USA, before joining the Data Department at NERSC, where she focuses on data-intensive computing and research, including supercomputing for experimental science and machine learning at scale."
Watch the video: https://wp.me/p3RLHQ-kLV
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
REAL-TIME SIMULATION TECHNOLOGIES FOR POWER SYSTEMS DESIGN, TESTING, AND ANAL...Jithin T
This is the ppt that contains effective elementsof the IEEE research journel "REAL-TIME SIMULATION TECHNOLOGIES FOR POWER
SYSTEMS DESIGN, TESTING, AND ANALYSIS"
"Computing systems for AI workloads have evolved towards data-center clusters of GPUs and TPUs, with architectures optimized for performing linear algebra and tunable for variable precision. As new AI paradigms emerge, more distinct divergence between hardware architectures for powering AI and other workloads are observed. GPU manufacturers are developing different architectures and chipsets for the HPC/supercomputing, cloud, edge computing, and robotics domains. FPGA vendors are also joining this ecosystem (e.g., Intel FPGAs deployed within Microsoft Azure). Moving forward, many industries and services ranging from cloud computing to consumer electronics are making hardware-accelerated AI a prominent component in their portfolio.
In this talk, some examples of AI hardware architectures and available silicon technologies will be presented. The concept of co-design will be discussed. This makes the unique needs of an application domain transparent to the hardware design process. Finally, an overview of design automation tool flows will be presented to gain an understanding of how to support a high productivity framework for domain experts to design and deploy AI hardware."
How to create innovative architecture using VisualSim?Deepak Shankar
In this presentation, we will get you started on using VisualSim Architect to conduct performance analysis, power measurement and functional validation. You will learn advanced concepts of system modeling and how to apply VisualSim Architect for a variety of applications.
Highlights include the application for both System-on-Chip and Large Systems including Designing memory interfaces using DDR3 and LPDDR3.
VisualSim Architect is used by systems and semiconductor companies to validate and analyze the specification of the product. The environment offers an easy-to-use methodology, huge library of technology components, extremely fast simulator and a huge reports list.
How to create innovative architecture using ViualSim?Deepak Shankar
In this presentation, we will get you started on using VisualSim Architect to conduct performance analysis, power measurement and functional validation. You will learn advanced concepts of system modeling and how to apply VisualSim Architect for a variety of applications.
Highlights include the application for both System-on-Chip and Large Systems including Designing memory interfaces using DDR3 and LPDDR3.
VisualSim Architect is used by systems and semiconductor companies to validate and analyze the specification of the product. The environment offers an easy-to-use methodology, huge library of technology components, extremely fast simulator and a huge reports list.
Please find our webinar video - How to create innovative architecture using ViualSim? at the last slide.
How to create innovative architecture using VisualSim?Deepak Shankar
In this presentation, we will get you started on using VisualSim Architect to conduct performance analysis, power measurement and functional validation. You will learn advanced concepts of system modeling and how to apply VisualSim Architect for a variety of applications.
Highlights include the application for both System-on-Chip and Large Systems including Designing memory interfaces using DDR3 and LPDDR3.
VisualSim Architect is used by systems and semiconductor companies to validate and analyze the specification of the product. The environment offers an easy-to-use methodology, huge library of technology components, extremely fast simulator and a huge reports list.
How HPC and large-scale data analytics are transforming experimental scienceinside-BigData.com
In this deck from DataTech19, Debbie Bard from NERSC presents: Supercomputing and the scientist: How HPC and large-scale data analytics are transforming experimental science.
"Debbie Bard leads the Data Science Engagement Group NERSC. NERSC is the mission supercomputing center for the USA Department of Energy, and supports over 7000 scientists and 700 projects with supercomputing needs. A native of the UK, her career spans research in particle physics, cosmology and computing on both sides of the Atlantic. She obtained her PhD at Edinburgh University, and has worked at Imperial College London as well as the Stanford Linear Accelerator Center (SLAC) in the USA, before joining the Data Department at NERSC, where she focuses on data-intensive computing and research, including supercomputing for experimental science and machine learning at scale."
Watch the video: https://wp.me/p3RLHQ-kLV
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
REAL-TIME SIMULATION TECHNOLOGIES FOR POWER SYSTEMS DESIGN, TESTING, AND ANAL...Jithin T
This is the ppt that contains effective elementsof the IEEE research journel "REAL-TIME SIMULATION TECHNOLOGIES FOR POWER
SYSTEMS DESIGN, TESTING, AND ANALYSIS"
A Dataflow Processing Chip for Training Deep Neural Networksinside-BigData.com
In this deck from the Hot Chips conference, Chris Nicol from Wave Computing presents: A Dataflow Processing Chip for Training Deep Neural Networks.
Watch the video: https://wp.me/p3RLHQ-k6W
Learn more: https://wavecomp.ai/
and
http://www.hotchips.org/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
"Digital Systems Design" undergraduate course slides, By Dr. Reza Sameni, presented at Shiraz University, Shiraz, IRAN, Revision: June 2018, WEB: www.sameni.info, EMAIL: reza.sameni@gmail.com
Mirabilis_Design AMD Versal System-Level IP LibraryDeepak Shankar
Mirabilis Design provides the VisualSim Versal Library that enable System Architect and Algorithm Designers to quickly map the signal processing algorithms onto the Versal FPGA and define the Fabric based on the performance. The Versal IP support all the heterogeneous resource.
Various processor architectures are described in this presentation. It could be useful for people working for h/w selection and processor identification.
State of the network
Plenty of bandwidth
Optical core
Increasing demand for services
Gateways
Network Service nodes
Content Switches
Network Caches
Evolving network
Get ready to dive into the exciting world of IoT data processing! 🌐📊
Join us for a thought-provoking webinar on "Processing: Turning IoT Data into Intelligence" hosted by industry visionary Deepak Shankar, founder of Mirabilis Design. Discover how to harness the potential of IoT devices by strategically choosing processors that optimize power, performance, and space.
In this engaging session, you'll explore key insights:
✅ Impact of processor architecture on Power-Performance-Area optimization
✅ Enabling AI and ML algorithms through precise compute and storage requirements
✅ Future trends in IoT hardware innovation
✅ Strategies for extending battery life and cost prediction through system design
Don't miss the chance to learn how to leverage a single IoT Edge processor for multiple applications and much more. This is your opportunity to gain a competitive edge in the evolving IoT landscape.
Architectural tricks to maximize memory bandwidthDeepak Shankar
Deepak Shankar, CEO and Founder of Mirabilis Deign Inc. hosted a webinar(Feb 17,2016) on the architectural possibilities to improve memory bandwidth. This webinar highlighted that memory plays a role in impacting the performance & power consumption of a system.
Webinar: Detecting Deadlocks in Electronic Systems using Time-based SimulationDeepak Shankar
Webinar: Detecting Deadlocks in Electronic Systems
Date: Nov 13th, 2019
Europe/ India Time: 11 AM CEST / 2:30 PM IST
US Time: 10 AM PT/ 1 PM ET
Register For the Webinar
Join Deepak Shankar, Founder of Mirabilis Design,
on Deadlock Detection of task graphs, using Discrete-Event Simulation.
on Thursday Nov 13th 2019
Europe/ India Time: 11 AM CEST / 2:30 PM IST
US Time: 10 AM PT/ 1 PM ET
Register For the Webinar
In Part One on Functional Analysis and Safety, we covered architecture modeling, fault injection, identification and resolution. View this Webinar, at the Mirabilis Design Video Channel. In Part Two, we focus on detecting deadlocks in systems that are time-variant. Traditional methods such as Ho-Ramamoorthy check for deadlocks in static directed graphs. In real systems, deadlocks occur from dependents missing deadlines, non-availability of resources from dependency and processing needs, multiple concurrent resource requests, criss-cross requests, stringent flow control, limited credit policies and buffer overflow. These require a dynamic, time-based simulation model to evaluate and detect deadlocks. In this Webinar, we use VisualSim Architect to assemble the task graph of the electronic; run use-cases and traffic through a time-based simulation; and evaluate the generated report to detect the source of the deadlocks.
During the webinar, you will learn to
1. Construct the system behavior using a system modeling environment
2. Run traffic and use-cases to create real-world operation
3. Evaluate the timing and resource consumption data to detect deadlocks
4. Determine the cause of the deadlocks using process and resource information
We will evaluate the simulated outcomes of an application to observe the functional coverage and design bottlenecks. Data Sampling with different test case are used to validate the correctness of the design. Example of deadlock scenarios are Multi-Core Cache Coherence, protocol and baseband Task Graphs, preemptive shared Bus and external resources such as printer, cameras and electrical drives.
"El álgebra lineal es una herramienta fundamental en muchos campos de la ciencia y la tecnología. Es particularmente importante en la física, la ingeniería, la informática y la estadística. La capacidad de manipular eficientemente grandes cantidades de datos y matrices complejas es esencial en estas áreas para la resolución de problemas y la toma de decisiones.
A priori, puede dar la sensación de que estamos muy lejos del uso del álgebra lineal en nuestro día a día. Sin embargo, algunas técnicas como la descomposición en valores singulares y la regresión lineal para entrenar modelos y hacer predicciones precisas están detrás de la inteligencia artificial y el aprendizaje automático. ¿Te suena ChatGPT? Puede no parecerlo, pero el álgebra lineal también está detrás en algunos de sus procesos. Por este motivo, debemos seguir trabajando en este campo, ya que su importancia seguirá creciendo a medida que se generen y analicen grandes cantidades de datos en el mundo actual.
"
La pandemia de COVID-19 ha supuesto una proliferación de mapas y contramapas. Por ello, organizaciones de la sociedad civil y movimientos sociales han generado sus propias interpretaciones y representaciones de los datos sobre la crisis. Estos también han contribuido a visibilizar aspectos, sujetos y temas que han sido desatendidos o infrarrepresentados en las visualizaciones hegemónicas y dominantes. En este contexto, la presente ponencia se centra en el análisis de los imaginarios sociales relacionados con la elaboración de mapas durante la pandemia. Es decir, trata de indagar en la importancia de los mapas para el activismo digital, las potencialidades que se extraen de esta tecnología y los valores asociados a las visualizaciones creadas con ellos. El objetivo último es reflexionar sobre la vía emergente del activismo de datos, así como sobre la intersección entre los imaginarios sociales y la geografía digital.
More Related Content
Similar to Design Automation Approaches for Real-Time Edge Computing for Science Applications
A Dataflow Processing Chip for Training Deep Neural Networksinside-BigData.com
In this deck from the Hot Chips conference, Chris Nicol from Wave Computing presents: A Dataflow Processing Chip for Training Deep Neural Networks.
Watch the video: https://wp.me/p3RLHQ-k6W
Learn more: https://wavecomp.ai/
and
http://www.hotchips.org/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
"Digital Systems Design" undergraduate course slides, By Dr. Reza Sameni, presented at Shiraz University, Shiraz, IRAN, Revision: June 2018, WEB: www.sameni.info, EMAIL: reza.sameni@gmail.com
Mirabilis_Design AMD Versal System-Level IP LibraryDeepak Shankar
Mirabilis Design provides the VisualSim Versal Library that enable System Architect and Algorithm Designers to quickly map the signal processing algorithms onto the Versal FPGA and define the Fabric based on the performance. The Versal IP support all the heterogeneous resource.
Various processor architectures are described in this presentation. It could be useful for people working for h/w selection and processor identification.
State of the network
Plenty of bandwidth
Optical core
Increasing demand for services
Gateways
Network Service nodes
Content Switches
Network Caches
Evolving network
Get ready to dive into the exciting world of IoT data processing! 🌐📊
Join us for a thought-provoking webinar on "Processing: Turning IoT Data into Intelligence" hosted by industry visionary Deepak Shankar, founder of Mirabilis Design. Discover how to harness the potential of IoT devices by strategically choosing processors that optimize power, performance, and space.
In this engaging session, you'll explore key insights:
✅ Impact of processor architecture on Power-Performance-Area optimization
✅ Enabling AI and ML algorithms through precise compute and storage requirements
✅ Future trends in IoT hardware innovation
✅ Strategies for extending battery life and cost prediction through system design
Don't miss the chance to learn how to leverage a single IoT Edge processor for multiple applications and much more. This is your opportunity to gain a competitive edge in the evolving IoT landscape.
Architectural tricks to maximize memory bandwidthDeepak Shankar
Deepak Shankar, CEO and Founder of Mirabilis Deign Inc. hosted a webinar(Feb 17,2016) on the architectural possibilities to improve memory bandwidth. This webinar highlighted that memory plays a role in impacting the performance & power consumption of a system.
Webinar: Detecting Deadlocks in Electronic Systems using Time-based SimulationDeepak Shankar
Webinar: Detecting Deadlocks in Electronic Systems
Date: Nov 13th, 2019
Europe/ India Time: 11 AM CEST / 2:30 PM IST
US Time: 10 AM PT/ 1 PM ET
Register For the Webinar
Join Deepak Shankar, Founder of Mirabilis Design,
on Deadlock Detection of task graphs, using Discrete-Event Simulation.
on Thursday Nov 13th 2019
Europe/ India Time: 11 AM CEST / 2:30 PM IST
US Time: 10 AM PT/ 1 PM ET
Register For the Webinar
In Part One on Functional Analysis and Safety, we covered architecture modeling, fault injection, identification and resolution. View this Webinar, at the Mirabilis Design Video Channel. In Part Two, we focus on detecting deadlocks in systems that are time-variant. Traditional methods such as Ho-Ramamoorthy check for deadlocks in static directed graphs. In real systems, deadlocks occur from dependents missing deadlines, non-availability of resources from dependency and processing needs, multiple concurrent resource requests, criss-cross requests, stringent flow control, limited credit policies and buffer overflow. These require a dynamic, time-based simulation model to evaluate and detect deadlocks. In this Webinar, we use VisualSim Architect to assemble the task graph of the electronic; run use-cases and traffic through a time-based simulation; and evaluate the generated report to detect the source of the deadlocks.
During the webinar, you will learn to
1. Construct the system behavior using a system modeling environment
2. Run traffic and use-cases to create real-world operation
3. Evaluate the timing and resource consumption data to detect deadlocks
4. Determine the cause of the deadlocks using process and resource information
We will evaluate the simulated outcomes of an application to observe the functional coverage and design bottlenecks. Data Sampling with different test case are used to validate the correctness of the design. Example of deadlock scenarios are Multi-Core Cache Coherence, protocol and baseband Task Graphs, preemptive shared Bus and external resources such as printer, cameras and electrical drives.
"El álgebra lineal es una herramienta fundamental en muchos campos de la ciencia y la tecnología. Es particularmente importante en la física, la ingeniería, la informática y la estadística. La capacidad de manipular eficientemente grandes cantidades de datos y matrices complejas es esencial en estas áreas para la resolución de problemas y la toma de decisiones.
A priori, puede dar la sensación de que estamos muy lejos del uso del álgebra lineal en nuestro día a día. Sin embargo, algunas técnicas como la descomposición en valores singulares y la regresión lineal para entrenar modelos y hacer predicciones precisas están detrás de la inteligencia artificial y el aprendizaje automático. ¿Te suena ChatGPT? Puede no parecerlo, pero el álgebra lineal también está detrás en algunos de sus procesos. Por este motivo, debemos seguir trabajando en este campo, ya que su importancia seguirá creciendo a medida que se generen y analicen grandes cantidades de datos en el mundo actual.
"
La pandemia de COVID-19 ha supuesto una proliferación de mapas y contramapas. Por ello, organizaciones de la sociedad civil y movimientos sociales han generado sus propias interpretaciones y representaciones de los datos sobre la crisis. Estos también han contribuido a visibilizar aspectos, sujetos y temas que han sido desatendidos o infrarrepresentados en las visualizaciones hegemónicas y dominantes. En este contexto, la presente ponencia se centra en el análisis de los imaginarios sociales relacionados con la elaboración de mapas durante la pandemia. Es decir, trata de indagar en la importancia de los mapas para el activismo digital, las potencialidades que se extraen de esta tecnología y los valores asociados a las visualizaciones creadas con ellos. El objetivo último es reflexionar sobre la vía emergente del activismo de datos, así como sobre la intersección entre los imaginarios sociales y la geografía digital.
Designing RISC-V-based Accelerators for next generation Computers (DRAC) is a 3-year project (2019-2022) funded by the ERDF Operational Program of Catalonia 2014-2020. DRAC will design, verify, implement and fabricate a high performance general purpose processor that will incorporate different accelerators based on the RISC-V technology, with specific applications in the field of post-quantum security, genomics and autonomous navigation. In this talk, we will provide an overview of the main achievements in the DRAC project, including the fabrication of Lagarto, the first RISC-V processor developed in Spain.
This talk will begin introducing the uElectronics section of ESA at ESTEC and the general activities the group is responsible for. Then, it will go through some of the R+D on-going activities that the group is involved with, hand in hand with universities and/or companies. One of the major ones is related to the European rad-hard FPGAs that have been partially founded by ESA for several years and that will be playing a major role in the sector in the upcoming years. It´s also worth talking about the RTL soft IPs that are currently under development and that will allow us to keep on providing the European ecosystem with some key capabilities. The latter will be an overview of RISC-V space hardened on-going activities that might be replacing the current SPARC based processors available for our missions.
El objetivo de esta charla es presentar las últimas novedades incorporadas en la arquitectura ARM y describir las tendencias en la microarquitectura de los procesadores con arquitectura ARM. ARM es una empresa relativamente pequeña en comparación con otros gigantes del sector tecnológico. Sin embargo, la amplia implantación de su arquitectura, siendo ampliamente dominante en algunos sectores, y sus microarquitecturas, hacen que la tecnología ARM ocupe un lugar central en el desarrollo tecnológico del mundo actual. La tecnología ARM está presente prácticamente en todo el espectro tecnológico, desde los dispositivos más sencillos hasta el HPC y Cloud computing, pasando por los smartphones, automoción electrónica de consumo, etc
"Formal verification has been used by computer scientists for decades to prevent
software bugs. However, with a few exceptions, it has not been used by researchers
working in most areas of mathematics (geometry, algebra, analysis, etc.). In this
talk, we will discuss how this has changed in the past few years, and the possible
implications to the future of mathematical research, teaching and communication.
We will focus on the theorem prover Lean and its mathematical library
mathlib, since this is currently the system most widely used by mathematicians.
Lean is a functional programming language and interactive theorem prover based
on dependent type theory, with proof irrelevance and non-cumulative universes.
The mathlib library, open-source and designed as a basis for research level
mathematics, is one of the largest collections of formalized mathematics. It allows
classical reasoning, uses large- and small-scale automation, and is characterized
by its decentralized nature with over 200 contributors, including both computer
scientists and mathematicians."
"Part of the research community thinks that it is still early to tackle the development of quantum software engineering techniques. The reason is that how the quantum computers of the future will look like is still unknown. However, there are some facts that we can affirm today: 1) quantum and classical computers will coexist, each dedicated to the tasks at which they are most efficient. 2) quantum computers will be part of the cloud infrastructure and will be accessible through the Internet. 3) complex software systems will be made up of smaller pieces that will collaborate with each other. 4) some of those pieces will be quantum, therefore the systems of the future will be hybrid. 5) the coexistence and interaction between the components of said hybrid systems will be supported by service composition: quantum services.
This talk analyzes the challenges that the integration of quantum services poses to Service Oriented Computing."
In this talk, after a brief overview of AI concepts in particular Machine Learning (ML) techniques, some of the well-known computer design concepts for high performance and power efficiency are presented. Subsequently, those techniques that have had a promising impact for computing ML algorithms are discussed. Deep learning has emerged as a game changer for many applications in various fields of engineering and medical sciences. Although the primary computation function is matrix vector multiplication, many competing efficient implementations of this primary function have been proposed and put into practice. This talk will review and compare some of those techniques that are used for ML computer design.
Tras una breve introducción a la informática médica y unas pinceladas sobre conceptos prácticos de Inteligencia Artificial (posible definición consensuada, strong VS weak AI y técnicas y métodos comúnmente empleados), el bloque central de la charla muestra ejemplos prácticos (en forma de casos de éxito) de distintos desarrollos llevados a cabo por el grupo de Sistemas Informáticos de Nueva Generación (SING: http//sing-group.org/) en los ámbitos de (i) Informática clínica (InNoCBR, PolyDeep), (ii) Informática para investigación clínica (PathJam, WhichGenes), (iii) bioinformática traslacional (Genómica: ALTER, Proteómica: DPD, BI, BS, Mlibrary, Mass-Up, e integración de datos ÓMICOS: PunDrugs) y (iv) Informática en salud pública (CURMIS4th). Finalmente, se comenta brevemente la importancia que se espera tenga en un futuro inmediato la IA interpretable (XAI, Explainable Artificial Intelligence) y la participación humana (HITL. Human-In-The-Loop). La charla termina con una breve reflexión sobre las lecciones aprendidas por el ponente después de más de 16 años de desarrollo de sistemas inteligentes en el ámbito de la informática médica.
En esta conferencia se presentará una revisión del concepto de autonomía para robots móviles de campo y la identificación de desafíos para lograr un verdadero sistema autónomo, además de sugerir posibles direcciones de investigación. Los sistemas robóticos inteligentes, por lo general, obtienen conocimiento de sus funciones y del entorno de trabajo en etapa de diseño y desarrollo. Este enfoque no siempre es eficiente, especialmente en entornos semiestructurados y complejos como puede ser el campo de cultivo. Un sistema robótico verdaderamente autónomo debería desarrollar habilidades que le permitan tener éxito en tales entornos sin la necesidad de tener a-priori un conocimiento ontológico del área de trabajo y la definición de un conjunto de tareas o comportamientos predefinidos. Por lo que en esta conferencia se presentarán posibles estrategias basadas en Inteligencia Artificial que permitan perfeccionar las capacidades de navegación de robots móviles y que sean capaces de ofrecer un nivel de autonomía lo suficientemente elevado para poder ejecutar todas las tareas dentro de una misión casa-a-casa (home-to-home).
Quantum computing has become a noteworthy topic in academia and industry. The multinational companies in the world have been obtaining impressive advances in all areas of quantum technology during the last two decades. These companies try to construct real quantum computers in order to exploit their theoretical preferences over today’s classical computers in practical applications. However, they are challenging to build a full-scale quantum computer because of their increased susceptibility to errors due to decoherence and other quantum noise. Therefore, quantum error correction (QEC) and fault-tolerance protocol will be essential for running quantum algorithms on large-scale quantum computers.
The overall effect of noise is modeled in terms of a set of Pauli operators and the identity acting on the physical qubits (bit flip, phase flip and a combination of bit and phase flips). In addition to Pauli errors, there is another error named leakage errors that occur when a qubit leaves the defined computational subspace. As the location of leakage errors is unknown, these can damage even more the quantum computations. Thus, this talk will briefly provide quantum error models.
Los chatbots son un elemento clave en la transformación digital de nuestra sociedad. Están por todas partes: eCommerce, salud digital, asistencia a clientes, turismo,... Pero si habéis usado alguno, probablemente os habrá decepcionado. Lo confieso, la mayoría de los chatbots que existen son muy malos. Y es que no es nada fácil hacer un chatbot que sea realmente útil e inteligente. Un chatbot combina toda la complejidad de la ingeniería de software con la del procesamiento de lenguaje natural. Pensad que muchos chatbots hay que desplegarlos en varios canales (web, telegram, slack,...) y a menudo tienen que utilizar APIs y servicios externos, acceder a bases de datos internas o integrar modelos de lenguaje preentrenados (por ej. detectores de toxicidad), etc. Y el problema no es sólo crear el bot, si no también probarlo y evolucionarlo. En esta charla veremos los mayores desafíos a los que hay que enfrentarse cuando nos encargan un proyecto de desarrollo que incluye un chatbot y qué técnicas y estrategias podemos ir aplicando en función de las necesidades del proyecto, para conseguir, esta vez sí un chatbot que sepa de lo que habla.
Many HPC applications are massively parallel and can benefit from the spatial parallelism offered by reconfigurable logic. While modern memory technologies can offer high bandwidth, designers must craft advanced communication and memory architectures for efficient data movement and on-chip storage. Addressing these challenges requires to combine compiler optimizations, high-level synthesis, and hardware design.
In this talk, I will present challenges, solutions, and trends for generating massively parallel accelerators on FPGA for high-performance computing. These architectures can provide performance comparable to software implementations on high-end processors, and much higher energy efficiency thanks to logic customization.
The main challenge of concurrent software verification has always been in achieving modularity, i.e., the ability to divide and conquer the correctness proofs with the goal of scaling the verification effort. Types are a formal method well-known for its ability to modularize programs, and in the case of dependent types, the ability to modularize and scale complex mathematical proofs.
In this talk I will present our recent work towards reconciling dependent types with shared memory concurrency, with the goal of achieving modular proofs for the latter. Applying the type-theoretic paradigm to concurrency has lead us to view separation logic as a type theory of state, and has motivated novel abstractions for expressing concurrency proofs based on the algebraic structure of a resource and on structure-preserving functions (i.e., morphisms) between resources.
Microarchitectural attacks, such as Spectre and Meltdown, are a class of
security threats that affect almost all modern processors. These attacks exploit the side-effects resulting from processor optimizations to leak sensitive information and compromise a system’s security.
Over the years, a large number of hardware and software mechanisms for
preventing microarchitectural leaks have been proposed. Intuitively, more
defensive mechanisms are less efficient, while more permissive mechanisms may offer more performance but require more defensive programming. Unfortunately, there are no
hardware-software contracts that would turn this intuition into a basis for
principled co-design.
In this talk, we present a framework for specifying hardware/software security
contracts, an abstraction that captures a processor’s security guarantees in a
simple, mechanism-independent manner by specifying which program executions a
microarchitectural attacker can distinguish.
La aparición de vulnerabilidades por la falta de controles de seguridad es una de las causas por las que se demandan nuevos marcos de trabajo que produzcan software seguro de forma predeterminada. En la conferencia se abordará cómo transformar el proceso de desarrollo de software dando la importancia que merece la seguridad desde el inicio del ciclo de vida. Para ello se propone un nuevo modelo de desarrollo – modelo Viewnext-UEx – que incorpora prácticas de seguridad de forma preventiva y sistemática en todas las fases del proceso de ciclo de vida del software. El propósito de este nuevo modelo es anticipar la detección de vulnerabilidades aplicando la seguridad desde las fases más tempranas, a la vez que se optimizan los procesos de construcción del software. Se exponen los resultados de un escenario preventivo, tras la aplicación del modelo Viewnext-UEx, frente al escenario reactivo tradicional de aplicar la seguridad a partir de la fase de testing.
This lecture will address issues of trust in computer systems, artificial intelligence and attacks on these types of systems with practical examples. Artificial Intelligence has gained ground in several areas with different applications scenarios, but in the perspective of this lecture, the fundamental point of the discussion is: what does an artificial intelligence system should do from a security perspective and how does an intelligence system provide results on a given subject? Few people are really concerned about the behavior of these types of systems from a security point of view. If you like machine learning and security, I believe this lecture will show you interesting security problems in artificial intelligence systems.
El uso de energías renovables es clave para cumplir los objetivos de desarrollo sostenible de la Agenda 2030. Entre estas energías, la eólica es la segunda más utilizada debido a su alta eficiencia. Algunos estudios sugieren que la energía eólica será la principal fuente de generación en 2050. Por ello es conveniente seguir investigando en la aplicación de técnicas de control avanzadas en estos sistemas.
Entre estas técnicas avanzadas cabe destacar las redes neuronales y el aprendizaje por refuerzo combinadas con estrategias clásicas de control. Estas técnicas ya se han empleado con éxito en el modelado y el control de sistemas complejos.
Esta conferencia presentará la aplicación de redes neuronales y aprendizaje por refuerzo al control de aerogeneradores, centrándolo especialmente en el control de pitch. Se detallarán diferentes configuraciones con redes neuronales y otras técnicas aplicadas al control de pitch. Finalmente se propondrán algunas técnicas híbridas que combinen lógica difusa, tablas de búsqueda y redes neuronales, mostrando resultados que han permitido probar su utilidad para mejorar la eficiencia de las turbinas eólicas.
As the world's energy demand rises, so does the amount of renewable energy, particularly wind energy, in the supply. The life cycle of wind farms starting from manufacturing the components to decommission stage involve significant involvement of cost and the application of AI and data analytics are on reducing these costs are limited. With this conference talk, the audience expected to know some of the interesting applications of AI and data analytics on offshore wind. And, also highlight the future challenges and opportunities. This conference could be useful for students, academics and researcher who want to make next career in offshore wind but yet know where to start.
The evolution of semiconductor process technologies has enabled the design of low-cost, compact and high-performance embedded systems, which have enabled the concept of Internet of Things (IoT). In addition, technological advances in communication protocols and unsupervised Machine Learning (ML) techniques are leading to disruptive innovations. As a result, the IoT, a new era of massive numbers of smart connected devices, can enhance processes and enable new services in established industries, creating smart cities, e-health businesses, or industry 4.0.
However, major challenges remain in achieving this potential due to the inherent complexity of designing energy-efficient IoT architectures. Prof. Atienza will first present the challenges of ultra-low power (ULP) design and communication overhead in next-generation IoT devices in the context of Big Data processing. Then, the benefits of exploiting the latest knowledge of how mammalian nervous systems acquire, process, and share information between the internal systems to conceive future edge AI-enabled architectures for IoT will be discussed
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
Forklift Classes Overview by Intella PartsIntella Parts
Discover the different forklift classes and their specific applications. Learn how to choose the right forklift for your needs to ensure safety, efficiency, and compliance in your operations.
For more technical information, visit our website https://intellaparts.com
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Quality defects in TMT Bars, Possible causes and Potential Solutions.PrashantGoswami42
Maintaining high-quality standards in the production of TMT bars is crucial for ensuring structural integrity in construction. Addressing common defects through careful monitoring, standardized processes, and advanced technology can significantly improve the quality of TMT bars. Continuous training and adherence to quality control measures will also play a pivotal role in minimizing these defects.
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdfKamal Acharya
The College Bus Management system is completely developed by Visual Basic .NET Version. The application is connect with most secured database language MS SQL Server. The application is develop by using best combination of front-end and back-end languages. The application is totally design like flat user interface. This flat user interface is more attractive user interface in 2017. The application is gives more important to the system functionality. The application is to manage the student’s details, driver’s details, bus details, bus route details, bus fees details and more. The application has only one unit for admin. The admin can manage the entire application. The admin can login into the application by using username and password of the admin. The application is develop for big and small colleges. It is more user friendly for non-computer person. Even they can easily learn how to manage the application within hours. The application is more secure by the admin. The system will give an effective output for the VB.Net and SQL Server given as input to the system. The compiled java program given as input to the system, after scanning the program will generate different reports. The application generates the report for users. The admin can view and download the report of the data. The application deliver the excel format reports. Because, excel formatted reports is very easy to understand the income and expense of the college bus. This application is mainly develop for windows operating system users. In 2017, 73% of people enterprises are using windows operating system. So the application will easily install for all the windows operating system users. The application-developed size is very low. The application consumes very low space in disk. Therefore, the user can allocate very minimum local disk space for this application.
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Event Management System Vb Net Project Report.pdfKamal Acharya
In present era, the scopes of information technology growing with a very fast .We do not see any are untouched from this industry. The scope of information technology has become wider includes: Business and industry. Household Business, Communication, Education, Entertainment, Science, Medicine, Engineering, Distance Learning, Weather Forecasting. Carrier Searching and so on.
My project named “Event Management System” is software that store and maintained all events coordinated in college. It also helpful to print related reports. My project will help to record the events coordinated by faculties with their Name, Event subject, date & details in an efficient & effective ways.
In my system we have to make a system by which a user can record all events coordinated by a particular faculty. In our proposed system some more featured are added which differs it from the existing system such as security.
Water scarcity is the lack of fresh water resources to meet the standard water demand. There are two type of water scarcity. One is physical. The other is economic water scarcity.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
3. Computer Engineering Research
@Northwestern
• Housed within the Electrical and Computer
Engineering Department, co-owned by the
Computer Science Department
– Faculty have dual appointments
– PhD programs in both departments have
Computer Engineering tracks
• Design Automation / CAD and VLSI,
Computer Architecture, Internet of Things,
Embedded and Cyberphysical Systems, Big
Data & AI
2
4. Computer Engineering Research
@Northwestern
• Hardware assisted ML
• Neuromorphic computing
• Hardware security
• Photonic circuits in computer architectures
• Emphatic computing
• Batteryless computing
• Secure and verified software for automotive
application
• Adaptive architecture – compiler in the loop
3
5. My Research Interests
• Electronic Design Automation (EDA)
• Circuits and Systems
• High Performance Computing (HPC)
Systems
4
6. Research Activities
• Design Automation
– High-Level Synthesis
• Converting applications described
with programming languages to
hardware
• Circuits and Systems
– Thermal and Power
Management of ICs
• Temperature Sensors, 3D
Integration, Power and
Performance Modeling using
Machine Learning
5
7. Research Activities
• High Performance
Computing Systems
– FPGA-based
accelerators for HPC
applications
– Thermal and Power
Management in large
scale systems
6
8. HW Assisted ML
• Design Automation for real-time AI
Cyberinfrastructure for scientists who deploy HW
systems for ML
7
Physics
Material
Science
Astro-
physics
Particle
Tracking
Image Reconstruction
Real-time control of
microscopy
Accelerator
Control
Semantic
compression
Data filtering, compression, reconstruction,
feedback controllers
9. Handling Big Data
• Source of the computational challenge
– Particle acceleration experiment in high-energy
physics
• Billions of “event” happen every ~25ns creating Pb/
sec data rates
• “Interesting” events (~3% of all) need to be
recognized and filtered for further processing
– Sky survey data collected from observatories
stream in at rates ~Gb/s (or more)
• Limited energy (generators in South Pole) to move
data to datacenters
8
10. Handling Big Data
• Source of the computational challenge
– CPU-GPU systems perform streaming inference on
images captured by an electron microscope within
~300ms latency
– Desired goal is to perform inference on images
captured at 10s of millions of frames per second
under 50ms latency
– ~50ms latency could allow real-time control of the
EM during material synthesis
• Think of making defects for a quantum material,
deposition of nano layers, etc.
9
11. Projects – Complete/Underway
• READS – Accelerator Real-time Edge AI
for Distributed Systems
• Design of a reconfigurable autoencoder
algorithm for detector front-end ASICs
10
13. READS – Accelerator Real-time
Edge AI for Distributed Systems
• Goal: integrate ML into accelerator operations
• Challenges: Beam Loss
– High Energy Physics (HEP) experiment use proton
beams
– Particles get lost through interactions with the beam
vacuum pipe
– Intensity/pace of beam extraction affects radiation in
the environment
– Human operators tune parameters of hundreds to
thousands of devices inside the accelerator complex
– If beam loss (“leak”) is at extreme levels, system
needs to shut-down – loss of access to facility
12
14. READS – Accelerator Real-time
Edge AI for Distributed Systems
13
• Muon Complex
• Two rings are on top of
each other: Main
Injector Ring and the
Recycler Ring
• Protons are accelerated
within the Booster
• Injected into the
Recycler Ring and Main
Injector Ring
18. READS – Accelerator Real-time
Edge AI for Distributed Systems
• System Design Goal: PID Controller gains
need to be optimized in real-time ~ms
• The ML Processor receives inputs from
sensors
– beam position monitor (BPM)
– beam loss monitor (BLM)
17
19. READS – Accelerator Real-time
Edge AI for Distributed Systems
• System Architecture:
18
20. READS – Accelerator Real-time
Edge AI for Distributed Systems
• System Architecture:
– Edge device continuously optimizes the online
control agent
– Data streamed to a cloud system for large
scale training
19
21. READS – Accelerator Real-time
Edge AI for Distributed Systems
• Algorithm-
Architecture Co-
design
– A common
toolchain to
program FPGA
devices as well
as create
interfaces
20
22. READS – Accelerator Real-time
Edge AI for Distributed Systems
• Algorithm-Architecture Co-design
– Create modular neural network components
• Regroup, recombine
– Establish physics/science-aware methodology
• Hardware-aware quantization and pruning techniques
• Homogeneous versus heterogeneous quantization
• Quantization-aware training
• Control resource re-use trade-offs of high level
synthesis tools
• Differentiate between the relative hardware cost of
storage, DSP, interconnect
21
24. Target - Arria10 HPS + FPGA (on-chip ram)
IP generated for the ML-module by hls4ml tool flow
hls_input (on-
chip ram)
hls_output
(on-chip
ram)
irq
(PIO)
hls4ml_ct
rl
(VHDL)
mm_mast
er
mm_mast
er
HPS
h2f_axi_master h2f_axi_master h2f_axi_master
h2f_irq1
start, done
signal etc
+ Testbench
of hls4ml
READS Project
25. Hardware Optimization for the ML-IP
– The current model implemented as IP consists of
• Dense Layer – ReLU – Dense Layer - Sigmoid
– 259 inputs and 518 outputs
• 16 bits ac_fixed values
– Representation still under investigation
– HLS (hls4ml) tool was optimized to explore resource
sharing for computation such as dense layer
– HLS tool was not equipped with features to optimize other
layers such as Sigmoid
• When we are dealing with a model with 100s of outputs situation
changes
– Logic Synthesis stage (Quartus)
• Large scale models introduce clock timing violations
that were not observed in smaller models
24
26. Project #2: Reconfigurable autoencoder for
detector front-end ASICs
• Edge computing for particle collider experiments
• Data collected from a large number of photon
detectors are compressed to representative
information of the “shape”
– Charge measurements from the detectors are
compressed to a radiation pattern
– 6 million detector channels sending data at 40MHz
– The data from the original space is compressed to
lower dimensionality by the edge ASIC, transmitted,
then decoded on the receiving end
25
27. Reconfigurable autoencoder for detector
front-end ASICs
• System constraints
– Low power
• Will be part of a larger system with power budget
– Radiation tolerant
– Reprogrammable weights through accessible
registers to enable updates and customization
26
28. Reconfigurable autoencoder for detector
front-end ASICs
• Autoencoder:
neural network
with single
convolutional
layer followed
by a dense
layer
27
31. Reconfigurable autoencoder for detector
front-end ASICs
• Network properties
– CNN layer: eight 3x3x3 kernel matrices resulting in 128
outputs
– ReLu activation after CNN and after final dense layer
– 6-bits weights
– Dense layer produces 16 10-bit outputs
– Chip can be reconfigured to produce as low as total of
64 bits in output
30
36. Reconfigurable autoencoder for detector
front-end ASICs
35
• Protection against Single Event Upsets
Partial TMR: suppresses error without
correction
Full TMR: both suppresses and corrects
error
37. Reconfigurable autoencoder for detector
front-end ASICs
36
• Partial TMR can also be implemented in
a few different hybrid modes
– Applied to only registers
– Include clock lines
– Apply to a sub-block
38. Reconfigurable autoencoder for detector
front-end ASICs
37
• Exploring trade-offs in TMR
– Area versus TMR coverage
– Leakage Power versus TMR coverage
– Fanout (routability) versus TMR coverage
– Timing (impact on slack)
40. Adaptive ML accelerators
• Why adaptation?
– Varying energy constraints
– Input variability
– Translation to new platform/device
– Transfer Learning
39
41. Adaptive ML accelerators
• Some directions
– Emulate loss through drop out/connect
• Techniques in ML literature equate these phenomena to
forcing weights to zero
• Loss in (because of) hardware may look very different
– Continuous diagnostics
• Evaluation of network certainty
• Concept of surprise
– Detection of temporal dominance of classes
• Reconfigure optimized version for dominant class
40
42. Adaptive ML accelerators
• Some directions
– Critical path based hardware-aware resource
management
• Class-based CP: using contributions of a neuron to a
specific class
– Mean Absolute Activation, first order Taylor Approximations
• Generalized CP: relative participation of output channels at
the routing of the output from a layer
– Distribute resources (e.g., total number of bits
allocated for weights) according to a criticality metric
41
43. Adaptive ML accelerators
• Some directions
– Characterize circuits (e.g., SRAM) to create models
• Associate voltage drop/power outage with cell decay
• Create characteristic bit masks for weights
– Neural network architecture search
• Lessons learned from design space exploration in
high-level synthesis
• Search for a new cell from basic building blocks
– Input: a set of convolutions and pooling of varying size
– Think of it as your module library
42
44. Adaptive ML accelerators
• Some directions
– Partition system into two part
• Part1: Fixed ML architecture
• Part2: ML architecture with capability to train within the edge
device
– Training consumes resources and energy on
the limited edge device
• Need to explore the trade-off carefully
• Co-design ML architecture and training hardware
43
45. Future Directions
• Fluid implementations
– Quickly re-targetable from software to hardware
• Scientific domains offer a vast space of
computational challenges
– They need interpretable systems
– Laws of nature apparent in the system’s output
• Multiple paths need to converge
– Photonic circuits – not much automated
– FPGA
– Neuromorphic – not much automated
– Quantum
44
46. Collaborators
• Northwestern University
– Han Liu (CS), Kristian Hahn (Physics)
– Manuel Blanco Valentin, Rui Shi, Deniz Ulusel,
Bincong Ye, Yingyi Luo (now at Google), Sid Joshi
(now at Intel)
• Fermi National Laboratories
– Nhan Tran, Farah Fahim, Christian Herwig, Kiyomi
Seiya, et al.
• Columbia University
– Giuseppe di Guglielmo
• Lehigh University
– Josh Agar
45