This presentation entitled "Seamless FPGA deployment over Spark in cloud computing: A use case on machine learning hardware acceleration" was conducted in the 14th International Symposium, ARC 2018, Santorini, Greece, May 2-4, 2018.
This document discusses accelerating machine learning applications over Spark on Pynq. It describes using FPGAs to accelerate logistic regression, a popular machine learning technique. The authors developed SPynq, which integrates Spark with Pynq to run Spark applications on Zynq systems. Their implementation achieves up to 21x higher energy efficiency compared to software-only solutions. Evaluation shows the FPGA accelerator provides significant speedups over CPUs and ARM for logistic regression. The work is part of the VINEYARD project, which aims to build an integrated platform for energy-efficient data centers using programmable hardware accelerators.
Presentation made at the 14th International Symposium "Applied Reconfigurable Computing" (ARC 2018) in Santorini, Greece, on May 2-4, 2018, about a novel scheme for the acceleration of Alternating Least Squares-based (ALS) collaborative ltering for recommendation engines that can be used to speedup signicantly the processing time and also reduce the energy consumption of computing platforms.
HKG18-300K2 - Keynote: Tomas Evensen - All Programmable SoCs? – Platforms to ...Linaro
1) All Programmable SoCs provide platforms to enable future embedded machine learning at the edge by accelerating workloads and providing higher performance with lower latency than CPUs alone.
2) FPGAs act as "chameleon chips" that can be configured for different applications like parallel compute, flexible dataflows, and memory hierarchies through programmable logic and hardened blocks.
3) Xilinx is partnering with Avnet to release the Ultra96 board equipped with a Zynq UltraScale+ MPSoC for $249, targeting developers for machine learning, IoT, and other applications through an open software stack.
HKG18-301 - Dramatically Accelerate 96Board Software via an FPGA with Integra...Linaro
Session ID: HKG18-301
Session Name: HKG18-301 - Dramatically Accelerate 96Board Software via an FPGA with Integrated Processors
Speaker: Glenn Steiner
Track: LITE
★ Session Summary ★
Key Takeaways:
With the drive to increase integration, reduce system costs, accelerate performance, and enhance reliability, software developers are discovering the processor they would like to target is simply not fast enough. This session will help you the system architect or software developer understand how you can architect and develop software on an FPGA integrated processor, and accelerate software code via FPGA accelerators.
Abstract:
As a software developer, in order to meet system level performance requirements, you may have realized that your next software project will be targeting a processor inside of an FPGA. How will this impact your development process and what benefits might you gain with this tight integration of processor and FPGA? Starting from the basics of what FPGAs are (in terms of software programming), this session will provide a simple to understand primer of what modern FPGAs with embedded processors can do. We will wrap up with examples of how high level synthesis tools can move software to programmable logic hardware enabling dramatic software acceleration.
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/hkg18/hkg18-301/
Presentation: http://connect.linaro.org.s3.amazonaws.com/hkg18/presentations/hkg18-301.pdf
Video: http://connect.linaro.org.s3.amazonaws.com/hkg18/videos/hkg18-301.mp4
---------------------------------------------------
★ Event Details ★
Linaro Connect Hong Kong 2018 (HKG18)
19-23 March 2018
Regal Airport Hotel Hong Kong
---------------------------------------------------
Keyword: LITE
'http://www.linaro.org'
'http://connect.linaro.org'
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
This document announces an AI workshop on August 31, 2018 in Warsaw, Poland hosted by the University of Warsaw and OpenPOWER Academia. The workshop will provide an introduction to artificial intelligence using POWER9 systems, including demonstrations of deep learning tools and techniques. Attendees will learn about OpenPOWER/POWER9 systems, PowerAI tools, and have hands-on exercises for using AI on OpenPOWER systems.
The document discusses how DDN A3I storage solutions and Nvidia's SuperPOD platform can enable HPC at scale. It provides details on DDN's A3I appliances that are optimized for AI and deep learning workloads and validated for Nvidia's DGX-2 SuperPOD reference architecture. The solutions are said to deliver the fastest performance, effortless scaling, reliability and flexibility for data-intensive workloads.
Integration of mixed-criticality subsystems on multicore and manycore processorsBabak Sorkhpour
This document summarizes an international workshop on integrating mixed-criticality subsystems on multicore and manycore processors. The workshop addressed several challenges in integrating mixed-criticality systems, including extra functional requirements, architectures, development tools, and certification. Several European projects were presented that are continuing research from previously completed projects in this area to address these challenges, with a focus on requirements, computer architectures, software support, and community building efforts.
Virtualization and hypervisor solutions for mixed-criticality systems based o...Babak Sorkhpour
This document discusses virtualization techniques for mixed-criticality systems using heterogeneous multicore processors. It begins with an introduction that outlines problems with complex embedded systems like certification and dependability. It then describes solutions like partitioning kernels and hypervisors that provide virtual execution environments and isolation. Examples of virtualization techniques are given, like the XtratuM hypervisor, that provide temporal and spatial isolation between partitions. Use cases like avionics and automotive systems are discussed where virtualization can help integrate applications with different criticality levels.
This document discusses accelerating machine learning applications over Spark on Pynq. It describes using FPGAs to accelerate logistic regression, a popular machine learning technique. The authors developed SPynq, which integrates Spark with Pynq to run Spark applications on Zynq systems. Their implementation achieves up to 21x higher energy efficiency compared to software-only solutions. Evaluation shows the FPGA accelerator provides significant speedups over CPUs and ARM for logistic regression. The work is part of the VINEYARD project, which aims to build an integrated platform for energy-efficient data centers using programmable hardware accelerators.
Presentation made at the 14th International Symposium "Applied Reconfigurable Computing" (ARC 2018) in Santorini, Greece, on May 2-4, 2018, about a novel scheme for the acceleration of Alternating Least Squares-based (ALS) collaborative ltering for recommendation engines that can be used to speedup signicantly the processing time and also reduce the energy consumption of computing platforms.
HKG18-300K2 - Keynote: Tomas Evensen - All Programmable SoCs? – Platforms to ...Linaro
1) All Programmable SoCs provide platforms to enable future embedded machine learning at the edge by accelerating workloads and providing higher performance with lower latency than CPUs alone.
2) FPGAs act as "chameleon chips" that can be configured for different applications like parallel compute, flexible dataflows, and memory hierarchies through programmable logic and hardened blocks.
3) Xilinx is partnering with Avnet to release the Ultra96 board equipped with a Zynq UltraScale+ MPSoC for $249, targeting developers for machine learning, IoT, and other applications through an open software stack.
HKG18-301 - Dramatically Accelerate 96Board Software via an FPGA with Integra...Linaro
Session ID: HKG18-301
Session Name: HKG18-301 - Dramatically Accelerate 96Board Software via an FPGA with Integrated Processors
Speaker: Glenn Steiner
Track: LITE
★ Session Summary ★
Key Takeaways:
With the drive to increase integration, reduce system costs, accelerate performance, and enhance reliability, software developers are discovering the processor they would like to target is simply not fast enough. This session will help you the system architect or software developer understand how you can architect and develop software on an FPGA integrated processor, and accelerate software code via FPGA accelerators.
Abstract:
As a software developer, in order to meet system level performance requirements, you may have realized that your next software project will be targeting a processor inside of an FPGA. How will this impact your development process and what benefits might you gain with this tight integration of processor and FPGA? Starting from the basics of what FPGAs are (in terms of software programming), this session will provide a simple to understand primer of what modern FPGAs with embedded processors can do. We will wrap up with examples of how high level synthesis tools can move software to programmable logic hardware enabling dramatic software acceleration.
---------------------------------------------------
★ Resources ★
Event Page: http://connect.linaro.org/resource/hkg18/hkg18-301/
Presentation: http://connect.linaro.org.s3.amazonaws.com/hkg18/presentations/hkg18-301.pdf
Video: http://connect.linaro.org.s3.amazonaws.com/hkg18/videos/hkg18-301.mp4
---------------------------------------------------
★ Event Details ★
Linaro Connect Hong Kong 2018 (HKG18)
19-23 March 2018
Regal Airport Hotel Hong Kong
---------------------------------------------------
Keyword: LITE
'http://www.linaro.org'
'http://connect.linaro.org'
---------------------------------------------------
Follow us on Social Media
https://www.facebook.com/LinaroOrg
https://www.youtube.com/user/linaroorg?sub_confirmation=1
https://www.linkedin.com/company/1026961
This document announces an AI workshop on August 31, 2018 in Warsaw, Poland hosted by the University of Warsaw and OpenPOWER Academia. The workshop will provide an introduction to artificial intelligence using POWER9 systems, including demonstrations of deep learning tools and techniques. Attendees will learn about OpenPOWER/POWER9 systems, PowerAI tools, and have hands-on exercises for using AI on OpenPOWER systems.
The document discusses how DDN A3I storage solutions and Nvidia's SuperPOD platform can enable HPC at scale. It provides details on DDN's A3I appliances that are optimized for AI and deep learning workloads and validated for Nvidia's DGX-2 SuperPOD reference architecture. The solutions are said to deliver the fastest performance, effortless scaling, reliability and flexibility for data-intensive workloads.
Integration of mixed-criticality subsystems on multicore and manycore processorsBabak Sorkhpour
This document summarizes an international workshop on integrating mixed-criticality subsystems on multicore and manycore processors. The workshop addressed several challenges in integrating mixed-criticality systems, including extra functional requirements, architectures, development tools, and certification. Several European projects were presented that are continuing research from previously completed projects in this area to address these challenges, with a focus on requirements, computer architectures, software support, and community building efforts.
Virtualization and hypervisor solutions for mixed-criticality systems based o...Babak Sorkhpour
This document discusses virtualization techniques for mixed-criticality systems using heterogeneous multicore processors. It begins with an introduction that outlines problems with complex embedded systems like certification and dependability. It then describes solutions like partitioning kernels and hypervisors that provide virtual execution environments and isolation. Examples of virtualization techniques are given, like the XtratuM hypervisor, that provide temporal and spatial isolation between partitions. Use cases like avionics and automotive systems are discussed where virtualization can help integrate applications with different criticality levels.
For the full video of this presentation, please visit:
https://www.edge-ai-vision.com/2020/12/vitis-and-vitis-ai-application-acceleration-from-cloud-to-edge-a-presentation-from-xilinx/
For more information about edge AI and computer vision, please visit:
https://www.edge-ai-vision.com
Vinod Kathail, Fellow and Chief Architect at Xilinx, presents the “Vitis and Vitis AI: Application Acceleration from Cloud to Edge” tutorial at the September 2020 Embedded Vision Summit.
Xilinx SoCs and FPGAs provide significant advantages in throughput, latency, and energy efficiency for production deployments of compute-intensive applications when compared to CPUs and GPUs. Over the last decade, FPGAs have evolved into highly configurable devices that provide on-chip heterogeneous multi-core CPUs, domain-specific programmable accelerators and “any-to-any” interface connectivity.
Today, the Xilinx Vitis Unified Software Platform supports high-level programming in C, C++, OpenCL, and Python, enabling developers to build and seamlessly deploy applications on Xilinx platforms including Alveo cards, FPGA instances in the cloud, and embedded devices. Moreover, Vitis enables the acceleration of large-scale data processing and machine learning applications using familiar high-level frameworks, such as TensorFlow and SPARK. This presentation provides an overview of the Vitis Software platform and the accelerated Vitis Vision Library, which enables customizable functions such as image signal processing, adaptable AI inference, 3D reconstruction and motion analysis.
This document discusses the advantages of using field-programmable gate arrays (FPGAs) over application-specific integrated circuits (ASICs) for software-defined networking (SDN). It notes that FPGAs have more logic gates, lower development costs, open development environments, and enable longer product lifecycles than ASICs. This makes FPGAs a more cost-effective solution for high-performance SDN applications like core routers, which typically use expensive ASICs. The total market cap of FPGA companies is also significantly lower than ASIC companies, representing an opportunity for SDN solutions based on FPGAs.
This work is about how both private enterprise and government wish to improve their data value and how they deal with this issue. The talk summarizes the way of thinking about Big Data, Open Data and their use by organizations or individuals. Big Data is explained from collecting, storing, analyzing and put in value. This data is collected from numerous sources including sensor networks, government data holdings, company market databases, and public profiles on social networking sites. Organizations use many data analytical techniques to study both structured and unstructured data. Due to the volume, velocity and variety of data, some specific techniques have been developed. MapReduce, Hadoop and other related as RHadoop are trending topic nowadays.
Data which come from government must be open. Every day more and more cities and countries are opening their data. Open Data is then presented as a specific case of public data with a special role in Smartcity. The main goal of Big and Open Data in Smartcity is to develop systems which can be useful for citizens. In this sense RMap (Mapa de Recursos) is shown as an Open Data application, an open system for Madrid City Council, avalaible for smarthphones and totally developed by the researching group G-TeC (www.tecnologiaUCM.es).
EPCC is a supercomputing centre at the University of Edinburgh that has been self-funded for over 28 years. It has over 110 staff and £5 million in annual turnover. EPCC supports multi-disciplinary research through access to its high performance computing facilities, training courses, and collaborative projects. It houses various supercomputing systems totaling over 150,000 CPU cores for researchers to use. EPCC also works with over 1000 companies through technology transfer and industrial collaborations in areas like simulation, data processing, and cloud computing. One example is its partnership with Rolls-Royce on a £15 million virtual gas turbine engine simulation project.
Data Center Interconnection for Optimum Service: FloraHolland safeguards world famous flower auctions and cuts costs by 20 percent with Cisco Nexus-enabled innovations.
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
In this deck from the Stanford HPC Conference, Nick Nystrom and Paola Buitrago provide an update from the Pittsburgh Supercomputing Center.
Nick Nystrom is Chief Scientist at the Pittsburgh Supercomputing Center (PSC). Nick is architect and PI for Bridges, PSC's flagship system that successfully pioneered the convergence of HPC, AI, and Big Data. He is also PI for the NIH Human Biomolecular Atlas Program’s HIVE Infrastructure Component and co-PI for projects that bring emerging AI technologies to research (Open Compass), apply machine learning to biomedical data for breast and lung cancer (Big Data for Better Health), and identify causal relationships in biomedical big data (the Center for Causal Discovery, an NIH Big Data to Knowledge Center of Excellence). His current research interests include hardware and software architecture, applications of machine learning to multimodal data (particularly for the life sciences) and to enhance simulation, and graph analytics.
Watch the video: https://youtu.be/LWEU1L1o7yY
Learn more: https://www.psc.edu/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
IoT Slam Keynote: Harnessing the Flood of Data with Heterogeneous Computing a...Ryft
This presentation was delivered as the closing keynote for the 2015 IoT Slam virtual conference. During the presentation, Ryft VP of Engineering, Pat McGarry, took a close look at how the IoT revolution is changing data analytics and driving the move of data analysis to the network’s edge where the data is being created. - See more at: http://www.ryft.com/blog/2015-iot-slam-keynote-harnessing-flood-of-iot-data-with-heterogenenous-computing-at-the-edge#sthash.x1Anoapb.dpuf
Introducing the Vitis Unified Software Platform for Programming FPGAsinside-BigData.com
Since their beginnings, FPGA's have been notorious for being hard to program. That could be changing with the new Vitis Unified Software Platform from Xilinx. Five years in the making, the Vitis unified software platform is designed to allow a whole new user base of software engineers and AI scientists to take advantage of the power of hardware adaptability.
"The Vitis unified software platform automatically tailors the Xilinx hardware architecture to the software or algorithmic code without the need for hardware expertise. Rather than imposing a proprietary development environment, the Vitis platform plugs into common software developer tools and utilizes a rich set of optimized open source libraries, enabling developers to focus on their algorithms. Vitis is separate to the Vivado Design Suite, which will still be supported for those who want to program using hardware code, but Vitis can also boost the productivity of hardware developers by packaging hardware modules as software-callable functions.
With exponentially increasing compute needs, engineers and scientists are often limited by the fixed nature of silicon,” said Victor Peng, president and chief executive officer, Xilinx. “Xilinx has created a singular environment that enables programmers and engineers from all disciplines to co-develop and optimize both their hardware and software, using the tools and frameworks they already know and understand. This means that they can adapt their hardware architecture to their application without the need for new silicon.”
Learn more: https://www.xilinx.com/products/design-tools/vitis.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Open Source Edge Computing Platforms - OverviewKrishna-Kumar
IEEE 11th International Conference - COMSNETS 2019 - Last MilesTalk - Jan 2019. This talk is for Beginner or intermediate levels only. Kubernetes and related edge platforms are discussed.
SAMOS 2018: LEGaTO: first steps towards energy-efficient toolset for heteroge...LEGATO project
The LEGaTO project received funding from the European Union's Horizon 2020 programme to create a software stack that optimizes for energy efficiency on heterogeneous computing platforms. The project aims to start with mature European software and optimize it to support energy-efficient computation on hardware with CPUs, GPUs, FPGAs, and FPGA-based dataflow engines. Key partners include universities and companies developing hardware and software. The project will develop programming models, runtime systems, and use cases in areas like healthcare, smart homes, and machine learning to demonstrate the stack.
This document provides an update on HPC-Midlands, a supercomputing service for research and industry. It discusses HPC-Midlands' expertise in high-performance computing, its ability to provide supercomputing capacity on demand via the JANET network, and details of its new 3,000 core supercomputer with 15TB of RAM and 120TB of storage. The document also outlines next steps, including addressing industrial users' requirements around software licensing, virtualization, and compatibility with their choice of operating systems and schedulers.
Presentation by Philippe O.A. Navaux, professor at the Universidade Federal of Rio Grande do Sul and Computer Science Area Director of CAPES at Cloudscape Brazil 2017 & WCN 2017
1) The Helix Nebula Science Cloud aims to be the leading European hybrid cloud platform for open innovation, open science, and open access for organizations and individual researchers.
2) Open Telekom Cloud will be a key partner and resource contributor to the Helix Nebula Science Cloud, providing its scalable, secure, and cost-effective public cloud services.
3) Research and development focuses of the Helix Nebula Science Cloud include integrating large-scale data access and handling functions in a unified and efficient manner across hybrid clouds, extending service level agreements, and improving identity and access management for privacy and security.
The document describes Expedient's Dublin data center, which has 3 phases totaling 60,000 square feet and 1725 cabinets. Phase 1 has 29,000 square feet and 575 cabinets. The data center has 12 generators providing 18 MW of total power capacity and 2070 tons of cooling. Expedient operates the data center 24/7/365 and provides benefits such as scalable infrastructure and a strategically located, risk-adverse region.
End-to-End Big Data AI with Analytics ZooJason Dai
The document discusses Analytics Zoo, an open-source software platform for building end-to-end big data AI applications. It provides distributed deep learning frameworks like TensorFlow and PyTorch on Apache Spark. Analytics Zoo allows seamless scaling of AI models from laptop to distributed big data and includes features like automated machine learning, time series forecasting, and serving models in production. It aims to simplify development of end-to-end big data AI solutions.
Supermicro AI Pod that’s Super Simple, Super Scalable, and Super AffordableRebekah Rodriguez
The worlds of HPC and AI are evolving at a tremendous rate. The demands of modern-day applications put immense pressure on local IT teams and resources. More often than not, this pressure can come from requiring an AI strategy to speed up mission-critical applications - but this can come at a cost which can hinder adoption. In this webinar, Supermicro, together with International Computer Concepts (ICC) and Define Tech, will demonstrate their AI Super Pod that delivers on AI strategy needs without breaking the bank.
Ovh analytics data compute with apache spark as a service meetup ovh bordeauxMojtaba Imani
90% of the data in the world today has been created in the last two years. The world will be creating 163 zettabytes of data a year by 2025. So how do we want to process this volume of data?
Apache Spark is an open-source distributed general-purpose cluster computing framework that is trending today. But the problem is that how to create a computing cluster fast and efficient? Should I do all network configuration and cluster management myself? What should I do with my cluster if I don't need it anymore? Is my cluster secure?
After discovering Apache Spark principles and use cases, you will discover OVH Analytics Data Compute. A fast, secure, and efficient Spark Cluster as a Service which is going to give answers to all these questions.
For the full video of this presentation, please visit:
https://www.edge-ai-vision.com/2020/12/vitis-and-vitis-ai-application-acceleration-from-cloud-to-edge-a-presentation-from-xilinx/
For more information about edge AI and computer vision, please visit:
https://www.edge-ai-vision.com
Vinod Kathail, Fellow and Chief Architect at Xilinx, presents the “Vitis and Vitis AI: Application Acceleration from Cloud to Edge” tutorial at the September 2020 Embedded Vision Summit.
Xilinx SoCs and FPGAs provide significant advantages in throughput, latency, and energy efficiency for production deployments of compute-intensive applications when compared to CPUs and GPUs. Over the last decade, FPGAs have evolved into highly configurable devices that provide on-chip heterogeneous multi-core CPUs, domain-specific programmable accelerators and “any-to-any” interface connectivity.
Today, the Xilinx Vitis Unified Software Platform supports high-level programming in C, C++, OpenCL, and Python, enabling developers to build and seamlessly deploy applications on Xilinx platforms including Alveo cards, FPGA instances in the cloud, and embedded devices. Moreover, Vitis enables the acceleration of large-scale data processing and machine learning applications using familiar high-level frameworks, such as TensorFlow and SPARK. This presentation provides an overview of the Vitis Software platform and the accelerated Vitis Vision Library, which enables customizable functions such as image signal processing, adaptable AI inference, 3D reconstruction and motion analysis.
This document discusses the advantages of using field-programmable gate arrays (FPGAs) over application-specific integrated circuits (ASICs) for software-defined networking (SDN). It notes that FPGAs have more logic gates, lower development costs, open development environments, and enable longer product lifecycles than ASICs. This makes FPGAs a more cost-effective solution for high-performance SDN applications like core routers, which typically use expensive ASICs. The total market cap of FPGA companies is also significantly lower than ASIC companies, representing an opportunity for SDN solutions based on FPGAs.
This work is about how both private enterprise and government wish to improve their data value and how they deal with this issue. The talk summarizes the way of thinking about Big Data, Open Data and their use by organizations or individuals. Big Data is explained from collecting, storing, analyzing and put in value. This data is collected from numerous sources including sensor networks, government data holdings, company market databases, and public profiles on social networking sites. Organizations use many data analytical techniques to study both structured and unstructured data. Due to the volume, velocity and variety of data, some specific techniques have been developed. MapReduce, Hadoop and other related as RHadoop are trending topic nowadays.
Data which come from government must be open. Every day more and more cities and countries are opening their data. Open Data is then presented as a specific case of public data with a special role in Smartcity. The main goal of Big and Open Data in Smartcity is to develop systems which can be useful for citizens. In this sense RMap (Mapa de Recursos) is shown as an Open Data application, an open system for Madrid City Council, avalaible for smarthphones and totally developed by the researching group G-TeC (www.tecnologiaUCM.es).
EPCC is a supercomputing centre at the University of Edinburgh that has been self-funded for over 28 years. It has over 110 staff and £5 million in annual turnover. EPCC supports multi-disciplinary research through access to its high performance computing facilities, training courses, and collaborative projects. It houses various supercomputing systems totaling over 150,000 CPU cores for researchers to use. EPCC also works with over 1000 companies through technology transfer and industrial collaborations in areas like simulation, data processing, and cloud computing. One example is its partnership with Rolls-Royce on a £15 million virtual gas turbine engine simulation project.
Data Center Interconnection for Optimum Service: FloraHolland safeguards world famous flower auctions and cuts costs by 20 percent with Cisco Nexus-enabled innovations.
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
In this deck from the Stanford HPC Conference, Nick Nystrom and Paola Buitrago provide an update from the Pittsburgh Supercomputing Center.
Nick Nystrom is Chief Scientist at the Pittsburgh Supercomputing Center (PSC). Nick is architect and PI for Bridges, PSC's flagship system that successfully pioneered the convergence of HPC, AI, and Big Data. He is also PI for the NIH Human Biomolecular Atlas Program’s HIVE Infrastructure Component and co-PI for projects that bring emerging AI technologies to research (Open Compass), apply machine learning to biomedical data for breast and lung cancer (Big Data for Better Health), and identify causal relationships in biomedical big data (the Center for Causal Discovery, an NIH Big Data to Knowledge Center of Excellence). His current research interests include hardware and software architecture, applications of machine learning to multimodal data (particularly for the life sciences) and to enhance simulation, and graph analytics.
Watch the video: https://youtu.be/LWEU1L1o7yY
Learn more: https://www.psc.edu/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
IoT Slam Keynote: Harnessing the Flood of Data with Heterogeneous Computing a...Ryft
This presentation was delivered as the closing keynote for the 2015 IoT Slam virtual conference. During the presentation, Ryft VP of Engineering, Pat McGarry, took a close look at how the IoT revolution is changing data analytics and driving the move of data analysis to the network’s edge where the data is being created. - See more at: http://www.ryft.com/blog/2015-iot-slam-keynote-harnessing-flood-of-iot-data-with-heterogenenous-computing-at-the-edge#sthash.x1Anoapb.dpuf
Introducing the Vitis Unified Software Platform for Programming FPGAsinside-BigData.com
Since their beginnings, FPGA's have been notorious for being hard to program. That could be changing with the new Vitis Unified Software Platform from Xilinx. Five years in the making, the Vitis unified software platform is designed to allow a whole new user base of software engineers and AI scientists to take advantage of the power of hardware adaptability.
"The Vitis unified software platform automatically tailors the Xilinx hardware architecture to the software or algorithmic code without the need for hardware expertise. Rather than imposing a proprietary development environment, the Vitis platform plugs into common software developer tools and utilizes a rich set of optimized open source libraries, enabling developers to focus on their algorithms. Vitis is separate to the Vivado Design Suite, which will still be supported for those who want to program using hardware code, but Vitis can also boost the productivity of hardware developers by packaging hardware modules as software-callable functions.
With exponentially increasing compute needs, engineers and scientists are often limited by the fixed nature of silicon,” said Victor Peng, president and chief executive officer, Xilinx. “Xilinx has created a singular environment that enables programmers and engineers from all disciplines to co-develop and optimize both their hardware and software, using the tools and frameworks they already know and understand. This means that they can adapt their hardware architecture to their application without the need for new silicon.”
Learn more: https://www.xilinx.com/products/design-tools/vitis.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Open Source Edge Computing Platforms - OverviewKrishna-Kumar
IEEE 11th International Conference - COMSNETS 2019 - Last MilesTalk - Jan 2019. This talk is for Beginner or intermediate levels only. Kubernetes and related edge platforms are discussed.
SAMOS 2018: LEGaTO: first steps towards energy-efficient toolset for heteroge...LEGATO project
The LEGaTO project received funding from the European Union's Horizon 2020 programme to create a software stack that optimizes for energy efficiency on heterogeneous computing platforms. The project aims to start with mature European software and optimize it to support energy-efficient computation on hardware with CPUs, GPUs, FPGAs, and FPGA-based dataflow engines. Key partners include universities and companies developing hardware and software. The project will develop programming models, runtime systems, and use cases in areas like healthcare, smart homes, and machine learning to demonstrate the stack.
This document provides an update on HPC-Midlands, a supercomputing service for research and industry. It discusses HPC-Midlands' expertise in high-performance computing, its ability to provide supercomputing capacity on demand via the JANET network, and details of its new 3,000 core supercomputer with 15TB of RAM and 120TB of storage. The document also outlines next steps, including addressing industrial users' requirements around software licensing, virtualization, and compatibility with their choice of operating systems and schedulers.
Presentation by Philippe O.A. Navaux, professor at the Universidade Federal of Rio Grande do Sul and Computer Science Area Director of CAPES at Cloudscape Brazil 2017 & WCN 2017
1) The Helix Nebula Science Cloud aims to be the leading European hybrid cloud platform for open innovation, open science, and open access for organizations and individual researchers.
2) Open Telekom Cloud will be a key partner and resource contributor to the Helix Nebula Science Cloud, providing its scalable, secure, and cost-effective public cloud services.
3) Research and development focuses of the Helix Nebula Science Cloud include integrating large-scale data access and handling functions in a unified and efficient manner across hybrid clouds, extending service level agreements, and improving identity and access management for privacy and security.
The document describes Expedient's Dublin data center, which has 3 phases totaling 60,000 square feet and 1725 cabinets. Phase 1 has 29,000 square feet and 575 cabinets. The data center has 12 generators providing 18 MW of total power capacity and 2070 tons of cooling. Expedient operates the data center 24/7/365 and provides benefits such as scalable infrastructure and a strategically located, risk-adverse region.
End-to-End Big Data AI with Analytics ZooJason Dai
The document discusses Analytics Zoo, an open-source software platform for building end-to-end big data AI applications. It provides distributed deep learning frameworks like TensorFlow and PyTorch on Apache Spark. Analytics Zoo allows seamless scaling of AI models from laptop to distributed big data and includes features like automated machine learning, time series forecasting, and serving models in production. It aims to simplify development of end-to-end big data AI solutions.
Supermicro AI Pod that’s Super Simple, Super Scalable, and Super AffordableRebekah Rodriguez
The worlds of HPC and AI are evolving at a tremendous rate. The demands of modern-day applications put immense pressure on local IT teams and resources. More often than not, this pressure can come from requiring an AI strategy to speed up mission-critical applications - but this can come at a cost which can hinder adoption. In this webinar, Supermicro, together with International Computer Concepts (ICC) and Define Tech, will demonstrate their AI Super Pod that delivers on AI strategy needs without breaking the bank.
Ovh analytics data compute with apache spark as a service meetup ovh bordeauxMojtaba Imani
90% of the data in the world today has been created in the last two years. The world will be creating 163 zettabytes of data a year by 2025. So how do we want to process this volume of data?
Apache Spark is an open-source distributed general-purpose cluster computing framework that is trending today. But the problem is that how to create a computing cluster fast and efficient? Should I do all network configuration and cluster management myself? What should I do with my cluster if I don't need it anymore? Is my cluster secure?
After discovering Apache Spark principles and use cases, you will discover OVH Analytics Data Compute. A fast, secure, and efficient Spark Cluster as a Service which is going to give answers to all these questions.
OVH Analytics Data Compute - Apache Spark Cluster as a ServiceOVHcloud
You need Apache Spark computation over a big Apache Spark cluster but you don't have computers ?
You don't have enough time to create a cluster of computers and do all installations and configurations ?
You just need a cluster for few hours and not forever ?
Or you just want to try out easily the power of Apache Spark ? Discover OVH Analytics Data Compute!
How to scale your PaaS with OVH infrastructure?OVHcloud
ForePaaS provides a platform for data infrastructure automation that allows customers to collect, store, transform and analyze data across multiple cloud providers or on-premise in a unified manner. Key features of the ForePaaS platform include being end-to-end, multi-cloud, providing a marketplace for sharing elements of work, and offering automated infrastructure that scales based on customer needs. ForePaaS has partnered with OVH to leverage their public cloud, private cloud, and bare metal server offerings to power ForePaaS infrastructure globally.
The Implementing AI: High Performance Architectures webinar, hosted by KTN and eFutures, was the fourth event in the Implementing AI summer webinar series.
Every business is increasing the use of artificial intelligence to gain efficiency and to make better decisions. These new demands for data processing are not well delivered by traditional computer architectures. Enterprises, developers, data scientists, and researchers need new platforms that unify all AI workloads, simplifying infrastructure and accelerating ROI. This has led to the development of high performance and specialised hardware devices to meet these new demands.
The focus of this webinar was the impact of processing AI data on data centres - particularly from the technology perspective. The webinar had four presentations from experts, covering the opportunities, implementation techniques and Case Studies, followed by a panel Q&A session.
FPGA Hardware Accelerator for Machine Learning
Machine learning publications and models are growing exponentially, outpacing Moore's law. Hardware acceleration using FPGAs, GPUs, and ASICs can provide performance gains over CPU-only implementations for machine learning workloads. FPGAs allow for reprogramming after manufacturing and can accelerate parts of machine learning algorithms through customized hardware while sharing computations between the FPGA and CPU. Vitis AI is a software stack that optimizes machine learning models for deployment on Xilinx FPGAs, providing pre-optimized models, tools for optimization and quantization, and high-level APIs.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/xilinx/embedded-vision-training/videos/pages/may-2019-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Nick Ni, Director of Product Marketing at Xilinx, presents the "Xilinx AI Engine: High Performance with Future-proof Architecture Adaptability" tutorial at the May 2019 Embedded Vision Summit.
AI inference demands orders- of-magnitude more compute capacity than what today’s SoCs offer. At the same time, neural network topologies are changing too quickly to be addressed by ASICs that take years to go from architecture to production. In this talk, Ni introduces the Xilinx AI Engine, which complements the dynamically- programmable FPGA fabric to enable ASIC-like performance via custom data flows and a flexible memory hierarchy. This combination provides an orders-of-magnitude boost in AI performance along with the hardware architecture flexibility needed to quickly adapt to rapidly evolving neural network topologies.
This document discusses NVIDIA's AI technologies and products. It highlights NVIDIA's A100 GPU which provides high performance for large datasets and models with 80GB of HBM2e memory and 2TB/s of bandwidth. It discusses challenges of scaling AI and how NVIDIA addresses this with their AI platform which includes pre-trained models, frameworks, and analytics/training as well as inference applications. The document discusses different AI scenarios and benchmarks. It outlines NVIDIA's offerings for enabling enterprise transformation with AI including application frameworks and edge, data center, and cloud solutions. Finally, it provides examples of how NVIDIA AI is being applied in domains like retail, supply chain, and distribution centers.
At the technology meeting of the Association of Independent Research Centers (http://airi.org): An overview of recent Scientific Computing activities at Fred Hutch, Seattle
Flexible and Scalable Integration in the Automation Industry/Industrial IoTconfluent
Speaker: Kai Waehner, Technology Evangelist, Confluent
Kafka-Native, End-to-End IIoT Data Integration and Processing with Kafka Connect, KSQL, and PLC4X
IIoT / Industry 4.0 with Apache Kafka, Connect, KSQL, Apache PLC4X Kai Wähner
Data integration and processing is a huge challenge in Industrial IoT (IIoT, aka Industry 4.0 or Automation Industry) due to monolithic systems and proprietary protocols. Apache Kafka, its ecosystem (Kafka Connect, KSQL) and Apache PLC4X are a great open source choice to implement this integration end to end in a scalable, reliable and flexible way.
This blog post covers a high level overview about the challenges and a good, flexible architecture. At the end, I share a video recording and the corresponding slide deck. These provide many more details and insights.
Apache Kafka is the De-facto Standard for Real-Time Event Streaming. It provides
Open Source (Apache 2.0 License)
Global-scale
Real-time
Persistent Storage
Stream Processing
PCL4X allows vertical integration and to write software independent of PLCs using JDBC-like adapters for various protocols like Siemens S7, Modbus, Allen Bradley, Beckhoff ADS, OPC-UA, Emerson, Profinet, BACnet, Ethernet.
Github example: https://github.com/kaiwaehner/iiot-integration-apache-plc4x-kafka-connect-ksql-opc-ua-modbus-siemens-s7
More details: http://www.kai-waehner.de/blog/2019/09/02/iiot-data-integr…and-apache-plc4x/
Video Recording: https://youtu.be/RWKggid25ds
Are you ready to be edgy? Bringing applications to the edge of the networkMegan O'Keefe
This document discusses edge computing and provides an overview of key concepts:
1. Edge computing extends cloud computing to the edge of networks to address bandwidth and latency issues and enable real-time applications like VR/AR, machine learning, and IoT.
2. Edge computing deployments are shaped differently than cloud but still leverage containers and orchestration tools like Kubernetes.
3. The document demonstrates an edge application management platform called Optikon that uses Kubernetes to deploy and manage applications across edge clusters.
The document discusses industrial Internet of Things (IIoT). It describes how ARM technology spans from sensors to servers, enabling embedded intelligence from low-power devices to infrastructure. IIoT allows for greater visibility, analytics capabilities, and coordination of industrial processes. Key takeaways are that IIoT is applications spanning the physical and cloud environments using IP connectivity to the edge, treating devices as web services, and following standards. IIoT drives operational efficiency through situational awareness, predictive maintenance, and other benefits.
Applied Robotics provides robotics solutions using Xilinx FPGAs. They were established in 1989 and have 10 employees with extensive experience. Their VCS systems use high-performance Xilinx Zynq chips and support computer vision, edge AI, and deep learning. They have developed smaller systems like the VCS-Jr and have done projects funded by Innovate UK, including autonomous mining robots. They promote the use of FPGAs for tasks like computer vision due to benefits like lower power consumption and latency compared to CPUs.
Building Real-time Pipelines with FLaNK_ A Case Study with Transit DataTimothy Spann
Building Real-time Pipelines with FLaNK_ A Case Study with Transit Data
Building Real-time Pipelines with FLaNK: A Case Study with Transit Data
In this session, we will explore the powerful combination of Apache Flink, Apache NiFi, and Apache Kafka for building real-time data processing pipelines. We will present a case study using the FLaNK-MTA project, which leverages these technologies to process and analyze real-time data from the New York City Metropolitan Transportation Authority (MTA). By integrating Flink, NiFi, and Kafka, FLaNK-MTA demonstrates how to efficiently collect, transform, and analyze high-volume data streams, enabling timely insights and decision-making.
Takeaways:
Understanding the integration of Apache Flink, Apache NiFi, and Apache Kafka for real-time data processing
Insights into building scalable and fault-tolerant data processing pipelines
Best practices for data collection, transformation, and analytics with FLaNK-MTA as a reference
Knowledge of use cases and potential business impact of real-time data processing pipelines
https://github.com/tspannhw/FLaNK-MTA/tree/main
https://medium.com/@tspann/finding-the-best-way-around-7491c76ca4cb
apache nifi
apache kafka
apache flink
apache iceberg
apache parquet
real-time streaming
tim spann
principal developer advocate
cloudera
datainmotion.dev
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2014-embedded-vision-summit-khronos
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Neil Trevett, President of Khronos and Vice President at NVIDIA, presents the "OpenVX Hardware Acceleration API for Embedded Vision Applications and Libraries" tutorial at the May 2014 Embedded Vision Summit.
This presentation introduces OpenVX, a new application programming interface (API) from the Khronos Group. OpenVX enables performance and power optimized vision algorithms for use cases such as face, body and gesture tracking, smart video surveillance, automatic driver assistance systems, object and scene reconstruction, augmented reality, visual inspection, robotics and more.
OpenVX enables significant implementation innovation while maintaining a consistent API for developers. OpenVX can be used directly by applications or to accelerate higher-level middleware with platform portability. OpenVX complements the popular OpenCV open source vision library that is often used for application prototyping.
"Bull has emerged as a significant provider of midrange to supercomputing scale HPC systems, building a base first in its native France before expanding in Europe, Latin America, and Asia. Leveraging its own in-house expertise in R&D, Bull has unveiled a full range of HPC products."
Similar to Seamless FPGA deployment over Spark in cloud computing: A use case on machine learning hardware acceleration (20)
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Introducing Milvus Lite: Easy-to-Install, Easy-to-Use vector database for you...Zilliz
Join us to introduce Milvus Lite, a vector database that can run on notebooks and laptops, share the same API with Milvus, and integrate with every popular GenAI framework. This webinar is perfect for developers seeking easy-to-use, well-integrated vector databases for their GenAI apps.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?
Seamless FPGA deployment over Spark in cloud computing: A use case on machine learning hardware acceleration
1. Christoforos Kachris, Elias Koromilas, Ioannis Stamelos, Dimitrios Soudris
kachris@microlab.ntua.gr
ICCS-National Technical University of Athens
ARC 2018, Santorini
Seamless FPGA deployment over Spark in cloud
computing: A use case on machine learning
hardware acceleration
3. www.vineyard-h2020.eu
Power consumption in the data centers
3
• Currently Data
Centers consume
huge amounts of
energy
• Servers consume
around 30% of the
total power budget of
the IT infrastructure
Christoforos Kachris, ICCS, Greece
5. www.vineyard-h2020.eu
FPGAs at the spotlight
5
April 2015
Submission of VINEYARD proposal
2015
2016
2017
December 2016
Overall, Intel now has five different AI
platforms; FPGAs, the Xeon Phi, the
Nervana NNP, the Myriad X, and its
traditional Core processor. The Core
processor still performs most AI tasks.
7. www.vineyard-h2020.eu
FPGAs in Data Center
• Intel: “Two orders of magnitude faster than GPU by 2020”
($16.7 billion bet)
Broadwel Xeon with Arria 10
• Microsoft Bing with Altera Stratix V
• IBM SupperVessel with Power8 + Xilinx
• Xilinx SDAccel on Nimbix Cloud
• Google has released TPU only for Tensorflow – ISCA 2017
7
Christoforos Kachris, ICCS, Greece
8. www.vineyard-h2020.eu
Machine learning market size
• The machine learning market
size is expected to grow from
USD 1.41 Billion in 2017 to
USD 8.81 Billion by 2022, at a
Compound Annual Growth
Rate (CAGR) of 44.1%.
https://www.marketsandmarkets.com/PressReleas
es/machine-learning.asp
Christoforos Kachris, ICCS, Greece 8
10. www.vineyard-h2020.eu
Apache Spark
The largest open source project in
data processing.
• Structured Data
• Streaming Analytics
• Machine Learning
• Graph Computation
Provides an interface for
programming entire clusters with
implicit data parallelism and fault-
tolerance.
10
Christoforos Kachris, ICCS, Greece
11. www.vineyard-h2020.eu
Contributions
• The FPGA driver API is packed in a shared object library and
can be used in a transparent way hiding all the low level
details.
• We implemented top level APIs in Python for standalone and
Apache Spark integrated use, that are easy to be used and are
also easily maintained since the middle layer, our shared
library remains the same for all of the above.
Christoforos Kachris, ICCS, Greece 11
12. www.vineyard-h2020.eu
System stack
• Application Layer: This layer
hosts users’ applications. The
applications can run natively
using Python.
• Vineyard Layer: This layer hosts
the whole functionality of our
framework. The key element of
this layer is the implemented
shared library
• SDSoC-HLS API and FPGA
layerhared library
Christoforos Kachris, ICCS, Greece 12
13. www.vineyard-h2020.eu
Flow for data movement - RDDS
• Flow of the original
and optimized
method for the
DMA transfers to
the accelerator
Christoforos Kachris, ICCS, Greece 13
16. www.vineyard-h2020.eu
Pynq: Python Productivity for Zynq
• An open-source project from Xilinx that
makes it easy to design embedded
systems with Zynq MPSoCs.
• The APSoC is programmed using
Python.
• The code is developed and tested
directly on the PYNQ-Z1 board.
• The programmable logic circuits are
imported as hardware libraries and
programmed through their APIs in
essentially the same way as the
software libraries.
16
Christoforos Kachris, ICCS, Greece
29. www.vineyard-h2020.eu
Main goals
VINEYARD AIMS TO
• Build an integrated platform for energy-efficient data
centres based on novel programmable hardware
accelerators
• Develop a high-level programming framework and big
data infrastructure for allowing end-users to seamlessly
utilize these accelerators in heterogeneous computing
systems by employing typical data-centre programming
frameworks (i.e. Spark.).
• VINEYARD also foster the establishment of an
ecosystem that will empower open innovation based on
hardware accelerators as data-centre plugins for
marketplace, thereby facilitating innovative enterprises
(large industries, SMEs, and creative start-ups) to
develop novel solutions using VINEYARDS’s leading
edge developments.
29
Christoforos Kachris, ICCS, Greece
30. • Speedup your application seamlessly
• An integrated framework for the utilization of hardware
accelerators in HPC and data center seamlessly
Contact detais: kachris@microlab.ntua.gr