In this paper we tackle the problem of providing Quality of Service
guarantees to virtualized applications, focusing on computing and networking
guarantees. We propose a mechanism for providing temporal isolation based on
a CPU real time scheduling strategy. This allows not only to have control over
the individual virtual machine throughput, but also on the activation latency and
response-time by which virtualized software components react to external events.
We show experimental results gathered on a real system validating the approach.
Improving Responsiveness for Virtualized Networking Under Intensive Computing...tcucinotta
In this paper the problem of providing network response guarantees to multiple Virtual Machines (VMs) co-scheduled on the same set of CPUs is tackled, where the VMs may have to host both responsive real-time applications and batch compute-intensive workloads. When trying to use a real-time reservation-based CPU scheduler for providing stable performance guarantees to such a VM, the compute-intensive workload would be scheduled better with high time granularities, to increase performance and reduce system overheads, whilst the real-time workload would need lower time granularities in order to keep the response-time under acceptable levels. The mechanism that is proposed in this paper mixes both concepts, allowing the scheduler to dynamically switch between fine-grain and coarse-grain scheduling intervals depending on whether the VM is performing network operations or not. A prototype implementation of the proposed mechanism has been realized for the KVM hypervisor when running on Linux, modifying a deadline-based real-time scheduling strategy for the Linux kernel developed previously. The gathered experimental results show that the proposed technique is effective in controlling the response-times of the real-time workload inside a VM while at the same time it allows for an efficient execution of the batch compute-intensive workload.
Monitoring a SUSE Linux Enterprise Environment with System Center Operations ...Novell
Learn the architecture and how you can monitor a SUSE Linux Enterprise environment using cross-platform extensions, which will be part of Microsoft System Center Operations Manager 2007 R2, and the new Novell Linux Management Pack. The management pack extends the default cross-platform capability of System Center Operations Manager 2007 R2 and provides monitoring and management capability of key services running on SUSE Linux Enterprise.
Improving Responsiveness for Virtualized Networking Under Intensive Computing...tcucinotta
In this paper the problem of providing network response guarantees to multiple Virtual Machines (VMs) co-scheduled on the same set of CPUs is tackled, where the VMs may have to host both responsive real-time applications and batch compute-intensive workloads. When trying to use a real-time reservation-based CPU scheduler for providing stable performance guarantees to such a VM, the compute-intensive workload would be scheduled better with high time granularities, to increase performance and reduce system overheads, whilst the real-time workload would need lower time granularities in order to keep the response-time under acceptable levels. The mechanism that is proposed in this paper mixes both concepts, allowing the scheduler to dynamically switch between fine-grain and coarse-grain scheduling intervals depending on whether the VM is performing network operations or not. A prototype implementation of the proposed mechanism has been realized for the KVM hypervisor when running on Linux, modifying a deadline-based real-time scheduling strategy for the Linux kernel developed previously. The gathered experimental results show that the proposed technique is effective in controlling the response-times of the real-time workload inside a VM while at the same time it allows for an efficient execution of the batch compute-intensive workload.
Monitoring a SUSE Linux Enterprise Environment with System Center Operations ...Novell
Learn the architecture and how you can monitor a SUSE Linux Enterprise environment using cross-platform extensions, which will be part of Microsoft System Center Operations Manager 2007 R2, and the new Novell Linux Management Pack. The management pack extends the default cross-platform capability of System Center Operations Manager 2007 R2 and provides monitoring and management capability of key services running on SUSE Linux Enterprise.
The Wizard of OS: a Heartbeat for Legacy Multimedia Applicationstcucinotta
This presentation describes a technique for automagic provisioning of real-time scheduling guarantees for legacy multimedia applications, that does not require any modification to the application source-code.
Virtualised e-Learning with Real-Time Guarantees on the IRMOS Platformtcucinotta
In this paper we focus on how Quality of Service guarantees are provided to virtualised applications in the Cloud Computing infrastructure that is being developed in the context of the IRMOS European Project. Provisioning of proper timeliness guarantees to distributed real-time applications involves the careful use of real-time scheduling mechanisms at the virtual-machine hypervisor level, of QoS-aware networking protocols and of proper design methodologies and tools for stochastic modelling of the application. The paper focuses on how we applied these techniques to a case-study involving a real eLearning mobile content delivery application that has been integrated into the IRMOS platform and its achieved performance.
Study: The Future of VR, AR and Self-Driving CarsLinkedIn
We asked LinkedIn members worldwide about their levels of interest in the latest wave of technology: whether they’re using wearables, and whether they intend to buy self-driving cars and VR headsets as they become available. We asked them too about their attitudes to technology and to the growing role of Artificial Intelligence (AI) in the devices that they use. The answers were fascinating – and in many cases, surprising.
This SlideShare explores the full results of this study, including detailed market-by-market breakdowns of intention levels for each technology – and how attitudes change with age, location and seniority level. If you’re marketing a tech brand – or planning to use VR and wearables to reach a professional audience – then these are insights you won’t want to miss.
Self-tuning Schedulers for Legacy Real-Time Applicationstcucinotta
We present an approach for adaptive scheduling of soft real-time legacy applications (for which no timing information is exposed to the system). Our strategy is based on the combination of two techniques: 1) a real-time monitor that observes the sequence of events generated by the application to infer its activation period, 2) a feedback mechanism
that adapts the scheduling parameters to ensure a timely execution of the application. By a thorough experimental evaluation of an implementation of our approach, we show its performance and its efficiency.
Virtual Network Functions as Real-Time Containers in Private Cloudstcucinotta
This paper presents preliminary results from our on-going research for ensuring stable performance of co-located distributed cloud services in a resource-efficient way. It is based on using a real-time CPU scheduling policy to achieve a fine-grain control of the temporal interferences among real-time services running in co-located containers. We present results obtained applying the method to a synthetic application running within LXC containers on Linux, where a modified kernel has been used that includes our real-time scheduling policy.
More information about the paper is available at:
http://retis.sssup.it/~tommaso/papers/cloud18.php
SLAs in Virtualized Cloud Computing Infrastructures with QoS Assurancetcucinotta
Cloud Computing is gaining momentum as one of the technologies that promises to subvert our own idea of computing. With an increasing usage of cloud applications and their consequent dependency from connectivity, the nowadays Personal Computer is becoming merely a mobile device acting as a front-end to on-line applications and services. This huge paradigm shift in computing is witnessed for example by big market players who announced the imminent launch of innovative products and Operating Systems (like Chrome notebooks and the accompanying Chrome OS2. by Google), which are capable of projecting the user into the network in a few seconds by booting and starting immediately a web browser and (mostly) nothing else. In such a challenging scenario, more and more of the applications that we traditionally used locally on our PC are being hosted on cloud infrastructures and operated remotely through the Internet. This includes not only batch tasks, but also interactive applications which need to operate inherently with good levels of responsiveness.
In this paper, the challenging problem is discussed of how to ensure predictable levels of Quality of Service (QoS) to cloud applications across the multiple layers of a typical cloud infrastructure, and how a reasonable Service Level Agreement (SLA) management and enforcement policy might look like. The scope of this paper represents a hands-on experience that was gained by the authors realising the IRMOS real-time cloud-computing infrastructure in the context of the IRMOS European Project
There are a set of new real-time scheduling algorithms being developed for the Linux kernel, which provide temporal isolation among tasks.
These include an implementation of the POSIX sporadic server (SS) and a deadline-based scheduler. These are based on the specification of
the scheduling guarantees needed by the kernel in terms of a budget and a period.
This presentation aims to tackle the issues related to how to design a proper kernel-space / user-space interface for accessing such new
functionality. For the SS, a POSIX compliant implementation would break binary compatibility. However, the currently implemented API seems to be lacking some important features, like a sufficient level of extensibility. This would be required for example for adding further parameters in the future, e.g., deadlines different from periods, or soft (i.e., work-conserving) reservations, or how to mix power management in the looop (if ever).
Modeling and simulation of power consumption and execution times for real-tim...tcucinotta
In this work, we introduce a power-consumption model for heterogeneous multicore architectures that captures the variability of energy consumption based on processing workload type, in addition to the classical variables considered in the literature, like type and frequency of the CPU.
We motivate the approach presenting experimental results gathered on a Odroid-XU3 board equipped with an Arm big.LITTLE CPU, showing that power consumption has a non-negligible dependency on the workload type. We also present a model to define the execution time of the tasks, which depends on both the workload, and the CPU frequency and architecture. We present our modifications to the open-source RTSIM real-time scheduling simulator to extend its CPU power consumption and execution time duration models, integrating results taken from the real platform.
The presented work constitutes a useful base for future research in power-aware real-time scheduling on heterogeneous platforms.
For more information about the paper, see:
http://retis.sssup.it/~tommaso/papers/ewili18-rtsim.php
Mpls conference 2016-data center virtualisation-11-marchAricent
Aricent’s presentation on “Micro VNFs and Micro service environment” on next generation Virtualized Network Functions (VNFs) is heating up. In debate on micro services, carriers has requested communities to step up research on micro service deployments.
Aricent believes that existing VNFs, which comes directly from the physical appliances software are not rightly designed and are less suited for cloud operations. These first generation VNFs are replication of physical appliances, monolithic architecture and need more computational power. These are heavy with physical appliance platform features i.e. HA, ISSU, Nonstop Routing/Switching and they have lots of redundant code which may not be necessary on cloud. As cloud platform provides these feature through its inherent platform capabilities.
A checkpointing mechanism for virtual clusters using memory- bound time-multi...IJECEIAES
Transparent hypervisor-level checkpoint-restart mechanisms for virtual clusters (VCs) or clusters of virtual machines (VMs) offer an attractive fault tolerance capability for cloud data centers. However, existing mechanisms have suffered from high checkpoint downtimes and overheads. This paper introduces Mekha, a novel hypervisor-level, in-memory coordinated checkpoint-restart mechanism for VCs that leverages precopy live migration. During a VC checkpoint event, Mekha creates a shadow VM for each VM and employs a novel memory-bound timed-multiplex data (MTD) transfer mechanism to replicate the state of each VM to its corresponding shadow VM. We also propose a global ending condition that enables the checkpoint coordinator to control the termination of the MTD algorithm for every VM in a VC, thereby reducing overall checkpoint latency. Furthermore, the checkpoint protocols of Mekha are designed based on barrier synchronizations and virtual time, ensuring the global consistency of checkpoints and utilizing existing data retransmission capabilities to handle message loss. We conducted several experiments to evaluate Mekha using a message passing interface (MPI) application from the NASA advanced supercomputing (NAS) parallel benchmark. The results demonstrate that Mekha significantly reduces checkpoint downtime compared to traditional checkpoint mechanisms. Consequently, Mekha effectively decreases checkpoint overheads while offering efficiency and practicality, making it a viable solution for cloud computing environments.
Optimum Scalability Point for Parallelisable Real-Time Componentstcucinotta
Distributing the workload of computationally intensive software components across a set of homogeneous computing resources (nodes, hosts, processors, cores), for the purpose of allowing them to meet precise timing (response-time) constraints, is often a pain due to the difficulties in understanding how the software will actually scale. Often, such a problem is faced by recurring to a trial-and-error process. In this paper, a methodology is introduced to tackle the problem of finding the optimum number of processors for deploying parallelisable real-time software components. Basic building blocks of the methodology are: a generic performance model for the response-time of a parallel software component; a concrete procedure for tuning optimally the parameters of the model; the application of optimisation techniques that allow to compute what is the minimum allocation needed to meet precise deadline constraints, as well as the one that minimises the response-time; the consideration of multiple real-time (independent) components to be deployed on the same multi/many-core hardware.
Tommaso Cucinotta - Low-latency and power-efficient audio applications on Linuxlinuxlab_conf
Building Linux-based low-latency audio processing software for nowadays multi-core devices can be cumbersome. I’ll present some of our on-going research on the topic at the Real-Time Systems Lab of Scuola Superiore Sant’Anna, focusing on sound synthesis on Android where power-efficiency is a must.
The talk will provide basic background information on how the audio sub-system of Linux works, in terms of interactions between the Linux kernel and the ALSA sound architecture, including how user-space applications normally cope with low-latency requirements, touching briefly on design concepts behind the existence of the JACK low-latency framework. Then, a few concepts will be provided on the peculiarities of the Android audio processing pipeline, crossing the concepts with the due complications arising from the world of mobile and power-efficient devices. Throughout the talk, I’ll touch upon concepts behind our research efforts on the topic, describing how properly designed real-time CPU scheduling strategies can make a difference in what is achievable in this area.
An Evaluation of Adaptive Partitioning of Real-Time Workloads on Linuxtcucinotta
This paper provides an open implementation and an experimental evaluation of an adaptive partitioning approach for scheduling real-time tasks on symmetric multicore systems. The proposed technique is based on combining partitioned EDF scheduling with an adaptive migration policy that moves tasks across processors only when strictly needed to respect their temporal constraints. The implementation of the technique within the Linux kernel, via modifications to the SCHED DEADLINE code base, is presented. An extensive experimentation has been conducted by applying the technique on a real multi-core platform with several randomly generated synthetic task sets. The obtained experimental results highlight that the approach exhibits a promising performance to schedule real-time workloads on a real system, with a greatly reduced number of migrations compared to the original global EDF available in SCHED DEADLINE.
More information about the paper is available at:
http://retis.sssup.it/~tommaso/papers/isorc21.php
High Performance Parallel Computing with Clouds and Cloud Technologiesjaliyae
Infrastructure services (Infrastructure-as-a-service), provided by cloud vendors, allow any user to provision a large number of compute instances fairly easily. Whether leased from public clouds or allocated from private clouds, utilizing these virtual resources to perform data/compute intensive analyses requires employing different parallel runtimes to implement such applications. Among many parallelizable problems, most “pleasingly parallel” applications can be performed using MapReduce technologies such as Hadoop, CGL-MapReduce, and Dryad, in a fairly easy manner. However, many scientific applications, which have complex communication patterns, still require low latency communication mechanisms and rich set of communication constructs offered by runtimes such as MPI. In this paper, we first discuss large scale data analysis using different MapReduce implementations and then, we present a performance analysis of high performance parallel applications on virtualized resources.
The Wizard of OS: a Heartbeat for Legacy Multimedia Applicationstcucinotta
This presentation describes a technique for automagic provisioning of real-time scheduling guarantees for legacy multimedia applications, that does not require any modification to the application source-code.
Virtualised e-Learning with Real-Time Guarantees on the IRMOS Platformtcucinotta
In this paper we focus on how Quality of Service guarantees are provided to virtualised applications in the Cloud Computing infrastructure that is being developed in the context of the IRMOS European Project. Provisioning of proper timeliness guarantees to distributed real-time applications involves the careful use of real-time scheduling mechanisms at the virtual-machine hypervisor level, of QoS-aware networking protocols and of proper design methodologies and tools for stochastic modelling of the application. The paper focuses on how we applied these techniques to a case-study involving a real eLearning mobile content delivery application that has been integrated into the IRMOS platform and its achieved performance.
Study: The Future of VR, AR and Self-Driving CarsLinkedIn
We asked LinkedIn members worldwide about their levels of interest in the latest wave of technology: whether they’re using wearables, and whether they intend to buy self-driving cars and VR headsets as they become available. We asked them too about their attitudes to technology and to the growing role of Artificial Intelligence (AI) in the devices that they use. The answers were fascinating – and in many cases, surprising.
This SlideShare explores the full results of this study, including detailed market-by-market breakdowns of intention levels for each technology – and how attitudes change with age, location and seniority level. If you’re marketing a tech brand – or planning to use VR and wearables to reach a professional audience – then these are insights you won’t want to miss.
Self-tuning Schedulers for Legacy Real-Time Applicationstcucinotta
We present an approach for adaptive scheduling of soft real-time legacy applications (for which no timing information is exposed to the system). Our strategy is based on the combination of two techniques: 1) a real-time monitor that observes the sequence of events generated by the application to infer its activation period, 2) a feedback mechanism
that adapts the scheduling parameters to ensure a timely execution of the application. By a thorough experimental evaluation of an implementation of our approach, we show its performance and its efficiency.
Virtual Network Functions as Real-Time Containers in Private Cloudstcucinotta
This paper presents preliminary results from our on-going research for ensuring stable performance of co-located distributed cloud services in a resource-efficient way. It is based on using a real-time CPU scheduling policy to achieve a fine-grain control of the temporal interferences among real-time services running in co-located containers. We present results obtained applying the method to a synthetic application running within LXC containers on Linux, where a modified kernel has been used that includes our real-time scheduling policy.
More information about the paper is available at:
http://retis.sssup.it/~tommaso/papers/cloud18.php
SLAs in Virtualized Cloud Computing Infrastructures with QoS Assurancetcucinotta
Cloud Computing is gaining momentum as one of the technologies that promises to subvert our own idea of computing. With an increasing usage of cloud applications and their consequent dependency from connectivity, the nowadays Personal Computer is becoming merely a mobile device acting as a front-end to on-line applications and services. This huge paradigm shift in computing is witnessed for example by big market players who announced the imminent launch of innovative products and Operating Systems (like Chrome notebooks and the accompanying Chrome OS2. by Google), which are capable of projecting the user into the network in a few seconds by booting and starting immediately a web browser and (mostly) nothing else. In such a challenging scenario, more and more of the applications that we traditionally used locally on our PC are being hosted on cloud infrastructures and operated remotely through the Internet. This includes not only batch tasks, but also interactive applications which need to operate inherently with good levels of responsiveness.
In this paper, the challenging problem is discussed of how to ensure predictable levels of Quality of Service (QoS) to cloud applications across the multiple layers of a typical cloud infrastructure, and how a reasonable Service Level Agreement (SLA) management and enforcement policy might look like. The scope of this paper represents a hands-on experience that was gained by the authors realising the IRMOS real-time cloud-computing infrastructure in the context of the IRMOS European Project
There are a set of new real-time scheduling algorithms being developed for the Linux kernel, which provide temporal isolation among tasks.
These include an implementation of the POSIX sporadic server (SS) and a deadline-based scheduler. These are based on the specification of
the scheduling guarantees needed by the kernel in terms of a budget and a period.
This presentation aims to tackle the issues related to how to design a proper kernel-space / user-space interface for accessing such new
functionality. For the SS, a POSIX compliant implementation would break binary compatibility. However, the currently implemented API seems to be lacking some important features, like a sufficient level of extensibility. This would be required for example for adding further parameters in the future, e.g., deadlines different from periods, or soft (i.e., work-conserving) reservations, or how to mix power management in the looop (if ever).
Modeling and simulation of power consumption and execution times for real-tim...tcucinotta
In this work, we introduce a power-consumption model for heterogeneous multicore architectures that captures the variability of energy consumption based on processing workload type, in addition to the classical variables considered in the literature, like type and frequency of the CPU.
We motivate the approach presenting experimental results gathered on a Odroid-XU3 board equipped with an Arm big.LITTLE CPU, showing that power consumption has a non-negligible dependency on the workload type. We also present a model to define the execution time of the tasks, which depends on both the workload, and the CPU frequency and architecture. We present our modifications to the open-source RTSIM real-time scheduling simulator to extend its CPU power consumption and execution time duration models, integrating results taken from the real platform.
The presented work constitutes a useful base for future research in power-aware real-time scheduling on heterogeneous platforms.
For more information about the paper, see:
http://retis.sssup.it/~tommaso/papers/ewili18-rtsim.php
Mpls conference 2016-data center virtualisation-11-marchAricent
Aricent’s presentation on “Micro VNFs and Micro service environment” on next generation Virtualized Network Functions (VNFs) is heating up. In debate on micro services, carriers has requested communities to step up research on micro service deployments.
Aricent believes that existing VNFs, which comes directly from the physical appliances software are not rightly designed and are less suited for cloud operations. These first generation VNFs are replication of physical appliances, monolithic architecture and need more computational power. These are heavy with physical appliance platform features i.e. HA, ISSU, Nonstop Routing/Switching and they have lots of redundant code which may not be necessary on cloud. As cloud platform provides these feature through its inherent platform capabilities.
A checkpointing mechanism for virtual clusters using memory- bound time-multi...IJECEIAES
Transparent hypervisor-level checkpoint-restart mechanisms for virtual clusters (VCs) or clusters of virtual machines (VMs) offer an attractive fault tolerance capability for cloud data centers. However, existing mechanisms have suffered from high checkpoint downtimes and overheads. This paper introduces Mekha, a novel hypervisor-level, in-memory coordinated checkpoint-restart mechanism for VCs that leverages precopy live migration. During a VC checkpoint event, Mekha creates a shadow VM for each VM and employs a novel memory-bound timed-multiplex data (MTD) transfer mechanism to replicate the state of each VM to its corresponding shadow VM. We also propose a global ending condition that enables the checkpoint coordinator to control the termination of the MTD algorithm for every VM in a VC, thereby reducing overall checkpoint latency. Furthermore, the checkpoint protocols of Mekha are designed based on barrier synchronizations and virtual time, ensuring the global consistency of checkpoints and utilizing existing data retransmission capabilities to handle message loss. We conducted several experiments to evaluate Mekha using a message passing interface (MPI) application from the NASA advanced supercomputing (NAS) parallel benchmark. The results demonstrate that Mekha significantly reduces checkpoint downtime compared to traditional checkpoint mechanisms. Consequently, Mekha effectively decreases checkpoint overheads while offering efficiency and practicality, making it a viable solution for cloud computing environments.
Optimum Scalability Point for Parallelisable Real-Time Componentstcucinotta
Distributing the workload of computationally intensive software components across a set of homogeneous computing resources (nodes, hosts, processors, cores), for the purpose of allowing them to meet precise timing (response-time) constraints, is often a pain due to the difficulties in understanding how the software will actually scale. Often, such a problem is faced by recurring to a trial-and-error process. In this paper, a methodology is introduced to tackle the problem of finding the optimum number of processors for deploying parallelisable real-time software components. Basic building blocks of the methodology are: a generic performance model for the response-time of a parallel software component; a concrete procedure for tuning optimally the parameters of the model; the application of optimisation techniques that allow to compute what is the minimum allocation needed to meet precise deadline constraints, as well as the one that minimises the response-time; the consideration of multiple real-time (independent) components to be deployed on the same multi/many-core hardware.
Tommaso Cucinotta - Low-latency and power-efficient audio applications on Linuxlinuxlab_conf
Building Linux-based low-latency audio processing software for nowadays multi-core devices can be cumbersome. I’ll present some of our on-going research on the topic at the Real-Time Systems Lab of Scuola Superiore Sant’Anna, focusing on sound synthesis on Android where power-efficiency is a must.
The talk will provide basic background information on how the audio sub-system of Linux works, in terms of interactions between the Linux kernel and the ALSA sound architecture, including how user-space applications normally cope with low-latency requirements, touching briefly on design concepts behind the existence of the JACK low-latency framework. Then, a few concepts will be provided on the peculiarities of the Android audio processing pipeline, crossing the concepts with the due complications arising from the world of mobile and power-efficient devices. Throughout the talk, I’ll touch upon concepts behind our research efforts on the topic, describing how properly designed real-time CPU scheduling strategies can make a difference in what is achievable in this area.
An Evaluation of Adaptive Partitioning of Real-Time Workloads on Linuxtcucinotta
This paper provides an open implementation and an experimental evaluation of an adaptive partitioning approach for scheduling real-time tasks on symmetric multicore systems. The proposed technique is based on combining partitioned EDF scheduling with an adaptive migration policy that moves tasks across processors only when strictly needed to respect their temporal constraints. The implementation of the technique within the Linux kernel, via modifications to the SCHED DEADLINE code base, is presented. An extensive experimentation has been conducted by applying the technique on a real multi-core platform with several randomly generated synthetic task sets. The obtained experimental results highlight that the approach exhibits a promising performance to schedule real-time workloads on a real system, with a greatly reduced number of migrations compared to the original global EDF available in SCHED DEADLINE.
More information about the paper is available at:
http://retis.sssup.it/~tommaso/papers/isorc21.php
High Performance Parallel Computing with Clouds and Cloud Technologiesjaliyae
Infrastructure services (Infrastructure-as-a-service), provided by cloud vendors, allow any user to provision a large number of compute instances fairly easily. Whether leased from public clouds or allocated from private clouds, utilizing these virtual resources to perform data/compute intensive analyses requires employing different parallel runtimes to implement such applications. Among many parallelizable problems, most “pleasingly parallel” applications can be performed using MapReduce technologies such as Hadoop, CGL-MapReduce, and Dryad, in a fairly easy manner. However, many scientific applications, which have complex communication patterns, still require low latency communication mechanisms and rich set of communication constructs offered by runtimes such as MPI. In this paper, we first discuss large scale data analysis using different MapReduce implementations and then, we present a performance analysis of high performance parallel applications on virtualized resources.
XPDS13: Enabling Fast, Dynamic Network Processing with ClickOS - Joao Martins...The Linux Foundation
While virtualization technologies like Xen have been around for a long time, it is only in recent years that they have started to be targeted as viable systems for implementing middlebox processing (e.g., firewalls, NATs). But can they provide this functionality while yielding the high performance expected from hardware-based middlebox offerings? In this talk Joao Martins will introduce ClickOS, a tiny, MiniOS-based virtual machine tailored for network processing. In addition to the vm itself, Joao Martins will describe performance improvements done to the entire Xen I/O pipe. Finally, Joao Martins will discuss an evaluation showing that ClickOS can be instantiated in 30 msecs, can process traffic at 10Gb/s for almost all packet sizes, introduces delay of 40 microseconds and can run middleboxes at rates of 5 Mp/s.
Welocme to ViralQR, your best QR code generator.ViralQR
Welcome to ViralQR, your best QR code generator available on the market!
At ViralQR, we design static and dynamic QR codes. Our mission is to make business operations easier and customer engagement more powerful through the use of QR technology. Be it a small-scale business or a huge enterprise, our easy-to-use platform provides multiple choices that can be tailored according to your company's branding and marketing strategies.
Our Vision
We are here to make the process of creating QR codes easy and smooth, thus enhancing customer interaction and making business more fluid. We very strongly believe in the ability of QR codes to change the world for businesses in their interaction with customers and are set on making that technology accessible and usable far and wide.
Our Achievements
Ever since its inception, we have successfully served many clients by offering QR codes in their marketing, service delivery, and collection of feedback across various industries. Our platform has been recognized for its ease of use and amazing features, which helped a business to make QR codes.
Our Services
At ViralQR, here is a comprehensive suite of services that caters to your very needs:
Static QR Codes: Create free static QR codes. These QR codes are able to store significant information such as URLs, vCards, plain text, emails and SMS, Wi-Fi credentials, and Bitcoin addresses.
Dynamic QR codes: These also have all the advanced features but are subscription-based. They can directly link to PDF files, images, micro-landing pages, social accounts, review forms, business pages, and applications. In addition, they can be branded with CTAs, frames, patterns, colors, and logos to enhance your branding.
Pricing and Packages
Additionally, there is a 14-day free offer to ViralQR, which is an exceptional opportunity for new users to take a feel of this platform. One can easily subscribe from there and experience the full dynamic of using QR codes. The subscription plans are not only meant for business; they are priced very flexibly so that literally every business could afford to benefit from our service.
Why choose us?
ViralQR will provide services for marketing, advertising, catering, retail, and the like. The QR codes can be posted on fliers, packaging, merchandise, and banners, as well as to substitute for cash and cards in a restaurant or coffee shop. With QR codes integrated into your business, improve customer engagement and streamline operations.
Comprehensive Analytics
Subscribers of ViralQR receive detailed analytics and tracking tools in light of having a view of the core values of QR code performance. Our analytics dashboard shows aggregate views and unique views, as well as detailed information about each impression, including time, device, browser, and estimated location by city and country.
So, thank you for choosing ViralQR; we have an offer of nothing but the best in terms of QR code services to meet business diversity!
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
3. Introduction
“Traditional” worlds of computing
General-Purpose Computing (GPC) and servers
Low-cost, low parallelism degree equipment
Focused on mixing batch and interactive workloads
High-Performance Computing (HPC)
High-cost, massively parallel and vector-based equipment
Focused on batch computing, tightly coupled parallel tasks,
scientific applications
New trends
Affordable many-core systems also for GPC
Cloud-Computing: world-level scalability and replicability
Use of high-performance hardware in CC applications
Interest in mixing interactive/real-time and HPC workloads
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 3/24
4. Introduction
Virtualization is a key technology
For IaaS providers (Cloud Computing)
For server consolidation
Physical Host
Physical Host
OS
OS
...
Physical Host
Physical Host
OS
OS
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 4/24
5. Introduction
Virtualization is a key technology
For IaaS providers (Cloud Computing)
For server consolidation
Physical Host
Physical Host Physical Host
Physical Host
OS
OS VM/OS
VM/OS
... ...
Physical Host VM/OS
VM/OS
Physical Host
OS
OS
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 5/24
6. Need for Performance Isolation
Resource sharing
→ Temporal interference
Physical Host
Physical Host
VM
VM
VM
VM
VM
VM
VM
VM
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 6/24
7. Need for Performance Isolation
~30ms
Resource sharing VM Alone
→ Temporal interference
Physical Host
Physical Host
VM
VM
VM
VM
VM
VM
VM
VM
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 7/24
8. Need for Performance Isolation
~30ms
Resource sharing VM Alone
→ Temporal interference
Physical Host
Physical Host
VM
VM
VM
VM
~120ms
2 VMs
VM
VM
VM
VM
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 8/24
9. Possible Solution
Hardware replication and Physical Host
Physical Host
static partitioning VM
VM
Computing
Multi-core (1 core per VM)
VM
VM
Networking
Multiple network adapters
(1 network adapter per VM) VM
VM
Multi-queue adapters
Drawbacks VM
VM
Limitation of flexibility
Under-utilization of resources
e.g., with real-time/interactive workloads
(cloud computing)
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 9/24
10. Possible Solution
Another approach
Let multiple VMs use the same resources
Use proper resource scheduling strategies
For example
Computing
Xen credit-based and SEDF schedulers
Networking
QoS-aware protocols (IntServ, MPLS)
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 10/24
11. What is still missing ?
Most approaches
Investigate on high-level load balancing techniques
Without the necessary care for fine-grain resource scheduling
Focus on fairness among multiple VMs
Only a few works
Focus on providing precise QoS guarantees to VMs
For example
Gupta et al., “[...] Performance Isolation […] in Xen”
Cherkasova et al., “Comparison of the 3 CPU Schedulers in Xen”
Our focus
Network performance isolation among concurrent VMs
Virtualization with a host OS
Specifically, on Linux as host OS with KVM
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 11/24
13. Proposed Approach
Use of real-time CPU scheduling
As basic mechanism for isolating VMs concurrently
running on the same CPU and core
IRMOS Real-Time Scheduler
For the Linux kernel
Provides hierarchical EDF/FP scheduling
EDF-based resource reservations
– (Q, P): a budget Q is granted every period P
– Both a guarantee and a limitation
FP scheduling within each EDF reservation
Provides temporal isolation
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 13/24
14. What can we achieve ?
(by CPU real-time scheduling)
CPU real-time scheduling achieves
Performance isolation of
compute-intensive VMs
Shown in other papers
What about network-intensive VMs ?
Problems
On the host OS (Linux)
network-intensive VMs impose a big interrupt workload
– difficult to quantify and keep under control
– steals CPU from reservations of other VMs
How can we fix this ?
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 14/24
15. Preliminary Solution
Preliminary solution (subject of this paper)
Temporal isolation of VMs by real-time scheduling
Scheduling parameters tuned according to benchmarked figures
– Reservation period set according to responsiveness requirements
– Budget needed for computations, plus
– Budget needed for sustaining network traffic
Budget over-provisioning
According to the foreseen interference due to
network-intensive VMs
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 15/24
17. Experimental Results
(Q6600 @ 2.4 GHz, 1-Gbit Card)
Achievable network throughput as a function of
the CPU share reserved to the VM
Measured via iperf: almost linear, as expected
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 17/24
18. Experimental Results
(2 VMs on the same core)
Throughput as a function of its own budget
13% drop when
own reservation is 35% (from ~300 Mbps to ~260 Mbps)
and reservation of other VM increased from 10% to 40%
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 18/24
19. Experimental Results
(2 VMs on the same core)
Throughput as a function of its own budget
~14% budget over-provisioning needed
for sustaining a ~300 Mbps throughput
when reservation of other VM increased from 10% to 40%
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 19/24
20. Experimental Results
(2 VMs on the same core)
Throughput as a function of the budget reserved
to the other interfering VM
Drop due to compute-intensive interfering VM lower than
the one due to network-intensive interfering VM
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 20/24
21. Experimental Results
(application-level benchmark)
Download time for a 100 KB file from Apache
Periodic download requests every 20ms
Response-times may be kept much more stable
by real-time scheduling
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 21/24
22. Conclusions and Future Work
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 22/24
23. Conclusions and Future Work
Conclusions
We showed how to achieve network
performance isolation by CPU
real-time scheduling
Technique to be used jointly with
traffic-shaping techniques
Planned Future Work
Experiment with PREEMPT_RT
IRQs handled in kernel threads
Investigate on task-level isolation of network traffic of
different VMs (via real-time scheduling)
Make KVM a QoS-aware hypervisor
Investigate on the use of Adaptive Reservations
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 23/24
24. Thanks for your attention!
Questions ?
Tommaso Cucinotta – ReTiS Lab – Scuola Superiore Sant'Anna – Pisa – Italy 24/24