The demand for managing a large amount of data in a scalable yet reliable and cost-effective way has became more and more relevant in this day and age. Ceph, a software-defined storage, provides an original solution for this problem and guarantees a resilient and self-healing way for managing large amount of data up to the Exabyte level. In this session I will talk about a new feature introduced in oVirt 3.6 which provides the ability to integrate with Red Hat Ceph storage using Cinder, a storage service used mainly for OpenStack. This integration reveals new opportunities and tools for storage management in a scalable and virtualized way and also opens the door for interesting future integrations with other storage providers.
In this session I will describe how oVirt, an open source virtualization management platform, has extended and elevated its storage virtualization management capabilities by integrating with Cinder, a storage service, to manage resources from the Ceph Storage. oVirt 3.6 revolutionize the way it manages virtualized storage to be much more scalable and flexible, and opens the door for future integrations with well known storage providers such as NetApp, EMC, HP and more.
Simone Tiraboschi, Senior Software Engineer presso Red Hat, presenta oVirt, che, secondo Red Hat, è la risposta open-source alle esigenze di virtualizzazione in ambito enterprise. oVirt è una piattaforma di virtualizzazione con molteplici funzionalità; è basata su KVM, fornisce una semplice interfaccia WEB di amministrazione ed un accesso programmatico via API. Garantisce alta affidabilità.
Even the best system administrator cannot always avoids any and
every disaster that may plague his data center, but he should have a
contingency plan to recover from one - and an administrator that
manages his virtual data centers with oVirt is of course no different.
This session will cover the new features introduced in oVirt 3.5.0 to
handle such scenarios and will showcase how stringing together a set
of building blocks can produce a well rounded solution for disaster
scenarios.
The demand for managing a large amount of data in a scalable yet reliable and cost-effective way has became more and more relevant in this day and age. Ceph, a software-defined storage, provides an original solution for this problem and guarantees a resilient and self-healing way for managing large amount of data up to the Exabyte level. In this session I will talk about a new feature introduced in oVirt 3.6 which provides the ability to integrate with Red Hat Ceph storage using Cinder, a storage service used mainly for OpenStack. This integration reveals new opportunities and tools for storage management in a scalable and virtualized way and also opens the door for interesting future integrations with other storage providers.
In this session I will describe how oVirt, an open source virtualization management platform, has extended and elevated its storage virtualization management capabilities by integrating with Cinder, a storage service, to manage resources from the Ceph Storage. oVirt 3.6 revolutionize the way it manages virtualized storage to be much more scalable and flexible, and opens the door for future integrations with well known storage providers such as NetApp, EMC, HP and more.
Simone Tiraboschi, Senior Software Engineer presso Red Hat, presenta oVirt, che, secondo Red Hat, è la risposta open-source alle esigenze di virtualizzazione in ambito enterprise. oVirt è una piattaforma di virtualizzazione con molteplici funzionalità; è basata su KVM, fornisce una semplice interfaccia WEB di amministrazione ed un accesso programmatico via API. Garantisce alta affidabilità.
Even the best system administrator cannot always avoids any and
every disaster that may plague his data center, but he should have a
contingency plan to recover from one - and an administrator that
manages his virtual data centers with oVirt is of course no different.
This session will cover the new features introduced in oVirt 3.5.0 to
handle such scenarios and will showcase how stringing together a set
of building blocks can produce a well rounded solution for disaster
scenarios.
Users need the ability to increase storage space available in oVirt storage data domains (ISCSI & FC) without increasing the number of LUNs presented to it.
Feature Page: http://www.ovirt.org/Features/LUN_Resize
Step by Step - Reusing old features to build new onesAllon Mureinik
Designing monolithic infrastructures is a common mistake in large projects. However, more often than not, these infrastructures are too generic, make false assumptions or are simply delivered too late for feature developers to use, becoming "white elephants".
This presentation is a case study of the work done by my team to deliver Live Merging of Snapshots oVirt from the initial steps in oVirt 3.1.0 to the full delivery in 3.5.0, and how good design can be feature-driven, building infra-structures step by step, while gaining small wins during the process.
Containers for the Enterprise: Delivering OpenShift on OpenStack for Performa...Stephen Gordon
Imagine being able to stand up thousands of tenants with thousands of apps, running thousands of Docker-formatted container images and routes, all on a self-healing cluster. Now, take that one step further with all of those images being updatable through a single upload to the registry, and with zero downtime. In this session, Steve Gordon of the Red Hat OpenStack Platform team will show you just that. Steve will walk through a recent benchmarking deployment using the Cloud Native Computing Foundation’s (CNCF) new 1,000 node cluster with OpenStack and Red Hat’s OpenShift Container Platform, the enterprise-ready Kubernetes for developers.
This is the slide deck I used for the developer workshop I presented the first day of OpenStack Day India 2016. It gives and overview of how to be a contributor to OpenStack with a walk through of the various steps to get started and tips and tricks for working with the development process.
A Container Stack for Openstack - OpenStack Silicon ValleyStephen Gordon
OpenStack is an Infrastructure as a Service offering that provides a powerful abstraction layer for interacting with your datacenter infrastructure, supported by a wide array of pluggable drivers for existing physical and virtual infrastructure investments. In this session, you’ll learn how OpenStack is evolving to integrate with the Linux, Docker, Kubernetes stack to provide the ideal infrastructure platform for modern containerized applications. You’ll learn how you can modernize application delivery using the Linux, Docker, Kubernetes stack provided by Red Hat while seamlessly using the authentication, network, and storage infrastructure services provided by an underlying OpenStack cloud.
Containerd internals: building a core container runtimeDocker, Inc.
In this talk, we’ll briefly overview of the OpenWhisk serverless (function-as-a-service) framework that initially used the full Docker container engine as the execution vehicle for invoking user functions via containers. After several performance and stability challenges, this project decided to assess the various layers of the Docker engine (containerd and runC) as potential options for the function invoker. Out of that work came an open source project, bucketbench, that can be used to generate benchmarks of container lifecycle operations (e.g., start, stop, kill, remove, pause, unpause) and compare multithreaded operation throughput and stability of each optional engine.
This talk will provide details on the bucketbench project, explain how it has been used to generate performance data for these container runtimes, and shares lessons learned along the way that greatly impact container runtime performance, including bottlenecks in the Linux kernel.
In this talk you’ll learn how you can use bucketbench for your own performance tuning or assessment of container runtimes and how you can collaborate on improvements to the bucketbench project.
KVM (Kernel-based Virtual Machine) is a full virtualization solution built into the Linux kernel. OpenStack Foundation user surveys consistently indicate that KVM is the most commonly used Hypervisor for OpenStack deployments, managed using the Libvirt driver for OpenStack Compute (Nova). Despite this sustained popularity development of the driver, and indeed the underlying Hypervisor itself, continues at a frantic pace.
This presentation will help you make sense of it all starting with an overview of the way Nova, Libvirt, and KVM interact before analysing progress made in Kilo on utilizing key Libvirt/KVM features in Nova including:
Instance vCPU pinning
Huge page backed instances
Enhanced NUMA topology awareness
...and more! The session will close with a discussion of how in addition to exposing existing Libvirt/KVM features emerging OpenStack use cases - such as Network Function Virtualization (NFV) and High Performance Computing (HPC) - are driving open innovation in the Libvirt, QEMU, and KVM projects themselves.
Can we leverage the resource of public cloud for gaming, streaming, transcoding, machine learning and visualized CAD application on demand? Yes if it provides the capability and infrastructure to utilize GPUs. Can we get high performance networking in the cloud as what I have in the bare metal environment? Yes with SR-IOV. How to achieve them? In this presentation we describe Discrete Device Assignment (also known as PCI Pass-through) support for GPU and network adapter in Linux guest and SR-IOV architectures of Linux guest with near-native performance profile running on Hyper-V. We also will share how to integrate accelerated graphics and networking capabilities in Microsoft Azure infrastructure.
A review about the scenarios with most expectations always from an OpenNebula’s
integration point of view. What can we expect nowadays of both technologies and
what should they improve in the future.
Talk held at the OpenNebula Techday Barcelona 2017:
https://opennebula.org/community/techdays/techday-barcelona-2017/
OpenStack Tokyo Meeup - Gluster Storage DayDan Radez
November 2012 Tokyo OpenStack meetup was dedicated to using Gluster storage. This presentation showed the fuse mount method to integrating gluster into OpenStack. There are new drivers that have been developed that make mounting gluster volumes to instances more efficient. This presentation doesn't show how to use them.
oVirt and OpenStack look kind of similar from a distance. But they cater to different use-cases. That said, they do have some common needs. How can they work together? And when is it better to use one over the other?
Users need the ability to increase storage space available in oVirt storage data domains (ISCSI & FC) without increasing the number of LUNs presented to it.
Feature Page: http://www.ovirt.org/Features/LUN_Resize
Step by Step - Reusing old features to build new onesAllon Mureinik
Designing monolithic infrastructures is a common mistake in large projects. However, more often than not, these infrastructures are too generic, make false assumptions or are simply delivered too late for feature developers to use, becoming "white elephants".
This presentation is a case study of the work done by my team to deliver Live Merging of Snapshots oVirt from the initial steps in oVirt 3.1.0 to the full delivery in 3.5.0, and how good design can be feature-driven, building infra-structures step by step, while gaining small wins during the process.
Containers for the Enterprise: Delivering OpenShift on OpenStack for Performa...Stephen Gordon
Imagine being able to stand up thousands of tenants with thousands of apps, running thousands of Docker-formatted container images and routes, all on a self-healing cluster. Now, take that one step further with all of those images being updatable through a single upload to the registry, and with zero downtime. In this session, Steve Gordon of the Red Hat OpenStack Platform team will show you just that. Steve will walk through a recent benchmarking deployment using the Cloud Native Computing Foundation’s (CNCF) new 1,000 node cluster with OpenStack and Red Hat’s OpenShift Container Platform, the enterprise-ready Kubernetes for developers.
This is the slide deck I used for the developer workshop I presented the first day of OpenStack Day India 2016. It gives and overview of how to be a contributor to OpenStack with a walk through of the various steps to get started and tips and tricks for working with the development process.
A Container Stack for Openstack - OpenStack Silicon ValleyStephen Gordon
OpenStack is an Infrastructure as a Service offering that provides a powerful abstraction layer for interacting with your datacenter infrastructure, supported by a wide array of pluggable drivers for existing physical and virtual infrastructure investments. In this session, you’ll learn how OpenStack is evolving to integrate with the Linux, Docker, Kubernetes stack to provide the ideal infrastructure platform for modern containerized applications. You’ll learn how you can modernize application delivery using the Linux, Docker, Kubernetes stack provided by Red Hat while seamlessly using the authentication, network, and storage infrastructure services provided by an underlying OpenStack cloud.
Containerd internals: building a core container runtimeDocker, Inc.
In this talk, we’ll briefly overview of the OpenWhisk serverless (function-as-a-service) framework that initially used the full Docker container engine as the execution vehicle for invoking user functions via containers. After several performance and stability challenges, this project decided to assess the various layers of the Docker engine (containerd and runC) as potential options for the function invoker. Out of that work came an open source project, bucketbench, that can be used to generate benchmarks of container lifecycle operations (e.g., start, stop, kill, remove, pause, unpause) and compare multithreaded operation throughput and stability of each optional engine.
This talk will provide details on the bucketbench project, explain how it has been used to generate performance data for these container runtimes, and shares lessons learned along the way that greatly impact container runtime performance, including bottlenecks in the Linux kernel.
In this talk you’ll learn how you can use bucketbench for your own performance tuning or assessment of container runtimes and how you can collaborate on improvements to the bucketbench project.
KVM (Kernel-based Virtual Machine) is a full virtualization solution built into the Linux kernel. OpenStack Foundation user surveys consistently indicate that KVM is the most commonly used Hypervisor for OpenStack deployments, managed using the Libvirt driver for OpenStack Compute (Nova). Despite this sustained popularity development of the driver, and indeed the underlying Hypervisor itself, continues at a frantic pace.
This presentation will help you make sense of it all starting with an overview of the way Nova, Libvirt, and KVM interact before analysing progress made in Kilo on utilizing key Libvirt/KVM features in Nova including:
Instance vCPU pinning
Huge page backed instances
Enhanced NUMA topology awareness
...and more! The session will close with a discussion of how in addition to exposing existing Libvirt/KVM features emerging OpenStack use cases - such as Network Function Virtualization (NFV) and High Performance Computing (HPC) - are driving open innovation in the Libvirt, QEMU, and KVM projects themselves.
Can we leverage the resource of public cloud for gaming, streaming, transcoding, machine learning and visualized CAD application on demand? Yes if it provides the capability and infrastructure to utilize GPUs. Can we get high performance networking in the cloud as what I have in the bare metal environment? Yes with SR-IOV. How to achieve them? In this presentation we describe Discrete Device Assignment (also known as PCI Pass-through) support for GPU and network adapter in Linux guest and SR-IOV architectures of Linux guest with near-native performance profile running on Hyper-V. We also will share how to integrate accelerated graphics and networking capabilities in Microsoft Azure infrastructure.
A review about the scenarios with most expectations always from an OpenNebula’s
integration point of view. What can we expect nowadays of both technologies and
what should they improve in the future.
Talk held at the OpenNebula Techday Barcelona 2017:
https://opennebula.org/community/techdays/techday-barcelona-2017/
OpenStack Tokyo Meeup - Gluster Storage DayDan Radez
November 2012 Tokyo OpenStack meetup was dedicated to using Gluster storage. This presentation showed the fuse mount method to integrating gluster into OpenStack. There are new drivers that have been developed that make mounting gluster volumes to instances more efficient. This presentation doesn't show how to use them.
oVirt and OpenStack look kind of similar from a distance. But they cater to different use-cases. That said, they do have some common needs. How can they work together? And when is it better to use one over the other?
oVirt UI Plugin Infrastructure and the oVirt-Foreman pluginOved Ourfali
In this presentation I show how one can use the new oVirt-Engine UI-Plugin infrastructure,
to add a new oVirt UI plugins, and show an example of a Foreman-UI-plugin, that allows querying Foreman information on oVirt entities.
Disaster Recovery Strategies Using oVirt's new Storage Connection Management ...Allon Mureinik
A short overview of oVirt 3.3's Storage Connection Management feature, and several examples how this feature can be used in Disaster Recovery strategies.
Practical information on how to Optimize Virtual Machines for High Performance by Boyan Krosnov, Chief Product Officer at StorPool Storage
Presentation delivered at OpenNebula TechDay Sofia on 25-th of February 2016
Elena Belotti, Web Designer e insegnante alla Plat1 Academy, spiega quanto sia importante apprendere le basi della logica di programmazione. Ha descritto perchè è importante e perchè farlo giocando, utilizzando, ad esempio, strumenti come Scratch.
Red Hat Gluster Storage - Direction, Roadmap and Use-CasesRed_Hat_Storage
Red Hat Gluster Storage is open, software-defined storage that helps you manage big, unstructured, and semistructured data. This product is based on the open source project GlusterFS, a distributed scale-out file system technology, and focuses on file sharing, analytics, and hyper-converged use cases.
In this session, you will:
See real-life case studies about Red Hat Gluster Storage’s usage in production environments, including ideal workloads.
Learn about the Red Hat Gluster Storage roadmap, including innovations from the GlusterFS community pipeline.
Gain insights into how the product will be integrated with Red Hat Enterprise Virtualization (including hyperconvergence), Red Hat Satellite, and Red Hat Enterprise Linux OpenStack Platform.
Mastering kvm virtualization- A complete guide of KVM virtualizationHumble Chirammal
Mastering KVM virtualization is a complete guide to understand KVM virtualization. Mastering KVM Virtualization is a culmination of all the knowledge we gained by
troubleshooting, configuring and fixing bug on KVM virtualization. We
authored this book for system administrators, DevOps practitioners and developers who have
a good hands-on knowledge of Linux and would like to sharpen their skills of open
source virtualization. The chapters in this book are written with a focus on practical
examples that should help you deploy a robust virtualization environment, suiting
your organization's needs. Our expectation is that, once you have finished the book,
you should have a good understanding of KVM virtualization, its tools to build
and manage diverse virtualization environments.
Virtualization with KVM (Kernel-based Virtual Machine)Novell
As a technical preview, SUSE Linux Enterprise Server 11 contains KVM, which is the next-generation virtualization software delivered with the Linux kernel. In this technical session we will demonstrate how to set up SUSE Linux Enterprise Server 11 for KVM, install some virtual machines and deal with different storage and networking setups.
To demonstrate live migration we will also show a distributed replicated block device (DRBD) setup and a setup based on iSCSI and OCFS2, which are included in SUSE Linux Enterprise Server 11 and SUSE Linux Enterprise 11 High Availability Extension.
Red Hat Enterprise Linux: Open, hyperconverged infrastructureRed_Hat_Storage
The next generation of IT will be built around flexible infrastructures and operational efficiencies, lowering costs and increasing overall business value in the organization.
A hyperconverged infrastructure that's built on Red Hat supported technologies--including Linux, Gluster storage, and oVirt virtualization manager--will run on commodity x86 servers using the performance of local storage, to deliver a cost-effective, modular, highly scalable, and secure hyperconverged solution.
DCEU 18: Use Cases and Practical Solutions for Docker Container Storage on Sw...Docker, Inc.
Mark Church - Product Manager, Docker
Don Stewart - Solutions Architect, Docker
Persistent storage has quickly advanced from something considered incompatible with containers to a mature set of solutions and patterns that have been thoroughly adopted by the industry. We’ll define the persistent characteristics of different use-cases and map these to some of the many solutions that exist for container storage. From this talk you’ll learn about the storage options available to users on Swarm, Kubernetes, on-premises, cloud, and how they work and compare to each other. You’ll also learn how to characterize different persistent application requirements and the solutions best for suited for them.
A Storage Orchestrator for Kubernetes
Rook turns distributed storage systems into self-managing, self-scaling, self-healing storage services. It automates the tasks of a storage administrator: deployment, bootstrapping, configuration, provisioning, scaling, upgrading, migration, disaster recovery, monitoring, and resource management.
Rook uses the power of the Kubernetes platform to deliver its services: cloud-native container management, scheduling, and orchestration.
gVisor, Kata Containers, Firecracker, Docker: Who is Who in the Container Space?ArangoDB Database
View the video of this webinar here: https://www.arangodb.com/arangodb-events/gvisor-kata-containers-firecracker-docker/
Containers* have revolutionized the IT landscape and for a long time. Docker seemed to be the default whenever people were talking about containerization technologies**. But traditional container technologies might not be suitable if strong isolation guarantees are required. So recently new technologies such as gVisor, Kata Container, or firecracker have been introduced to close the gap between the strong isolation of virtual machines and the small resource footprint of containers.
In this talk, we will provide an overview of the different containerization technologies, discuss their tradeoffs, and provide guidance for different use cases.
* We will define the term container in more detailed during the talk
** and yes we will also cover some of the pre-docker container space!
Containers and Nutanix - Acropolis Container ServicesNEXTtour
This presentation was given at the London Nutanix user group (NUG) on Oct 26 by Denis Guyadeen. If you would like to join a NUG, you can find more information here http://bit.ly/NTNXUG - Hope to see you at a community meeting!
JBoss Architect Forum London - October 2013 - Platform as a What?JBossArchitectForum
• State of the Container: From Tomcat to JEE and beyond
• In-Memory Computing: How can a Data Grid accelerate your applications?
• PaaS: Learn how Red Hat's OpenShift has helped PayPal increase developer productivity
Containers package code and runtime dependencies to offer greater portability to cloud-native applications. But containers are ephemeral by design. If containers fail, stateful applications lose all of their data, leaving your enterprise open to the risk of lost revenue and lower customer satisfaction.
As you consider deploying containers in production, you'll need enterprise-calibre persistent storage that's scalable, secure, and container-aware.
Red Hat is the ideal provider for versatile, multi-purpose storage for containerized applications. Red Hat offers storage for containers, letting you attach modern software-defined storage to container platforms or bridge to traditional storage. In addition, Red Hat offers storage in containers, orchestrated by Kubernetes, delivering storage services and applications out of the same containers.
Container native storage reaches a new level of storage capabilities on the OpenShift Container Platform. Container-native storage can now be used for all the key infrastructure pieces of OpenShift: the registry, logging, and metrics services.
Examining caching solutions that you can use in Windows Azure, such as the Windows Azure Cache service, In-role cache, Memcached, Redis, and Couchbase.
The presentation shows the different features of each caching solution, pros and cons, deployment strategies, and deployment steps.
The OpenEBS Hangout #4 was held on 22nd December 2017 at 11:00 AM (IST and PST) where a live demo of cMotion was shown . Storage policies of OpenEBS 0.5 were also explained
Storage 101: Rook and Ceph - Open Infrastructure Denver 2019Sean Cohen
Starting from the basics, we explore the advantages of using Rook as a Storage operator to serve Ceph storage, the leading Software-Defined Storage platform in the Open Source world. Ceph automates the internal storage management, while Rook automates the user-facing operations and effectively turns a storage technology into a service transparent to the user. The combination delivers an impressive improvement in UX and provides the ideal storage platform for Kubernetes.
A comprehensive examination of use cases and open problems will complement our review of the Rook architecture. We will deep-dive into what Rook does well, what it does not do (yet), and what trade-offs using a storage operator involves operationally. With live access to a running cluster, we will showcase Rook in action as we discuss its capabilities.
https://www.openstack.org/summit/denver-2019/summit-schedule/events/23515/storage-101-rook-and-ceph
How to Survive an OpenStack Cloud Meltdown with CephSean Cohen
What if you lost your datacenter completely in a catastrophe, but your users hardly noticed? Sounds like a mirage, but it’s absolutely possible.
This talk will showcase OpenStack features enabling multisite and disaster recovery functionalities. We’ll present the latest capabilities of OpenStack and Ceph for Volume and Image Replication using Ceph Block and Object as the backend storage solution, as well as look at the future developments they are driving to improve and simplify the relevant architecture use cases, such as Distributed NFV, an emerging use case that rationalizes your IT by using less control planes and allows you to spread your VNF on multiple datacenters and edge deployments.
In this session you will learn about wew OpenStack features enabling Multisite and distributed deployments, as well as review key use cases, architecture design and best practices to help operations avoid the OpenStack cloud Meltdown nightmare.
https://youtu.be/n2S7uNC_KMw
https://goo.gl/cRNGBK
3-2-1 Action! Running OpenStack Shared File System Service in ProductionSean Cohen
As OpenStack’s Shared File System Service is getting more and more adoption as one of top leading emerging projects in OpenStack deployments (according to the last OpenStack foundation user survey), we would like to share some of the key customers use cases such as DevOps, Containers and Enterprise Applications as well review the latest Newton release project updates towards delivering a production-grade deployments.
Slides from OpenStack Summit Barcelona,, October 25, 2016
Session video: https://www.youtube.com/watch?v=F5o-EbESNr8
Peanut Butter and jelly: Mapping the deep Integration between Ceph and OpenStackSean Cohen
Ceph is the most widely deployed storage technology used with OpenStack, most often because it's an open source, massively scalable, unified software-defined storage solution. Its popularity is also due to its unique and optimized technical integration with the OpenStack services and its pure-software approach to scaling. In this session, we'll review how Ceph is integrated into Nova, Glance, Keystone, Cinder, and Manila and demonstrate why using traditional storage products won’t give you the full benefits of an elastic cloud infrastructure. We’ll also cover the flexible deployment options, available through Red Hat Enterprise Linux OpenStack Platform and Red Hat Ceph Storage, for seamless operations and key scenarios like disaster recovery. We'll discuss architectural options for deploying a multisite OpenStack cluster and cover the varying levels of maturity in the OpenStack services for configuring multisite. This session will also show how other technologies are using OpenStack Ceph to increase performance and reduce power consumption, such as Intel SSDs. This will include reference architectures and best practices for Ceph and SSDs.
Protecting the Galaxy - Multi-Region Disaster Recovery with OpenStack and CephSean Cohen
IT organizations require a disaster recovery strategy addressing outages with loss of storage, or extended loss of availability at the primary site. Applications need to rapidly migrate to the secondary site and transition with little or no impact to their availability.This talk will cover the various architectural options and levels of maturity in OpenStack services for building multi-site configurations using the Mitaka release. We’ll present the latest capabilities for Volume, Image and Object Storage with Ceph as the backend storage solution, and look at the future developments the OpenStack and Ceph communities are driving to improve and simplify the relevant use cases.
Slides from OpenStack Austin Summit 2016 session: http://alturl.com/hpesz
Manila, an update from Liberty, OpenStack Summit - TokyoSean Cohen
Manila is a community-driven project that presents the management of file shares (e.g. NFS, CIFS, HDFS) as a core service to OpenStack. Manila currently works with a variety of storage platforms, as well as a reference implementation based on a Linux NFS server.
Manila is exploding with new features, use cases, and deployers. In this session, we'll give an update on the new capabilities added in the Liberty release:
• Integration with OpenStack Sahara
• Migration of shares across different storage back-ends
• Support for availability zones (AZs) and share replication across these AZs
• The ability to grow and shrink file shares on demand
• New mount automation framework
• and much more…
As well as provide a quick look of whats coming up in Mitaka release with Share Replication demo
The road to enterprise ready open stack storage as serviceSean Cohen
The OpenStack storage projects continue to mature each cycle exposing more and more Enterprise cloud storage infrastructure functionalities around high availability, security, business continuity, & provisioning, that redefines Enterprise storage to Storage as a service for both production, test & development cloud workloads.
Dude where's my volume, open stack summit vancouver 2015Sean Cohen
"Dude, where's my volume? A guide to storage backup, migration, and replication with OpenStack Cinder"
OpenStack Cinder now has a wide variety of options for moving and copying storage volumes, but it's not always clear which API calls are designed for which use cases. In this talk, we'll review the storage management workflows for disaster recovery, performance management, and day-to-day operational maintenance using Ceph as an example storage backend. We'll focus on both single and multi-site options for both end users and OpenStack administrators, so attendees should find ways to sleep easier at night knowing how to look after their data.
https://openstacksummitmay2015vancouver.sched.org/event/de8516a550835a338d09634143bed655?iframe=yes&w=i:0;&sidebar=yes&bg=no#?iframe=yes&w=i:0;&sidebar=yes&bg=no
When disaster strikes the cloud: Who, what, when, where and how to recoverSean Cohen
Enterprise applications needs to be able to survive large scale disasters. While some born-on-the-cloud applications have built-in disaster recovery functionality, non-born-on-the-cloud enterprise applications typically expect the infrastructure to provide disaster recovery support. OpenStack provides various building blocks that enable an OpenStack application to survive a disaster; these building blocks are being improved in Juno and Kilo. Some of these building blocks need to be enabled by the OpenStack cloud administrator and others need to be leveraged by the application deployer. In this presentation, we will review basic disaster recovery concepts covering when, where, and what is done at each stage of the application cloud life-cycle. We will describe the existing building blocks and we will explain the roles of cloud administrator and the cloud end-user, in enabling OpenStack applications to survive a disaster. We will then detail new features in Juno and coming in Kilo that will help enhance OpenStack's disaster recovery support. We will conclude by detailing the remaining gaps and present some tools that address these gaps, allowing an application to survive a disaster when running on an OpenStack cloud.
OpenStack Summit Session: https://youtu.be/Dj5sELG9keE
Deterministic capacity planning for OpenStack as elastic cloud infrastructureSean Cohen
Capacity planning for elastic cloud infrastructure platforms like OpenStack is critical for successful deployments. The proper sizing of compute resources within OpenStack allows for easier scheduling, optimal efficiency in hardware utilization, and consistency of resource allocation.
Google Compute Engine and Amazon Web Services offer deterministic compute resources designed to meet both cloud provider business requirements and cloud consumer service-level requirements. In this session, we'll explore these public provider approaches, extend them to OpenStack, and provide sizing data and tools to help with your deployment.
In this session, Keith Basil, Sean Cohen, and Tushar Katarki discuss:
-Approaches for providing consistent compute service levels in OpenStack.
-Building instance families for your workloads.
-Sizing compute node for OpenStack.
-Storage & Network sizing or elastic clouds
- Capacity planning tools & benchmarks
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Monitoring Java Application Security with JDK Tools and JFR Events
Kvm forum 2013 - future integration points for oVirt storage
1. The future
integration points for
oVirt storage
Birds of a feather session
Sean Cohen, Ayal Baron
Red Hat
KVM Forum, October 2013
1
2. Agenda
●
What's new in oVirt 3.3 Storage?
●
Storage directions
●
●
Storage Offloading
●
IO Performance
●
●
Software Defined Storage
Replication
The future integration points for oVirt Storage
●
●
DRDB
●
●
Openstack
Extending oVIrt API
3.4 & Beyond
2
3. What's new in oVirt 3.3 Storage?
Business Continuity:
●
●
Backup and Restore API for Independent Software Ve
ndors
(3.3.2)
Manage Storage Connections (Multipath & DR)(3.3.1)
Disk Management:
●
Enable online virtual drive resize
●
Virtio-SCSI support
●
Disks Block Alignment scan
●
Disk Hooks (for disk hot-plug/unplug)
3
4. Storage directions
Software Defined Storage (SDS)
●
●
●
GlusterFS native storage domain in oVirt 3.3
Converged Storage Hypervisor (self hosted)
Storage Offloading:
●
LibStorageMgmt integration
●
Ability to plugin external storage array into
oVirt/VDSM virtualization stack, in a vendor
neutral way
●
Can provide Array Offload capabilities, such as
Snapshots/Clones, xCopy
4
5. Storage directions
LibStorageMgmt
Storage Offloading
●
●
Open source, vendor agnostic
library which provides an API
for managing external storage
arrays
Current array support (varying
levels of functionality)
●
●
Linux software target
SMI-I compliant arrays
(NetApp, EMC etc)
5
6. Storage directions
Storage Offloading:
●
OpenStack CInder integration
●
Storage offload by design
●
LibSM Driver
●
Existing Eco System
IO Performance:
●
On Hypervisors running multiple VMs all writes
end up Random
●
Bad for:
● SSD
● SMR
6
7. Storage Replication directions
●
Array Based
●
●
3.3 DR support via Manage Storage Connections
Distributed
●
●
Ceph
●
●
Gluster
DRBD
Hypervisor based
●
●
QEMU-KVM
Cinder?
7
8. Future integration points
●
OpenStack Storage stack integration
●
Glance - consumed in oVirt 3.3
●
Cinder
●
●
●
Consume
Provide
Extending oVirt API
●
oVirt third-party UI plug-in framework:
●
●
NetApp VCS 1.0 integration available for oVirt 3.2
oVirt Backup & Restore API next phases:
●
●
Leverage qemu-qa Microsoft Windows Volume
Shadow Copy Service (VSS)
Leverage qemu block layer Change Block Tracking
(CBT) to cover incremental backups.
8
9. 3.4 & Beyond
●
Get rid of Storage Pool Manager
●
Proper iSCSI-multipathing
●
Single disk snapshots
●
Import existing storage domain
●
Read-only disks
●
Support VM Fleecing (qemu-kvm/Libvirt)
●
Snapshots Live Merge
9