Short presentation made for OpenStack London "Tokyo Aftermath" meetup, on current upstream activity in the OpenStack HA developers community around high availability for compute nodes.
Technical overview of how SUSE OpenStack Cloud uses Chef to implement highly available OpenStack infrastructure services.
Target audience: curious developers in the upstream openstack-chef community
These slides were extracted from internal HA training for SUSE OpenStack Cloud developers, and slightly modified for the benefit of the openstack‐chef community.
The document discusses high availability (HA) techniques in OpenStack. It covers HA concepts for both stateless and stateful services. For compute HA, it discusses server evacuation and instance migration without and with shared storage. It then covers different HA options for OpenStack controllers, including Pacemaker/Corosync/DRBD for active-passive HA and Galera for active-active MySQL HA. It also discusses using Keepalived, HAProxy and VRRP for load balancing and failover of API services. Finally, it presents a sample highly available OpenStack architecture and lists additional resources.
OpenShift v3 uses an overlay VXLAN network to connect pods within a project. Traffic between pods on a node uses Linux bridges, while inter-node communication uses the VXLAN overlay network. Services are exposed using a service IP and iptables rules to redirect traffic to backend pods. For external access, services are associated with router pods using a DNS name, and traffic is load balanced to backend pods by HAProxy in the router pod.
A study and practice of OpenStack release Kilo HA deployment. The Kilo document has some errors, and it's hardly find a detailed document to describe how to deploy a HA cloud based on Kilo release. Hope this slides can provide some clues.
The primary requirements for OpenStack based clouds (public, private or hybrid) is that they must be massively scalable and highly available. There are a number of interrelated concepts which make the understanding and implementation of HA complex. The potential for not implementing HA correctly would be disastrous.
This session was presented at the OpenStack Meetup in Boston Feb 2014. We discussed interrelated concepts as a basis for implementing HA and examples of HA for MySQL, Rabbit MQ and the OpenStack APIs primarily using Keepalived, VRRP and HAProxy which will reinforce the concepts and show how to connect the dots.
Dockerizing OpenStack for High AvailabilityDaniel Krook
This document discusses Dockerizing OpenStack high availability services. It begins by outlining existing challenges with OpenStack HA including complex configuration, scaling complexity, and lack of automation/visibility. It then discusses how Docker can help by allowing applications and dependencies to be packaged in lightweight containers, improving scaling, density, flexibility and reducing overhead. The document provides an example of running OpenStack services like Nova API in Docker containers for improved HA and manageability. It discusses sharing images in a private Docker registry and orchestrating container management.
Technical overview of how SUSE OpenStack Cloud uses Chef to implement highly available OpenStack infrastructure services.
Target audience: curious developers in the upstream openstack-chef community
These slides were extracted from internal HA training for SUSE OpenStack Cloud developers, and slightly modified for the benefit of the openstack‐chef community.
The document discusses high availability (HA) techniques in OpenStack. It covers HA concepts for both stateless and stateful services. For compute HA, it discusses server evacuation and instance migration without and with shared storage. It then covers different HA options for OpenStack controllers, including Pacemaker/Corosync/DRBD for active-passive HA and Galera for active-active MySQL HA. It also discusses using Keepalived, HAProxy and VRRP for load balancing and failover of API services. Finally, it presents a sample highly available OpenStack architecture and lists additional resources.
OpenShift v3 uses an overlay VXLAN network to connect pods within a project. Traffic between pods on a node uses Linux bridges, while inter-node communication uses the VXLAN overlay network. Services are exposed using a service IP and iptables rules to redirect traffic to backend pods. For external access, services are associated with router pods using a DNS name, and traffic is load balanced to backend pods by HAProxy in the router pod.
A study and practice of OpenStack release Kilo HA deployment. The Kilo document has some errors, and it's hardly find a detailed document to describe how to deploy a HA cloud based on Kilo release. Hope this slides can provide some clues.
The primary requirements for OpenStack based clouds (public, private or hybrid) is that they must be massively scalable and highly available. There are a number of interrelated concepts which make the understanding and implementation of HA complex. The potential for not implementing HA correctly would be disastrous.
This session was presented at the OpenStack Meetup in Boston Feb 2014. We discussed interrelated concepts as a basis for implementing HA and examples of HA for MySQL, Rabbit MQ and the OpenStack APIs primarily using Keepalived, VRRP and HAProxy which will reinforce the concepts and show how to connect the dots.
Dockerizing OpenStack for High AvailabilityDaniel Krook
This document discusses Dockerizing OpenStack high availability services. It begins by outlining existing challenges with OpenStack HA including complex configuration, scaling complexity, and lack of automation/visibility. It then discusses how Docker can help by allowing applications and dependencies to be packaged in lightweight containers, improving scaling, density, flexibility and reducing overhead. The document provides an example of running OpenStack services like Nova API in Docker containers for improved HA and manageability. It discusses sharing images in a private Docker registry and orchestrating container management.
High availability and fault tolerance of openstackDeepak Mane
This document discusses building a fault tolerant and highly available architecture for OpenStack. It proposes:
1. A master-master cluster architecture for MySQL and session-level replication for RabbitMQ to provide high availability for the database and message broker components.
2. Disk-level replication using DBRD for Glance, Swift, and Cinder to provide redundancy at the storage level.
3. Ensuring high availability for networking and the Horizon dashboard.
4. Developing predictive and reactive models to detect failures in Nova, Swift, and compute instances and enable recovery of all components.
The document recommends using Pacemaker for cluster-level management and Corosync for reliable messaging between cluster nodes.
[OpenStack Days Korea 2016] Track1 - Mellanox CloudX - Acceleration for Cloud...OpenStack Korea Community
1) Mellanox's CloudX platform enhances cloud performance through technologies like its Spectrum switch, ConnectX-4 adapters, and software solutions.
2) These solutions provide high-speed networking, efficient virtual networking through overlay acceleration, and data transfer technologies like RDMA.
3) CloudX reference architectures allow building efficient, high-performance, and scalable IaaS clouds using Mellanox interconnect solutions and off-the-shelf components.
Moving to Nova Cells without Destroying the WorldMike Dorman
This document discusses using cells in OpenStack Nova to scale deployments. Cells create a hierarchy with a top-level API cell and multiple compute cells. Each cell has its own database, message queue, and services. The document outlines planning the conversion, preparing the environment by expanding RabbitMQ and splitting services, configuring the compute and API cells, importing data to the API cell, and restarting services. It notes some caveats like limitations on certain notifications and objects between cells.
OpenNebulaConf 2016 - Measuring and tuning VM performance by Boyan Krosnov, S...OpenNebula Project
In this session we'll explore measuring VM performance and evaluating changes to settings or infrastructure which can affect performance positively. We'll also share the best current practice for architecture for high performance clouds from our experience.
OpenStack Nova - Developer IntroductionJohn Garbutt
This document provides an overview of Nova, OpenStack's compute service. It discusses Nova's architecture, code structure, API concepts, upgrade process, and how different groups work together as part of the upstream community. The new upgrade process aims to minimize downtime by expanding the database schema, restarting services individually, and signaling services to reload configuration. Collaboration across various groups with different perspectives is important to OpenStack's open development model.
The document discusses Ceph, an open source distributed storage platform that provides unified object, block, and file storage. It describes how the speaker's company Hostvn deployed Ceph in production, including using it with OpenStack. They started with a small proof-of-concept cluster using all SSD drives before expanding to a larger cluster with more nodes. Key lessons learned included keeping the design simple, monitoring performance closely during rebalancing, and realizing there is no one-size-fits-all model for Ceph deployment. Future plans include upgrading networking and replacing current storage with Ceph.
Jakub Pavlik discusses high availability versus disaster recovery in OpenStack clouds. He describes four types of high availability in OpenStack: physical infrastructure, OpenStack control services, virtual machines, and applications. For each type, he outlines concepts like active/passive and active/active configurations, specific technologies used like Pacemaker, Corosync, HAProxy, and MySQL Galera, and considerations for shared and non-shared storage. Finally, he provides examples of high availability architectures and methods used by different OpenStack vendors.
The document provides tips for OpenStack cloud transformation. It discusses (1) unifying CPU models on compute nodes for live migration, (2) clustering compute nodes by host aggregate, and (3) configuring options to slow down CPU during live migration for stability. Other tips include increasing HAProxy connection limits, enabling multiple network queues, implementing port level security, and ensuring adequate entropy for scale-out systems. The overall document offers best practices and configurations for improving performance and stability during OpenStack cloud transformations.
Presentation delivered at LinuxCon China 2017. Rethinking the Operating System.
A new wave of Operating Systems optimized for containers appeared on the horizon making us excited and puzzeled at the same time.
"Why do we need anything different for containers when traditional OSs served us well in the last 25+ years?" "Isn't Kubernetes just another package to install on top of my favorite distro?"" Will this obsolete my whole infrastructure?" are some of the questions this talk will shed some light on.
Explore the journey SUSE made in rethinking the OS: From a conservative linux distribution to a platform that goes hand in hand with the needs of Microservices.
You will get an insight at what lessons were learned during the intense development effort that lead to SUSE Containers as a Service Platform, how the obstacles along the way were lifted and why "Upstream first" is - and should always be - the rule.
Heart of the SwarmKit: Store, Topology & Object ModelDocker, Inc.
Heart of the SwarmKit: Store, Topology & Object Model by Aaron, Andrea, Stephen D (Docker)
Swarmkit repo - https://github.com/docker/swarmkit
Liveblogging: http://canopy.mirage.io/Liveblog/SwarmKitDDS2016
This document provides an overview of how to integrate Red Hat CloudForms 4 with OpenShift 3 to retrieve container metrics. It describes demonstrating the integration, installing and configuring CloudForms 4 to add OpenShift 3.1 as a container provider, configuring Hawkular on OpenShift to expose metrics via the Hawkular Metrics API, and modifying the Hawkular configuration to make it compatible with how CloudForms expects to retrieve metrics. The presentation contains steps for configuring CloudForms and OpenShift, creating service accounts, adding roles, and configuring routes to expose the Hawkular Metrics API on a port CloudForms can access.
Nowadays there is significant diversity in Infrastructure
as a Service (IaaS) clouds. The differences span from
virtualization technology and hypervisors, through storage
and network configuration, to the cloud management
APIs. These differences make migration of a VM (or
a set of VMs) from a private cloud into a public cloud,
or between different public clouds, complicated or even
impractical for many use-cases.
HVX is a virtualization platform that enables complete
abstraction of underlying cloud infrastructure from the
application virtual machines. HVX allows deployment
of existing VMs into the cloud without any modifications,
mobility between the clouds and easy duplication
of the entire deployment.
HVX can be deployed on almost any existing IaaS
cloud. Each instance of the HVX deployment packs in
a nested hypervisor, virtual hardware, network and storage
configuration.
Combined with image store and management APIs,
the HVX can be used for the creation of a virtual cloud
that utilizes existing cloud provider infrastructure as the
hardware rather than using physical servers, switches and
storage.
With the increasing maturity of OpenStack Neutron, and increasing support for a large number of services , we'll examine how SDNs can leverage the container technology for supporting "as a Service" components.
Securing & Monitoring Your K8s Cluster with RBAC and Prometheus”.Opcito Technologies
Opcito Technologies is a proud partner with Kubernetes, an open-source system for container orchestration.
We will be talking about:
• Features of Kubernetes 1.6
• RBAC Configurations
• RBAC Use Cases
• Running Prometheus in Kubernetes
• Prometheus Operator - Deployment, Cluster & Service Monitoring
Besides huge success in mobile, ARM is also ambitious in server field. Software ecosystem is now a barrier for wide deployment of ARM servers in data center. ARM Shanghai Workloads team is working on clouding and big data software enablement and optimization on ARM64 platform.
In this presentation, Yibo Cai will introduce the status and challenges of running OpenStack on ARM servers, with emphasis on OpenStack compute, storage and networking.
Nova compute and controller services provide virtual machines and coordinate nova services in OpenStack. Nova compute runs on each node and interacts with the hypervisor like KVM to launch VMs. The nova controller runs most nova services including the scheduler, which dispatches VM requests to nodes based on filters. Key components include the nova API, compute, and conductor services. Compute resources must be designed considering the number of processors, memory allocation, storage, resource pools, and over-commit ratios to optimize VM deployment in OpenStack.
Kubernetes currently has two load balancing mode: userspace and IPTables. They both have limitation on scalability and performance. We introduced IPVS as third kube-proxy mode which scales kubernetes load balancer to support 50,000 services. Beyond that, control plane needs to be optimized in order to deploy 50,000 services. We will introduce alternative solutions and our prototypes with detailed performance data.
In the container ecosystem, there is perhaps no technology that has received more focus and attention than orchestration and scheduling. Mesos, Kubernetes, and Swarm have established themselves as the leading technology choices in this space.
In this talk, Sheng will discuss what he learned from working directly with hundreds of users who have deployed one of these frameworks. He will look at how these frameworks will continue to evolve and if there’re any gaps and opportunities in container orchestration and scheduling. Sheng will make a case that there are still room for innovation and new orchestration and scheduling frameworks will be created in the future. He will discuss what new frameworks might look like--the features, functionalities, and attributes that differentiate them from the mainstream frameworks today.
Covers overview of CoreOS and current status of CoreOS projects. Presented at Open source meetup, Bangalore(http://www.meetup.com/Bangalore-Open-Source-Meetup/events/229763724/)
Deep dive into highly available open stack architecture openstack summit va...Arthur Berezin
This document summarizes a presentation on highly available OpenStack architecture. It discusses using Pacemaker and HAProxy for high availability enabling services. Shared databases like MariaDB Galera and message queues like RabbitMQ are made highly available. Individual OpenStack services like Keystone, Glance, Cinder, Nova, Neutron, and Horizon are made highly available through active-active clustering, load balancing, and fencing. The presentation covers topologies for controller, compute, network, and storage nodes. It provides examples of making individual services highly available and discusses ongoing work and future plans to improve high availability in OpenStack.
High availability and fault tolerance of openstackDeepak Mane
This document discusses building a fault tolerant and highly available architecture for OpenStack. It proposes:
1. A master-master cluster architecture for MySQL and session-level replication for RabbitMQ to provide high availability for the database and message broker components.
2. Disk-level replication using DBRD for Glance, Swift, and Cinder to provide redundancy at the storage level.
3. Ensuring high availability for networking and the Horizon dashboard.
4. Developing predictive and reactive models to detect failures in Nova, Swift, and compute instances and enable recovery of all components.
The document recommends using Pacemaker for cluster-level management and Corosync for reliable messaging between cluster nodes.
[OpenStack Days Korea 2016] Track1 - Mellanox CloudX - Acceleration for Cloud...OpenStack Korea Community
1) Mellanox's CloudX platform enhances cloud performance through technologies like its Spectrum switch, ConnectX-4 adapters, and software solutions.
2) These solutions provide high-speed networking, efficient virtual networking through overlay acceleration, and data transfer technologies like RDMA.
3) CloudX reference architectures allow building efficient, high-performance, and scalable IaaS clouds using Mellanox interconnect solutions and off-the-shelf components.
Moving to Nova Cells without Destroying the WorldMike Dorman
This document discusses using cells in OpenStack Nova to scale deployments. Cells create a hierarchy with a top-level API cell and multiple compute cells. Each cell has its own database, message queue, and services. The document outlines planning the conversion, preparing the environment by expanding RabbitMQ and splitting services, configuring the compute and API cells, importing data to the API cell, and restarting services. It notes some caveats like limitations on certain notifications and objects between cells.
OpenNebulaConf 2016 - Measuring and tuning VM performance by Boyan Krosnov, S...OpenNebula Project
In this session we'll explore measuring VM performance and evaluating changes to settings or infrastructure which can affect performance positively. We'll also share the best current practice for architecture for high performance clouds from our experience.
OpenStack Nova - Developer IntroductionJohn Garbutt
This document provides an overview of Nova, OpenStack's compute service. It discusses Nova's architecture, code structure, API concepts, upgrade process, and how different groups work together as part of the upstream community. The new upgrade process aims to minimize downtime by expanding the database schema, restarting services individually, and signaling services to reload configuration. Collaboration across various groups with different perspectives is important to OpenStack's open development model.
The document discusses Ceph, an open source distributed storage platform that provides unified object, block, and file storage. It describes how the speaker's company Hostvn deployed Ceph in production, including using it with OpenStack. They started with a small proof-of-concept cluster using all SSD drives before expanding to a larger cluster with more nodes. Key lessons learned included keeping the design simple, monitoring performance closely during rebalancing, and realizing there is no one-size-fits-all model for Ceph deployment. Future plans include upgrading networking and replacing current storage with Ceph.
Jakub Pavlik discusses high availability versus disaster recovery in OpenStack clouds. He describes four types of high availability in OpenStack: physical infrastructure, OpenStack control services, virtual machines, and applications. For each type, he outlines concepts like active/passive and active/active configurations, specific technologies used like Pacemaker, Corosync, HAProxy, and MySQL Galera, and considerations for shared and non-shared storage. Finally, he provides examples of high availability architectures and methods used by different OpenStack vendors.
The document provides tips for OpenStack cloud transformation. It discusses (1) unifying CPU models on compute nodes for live migration, (2) clustering compute nodes by host aggregate, and (3) configuring options to slow down CPU during live migration for stability. Other tips include increasing HAProxy connection limits, enabling multiple network queues, implementing port level security, and ensuring adequate entropy for scale-out systems. The overall document offers best practices and configurations for improving performance and stability during OpenStack cloud transformations.
Presentation delivered at LinuxCon China 2017. Rethinking the Operating System.
A new wave of Operating Systems optimized for containers appeared on the horizon making us excited and puzzeled at the same time.
"Why do we need anything different for containers when traditional OSs served us well in the last 25+ years?" "Isn't Kubernetes just another package to install on top of my favorite distro?"" Will this obsolete my whole infrastructure?" are some of the questions this talk will shed some light on.
Explore the journey SUSE made in rethinking the OS: From a conservative linux distribution to a platform that goes hand in hand with the needs of Microservices.
You will get an insight at what lessons were learned during the intense development effort that lead to SUSE Containers as a Service Platform, how the obstacles along the way were lifted and why "Upstream first" is - and should always be - the rule.
Heart of the SwarmKit: Store, Topology & Object ModelDocker, Inc.
Heart of the SwarmKit: Store, Topology & Object Model by Aaron, Andrea, Stephen D (Docker)
Swarmkit repo - https://github.com/docker/swarmkit
Liveblogging: http://canopy.mirage.io/Liveblog/SwarmKitDDS2016
This document provides an overview of how to integrate Red Hat CloudForms 4 with OpenShift 3 to retrieve container metrics. It describes demonstrating the integration, installing and configuring CloudForms 4 to add OpenShift 3.1 as a container provider, configuring Hawkular on OpenShift to expose metrics via the Hawkular Metrics API, and modifying the Hawkular configuration to make it compatible with how CloudForms expects to retrieve metrics. The presentation contains steps for configuring CloudForms and OpenShift, creating service accounts, adding roles, and configuring routes to expose the Hawkular Metrics API on a port CloudForms can access.
Nowadays there is significant diversity in Infrastructure
as a Service (IaaS) clouds. The differences span from
virtualization technology and hypervisors, through storage
and network configuration, to the cloud management
APIs. These differences make migration of a VM (or
a set of VMs) from a private cloud into a public cloud,
or between different public clouds, complicated or even
impractical for many use-cases.
HVX is a virtualization platform that enables complete
abstraction of underlying cloud infrastructure from the
application virtual machines. HVX allows deployment
of existing VMs into the cloud without any modifications,
mobility between the clouds and easy duplication
of the entire deployment.
HVX can be deployed on almost any existing IaaS
cloud. Each instance of the HVX deployment packs in
a nested hypervisor, virtual hardware, network and storage
configuration.
Combined with image store and management APIs,
the HVX can be used for the creation of a virtual cloud
that utilizes existing cloud provider infrastructure as the
hardware rather than using physical servers, switches and
storage.
With the increasing maturity of OpenStack Neutron, and increasing support for a large number of services , we'll examine how SDNs can leverage the container technology for supporting "as a Service" components.
Securing & Monitoring Your K8s Cluster with RBAC and Prometheus”.Opcito Technologies
Opcito Technologies is a proud partner with Kubernetes, an open-source system for container orchestration.
We will be talking about:
• Features of Kubernetes 1.6
• RBAC Configurations
• RBAC Use Cases
• Running Prometheus in Kubernetes
• Prometheus Operator - Deployment, Cluster & Service Monitoring
Besides huge success in mobile, ARM is also ambitious in server field. Software ecosystem is now a barrier for wide deployment of ARM servers in data center. ARM Shanghai Workloads team is working on clouding and big data software enablement and optimization on ARM64 platform.
In this presentation, Yibo Cai will introduce the status and challenges of running OpenStack on ARM servers, with emphasis on OpenStack compute, storage and networking.
Nova compute and controller services provide virtual machines and coordinate nova services in OpenStack. Nova compute runs on each node and interacts with the hypervisor like KVM to launch VMs. The nova controller runs most nova services including the scheduler, which dispatches VM requests to nodes based on filters. Key components include the nova API, compute, and conductor services. Compute resources must be designed considering the number of processors, memory allocation, storage, resource pools, and over-commit ratios to optimize VM deployment in OpenStack.
Kubernetes currently has two load balancing mode: userspace and IPTables. They both have limitation on scalability and performance. We introduced IPVS as third kube-proxy mode which scales kubernetes load balancer to support 50,000 services. Beyond that, control plane needs to be optimized in order to deploy 50,000 services. We will introduce alternative solutions and our prototypes with detailed performance data.
In the container ecosystem, there is perhaps no technology that has received more focus and attention than orchestration and scheduling. Mesos, Kubernetes, and Swarm have established themselves as the leading technology choices in this space.
In this talk, Sheng will discuss what he learned from working directly with hundreds of users who have deployed one of these frameworks. He will look at how these frameworks will continue to evolve and if there’re any gaps and opportunities in container orchestration and scheduling. Sheng will make a case that there are still room for innovation and new orchestration and scheduling frameworks will be created in the future. He will discuss what new frameworks might look like--the features, functionalities, and attributes that differentiate them from the mainstream frameworks today.
Covers overview of CoreOS and current status of CoreOS projects. Presented at Open source meetup, Bangalore(http://www.meetup.com/Bangalore-Open-Source-Meetup/events/229763724/)
Deep dive into highly available open stack architecture openstack summit va...Arthur Berezin
This document summarizes a presentation on highly available OpenStack architecture. It discusses using Pacemaker and HAProxy for high availability enabling services. Shared databases like MariaDB Galera and message queues like RabbitMQ are made highly available. Individual OpenStack services like Keystone, Glance, Cinder, Nova, Neutron, and Horizon are made highly available through active-active clustering, load balancing, and fencing. The presentation covers topologies for controller, compute, network, and storage nodes. It provides examples of making individual services highly available and discusses ongoing work and future plans to improve high availability in OpenStack.
OpenStack Best Practices and Considerations - terasky tech dayArthur Berezin
- Arthur Berezin presented on best practices for deploying enterprise-grade OpenStack implementations. The presentation covered OpenStack architecture, layout considerations including high availability, and best practices for compute, storage, and networking deployments. It provided guidance on choosing backend drivers, overcommitting resources, and networking designs.
오픈스택이 가진 기술에 대하여 설명합니다.
1. 오픈소스기반 OpenStack 클라우드 시스템
2. OpenStack 기술 개요 및 동향
3. OpenStack 의 Community 개발 체계
4. OpenStack HA를 위한 방안
5. OpenStack SDN 개발 동향
6. Neutron OVS-DPDK 가속화와 구현방안
The document provides instructions for updating and setting up the SMART Response classroom response system software. It describes how to update the SMART Product Update software, access and install the SMART Response software, set up a classroom and class lists, create quizzes and surveys with multiple question types, deliver the assessments to students, and export the results. The document is a comprehensive guide to getting started with and utilizing the key features of the SMART Response classroom software.
Open stack meetup 2014 11-13 - 101 + high availabilityRick Ashford
This document discusses building a highly available OpenStack cloud. It describes OpenStack components and the importance of high availability for control planes and guest instances. The document recommends using a control cluster across multiple availability zones for the control plane and placing workloads in high availability clusters within each zone to protect against failures.
The document discusses SUSE's technology preview of containerized OpenStack, called Airship. It provides an overview of Airship and how it uses containers and Kubernetes to deploy OpenStack services. Key benefits highlighted include improved scalability, ease of upgrades, and security of individual components through container isolation. The preview aims to demonstrate a unified lifecycle management approach for deploying and managing containerized OpenStack using the open source Airship project.
Deploying SUSE Cloud in a Multi-Hypervisor Enterprise EnvironmentRick Ashford
This document discusses deploying OpenStack in a multi-hypervisor enterprise environment. It notes that enterprises currently have large investments in virtualization infrastructure like VMware, and outlines some of the challenges of introducing an open-source cloud platform. It then describes how OpenStack's drivers allow it to support multiple hypervisors like VMware and Hyper-V, providing a single control plane to manage workloads on different infrastructure. Limitations of the hypervisor drivers are also discussed.
Using Ceph in a Private Cloud - Ceph Day Frankfurt Ceph Community
This document summarizes how to set up a Ceph cluster for private cloud storage using SUSE Cloud. It describes configuring over 10 storage nodes and 3 monitor nodes for the Ceph cluster. It explains integrating the external Ceph cluster with SUSE Cloud to provide block storage, image storage, and object storage services. It also covers setting up Ceph directly with SUSE Cloud using Crowbar to deploy all nodes.
Hands-On with Heat: Service Orchestration in SUSE CloudRick Ashford
This document provides an overview of three ways to deploy services in the cloud: manually through the dashboard, via scripting with APIs, and using the OpenStack Heat project. Heat allows pre-defining compute, network, and storage requirements and automating deployment of whole applications. It uses AWS CloudFormation templates and provides both REST and CloudFormation query APIs. While Heat has an upfront complexity, it offers easier management of complex, composite cloud applications in the long run.
CAPS: What's best for deploying and managing OpenStack? Chef vs. Ansible vs. ...Daniel Krook
Presentation at the OpenStack Summit in Tokyo, Japan on October 29, 2015.
http://sched.co/49vI
This talk will cover the pros and cons of four different OpenStack deployment mechanisms. Puppet, Chef, Ansible, and Salt for OpenStack all claim to make it much easier to configure and maintain hundreds of OpenStack deployment resources. With the advent of large-scale, highly available OpenStack deployments spread across multiple global regions, the choice of which deployment methodology to use has become more and more relevant.
Beyond the initial day-one deployment, when it comes to the day-two and beyond questions of updating and upgrading existing OpenStack deployments, it becomes all the more important choose the right tool.
Come join the Bluebox and IBM team to discuss the pros and cons of these approaches. We look at each of these four tools in depth, explore their design and function, and determine which scores higher than others to address your particular deployment needs.
Daniel Krook - Senior Software Engineer, Cloud and Open Source Technologies, IBM
Paul Czarkowski - Cloud Engineer at Blue Box, an IBM company
Daniel Krook - Senior Software Engineer, Cloud and Open Source Technologies, IBM
CAPS: What's best for deploying and managing OpenStack? Chef vs. Ansible vs. ...Animesh Singh
Chef, Puppet, Ansible, and Salt are popular configuration management tools for deploying and managing OpenStack. Each tool has its own strengths and weaknesses. Chef focuses on infrastructure automation and uses a Ruby DSL. Puppet uses a custom DSL and is focused on compliance. Ansible emphasizes orchestration and uses YAML playbooks. Salt uses a Python-based interface and focuses on remote execution and data collection at scale. All four tools provide options for deploying and managing OpenStack, with varying levels of documentation and community support.
Casos de uso para aplicaciones tradicionales en un mundo de contenedoresSUSE España
This document discusses using stateful applications on Kubernetes. It describes how containers are inherently stateless but Kubernetes provides tools like volumes and StatefulSets to manage state. Volume plugins allow connecting containers to various storage backends. PersistentVolumes and PersistentVolumeClaims provide an API for storage requests. StatefulSets ensure ordered deployment and stable networking for stateful applications. Real-world examples from SUSE include containerizing OpenStack and Cloud Foundry platforms.
This document provides an overview of the new features and updates in SUSE OpenStack Cloud 9, including support for multi-attached storage, Ironic improvements, and functionality from previous versions. It also discusses upcoming plans for SUSE OpenStack Cloud 10, such as containerized OpenStack deployment using Kubernetes and Airship, scalability improvements, and SDN integration.
The document discusses SUSE's portfolio and Container as a Service Platform (CaaSP). It provides an overview of how SUSE products like OpenStack Cloud, Enterprise Storage, and Cloud Application Platform integrate and deploy on CaaSP. This gives benefits like simplified management, upgrades, and reuse of skills across products. The document also outlines new versions and features for these products in upcoming years, including using CaaSP as a common deployment platform.
OpenNebulaConf2015 1.07 Cloud for Scientific Computing @ STFC - Alexander DibboOpenNebula Project
The Science and Technology Facilities Council is a UK Research Council which funds research and provides large facilities to the UK Scientific Community. This includes running a Tier 1 site for the LHC computing project, the JASMIN Super Data Cluster and a number of other HPC and HTC facilities. The Scientific Computing Department at the Rutherford Appleton Laboratory has been developing a cloud for use across both sites of the Department and in the wider scientific community. This is an OpenNebula backed by Ceph block storage. I will give a brief background of the project, describe our set up, some use cases and the work we have done around OpenNebula (including a simplified web front-end and a number of hooks to provide us with traceability). I will also discuss how we are creating an elastic boundary between our HTC batch farm and cloud.
Author Biography
I am a Systems Administrator in the Scientific Computing Department of the UK’s Science and Technology Facilities Council. I work as part of the cloud team and I also work on a number of Grid services including our HTC batch farm for the LHC computing project.
Prior to my position here I worked in IT at a SMB focusing on Storage and Virtualisation, in particular Hyper-V and VMWare.
OSMC 2010 | Insides SUSE Linux by Joachim WernerNETWAYS
SUSE Linux Enterprise is the most interoperable platform for mission-critical computing - both in traditional client-server and in virtual environments - from the desktop to the datacenter. In this talk some basic information about the data for the monitoring of SUSE LINUX and which opportunities for monitoring SUSE LINUX offers will be given.
The document discusses moving OpenStack to structured state management. It outlines use cases from deployers, developers, and users around ensuring reliability, debugging state transitions, and optimizing resource scheduling. Currently, state transitions are ad-hoc and distributed, making them hard to follow, recover from, and extend. The document proposes prototyping an orchestration solution to consolidate state management and make transitions and recoveries clearly defined. Key benefits would include less scattered state and recovery logic, faster provisioning, and improved scheduling capabilities.
This document provides an introduction to real-time systems and discusses approaches to making Linux a real-time operating system. It defines hard and soft real-time systems and explains why Linux is commonly used instead of dedicated real-time operating systems. The document then discusses two main solutions, PREEMPT_RT and Xenomai 3, which provide patches to make Linux meet timing constraints through different approaches. It also provides an overview of basic real-time concepts like scheduling algorithms, preemptive vs. non-preemptive scheduling, and interprocess communication.
Build Platform as a Service (PaaS) with SUSE Studio, WSO2 Middleware, and EC2 WSO2
To view recording of this webinar please use the below URL:
http://wso2.com/library/webinars/2015/04/build-platform-as-a-service-paas-with-suse-studio-wso2-middleware-and-ec2
Raise your Cloud program up a notch by delivering Platform as a Service (PaaS) to application teams. SUSE Studio, WSO2, and EC2 helps you assemble, build, and deploy portable application services instead of platform silos. In this session, Chris will demonstrate how you can rapidly build standardized application stack environments and offer a policy compliant Platform as a Service to your application project teams.
Mike Friesenegger from SUSE will give a presentation on using SUSE Cloud to help deploy SAP workloads. He will introduce SUSE Cloud and discuss use cases for deploying SAP applications with SUSE Cloud, including SAP application testing/evaluation and SAP system copying. He will then demonstrate SUSE Cloud.
1) Ceph is an open source distributed storage system that provides scalable, fault-tolerant storage and manages petabytes of data across clusters of commodity hardware.
2) It uses Object Storage Daemons (OSDs) that serve storage objects and replicate data across peers for redundancy. Multiple OSDs can be grouped in monitor nodes that track cluster state.
3) Ceph offers self-healing capabilities through redundancy and allows data to be placed close to applications for performance. It provides APIs and integration with clouds for flexible, software-defined storage.
Similar to Compute node HA - current upstream development (20)
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
1. Compute node HA
a.k.a. “can pets survive in OpenStack?”
Adam Spiers
Senior Software Engineer, Cloud & High Availability
aspiers@suse.com
OpenStack London Meetup, Wednesday 18th
November
– short update on upstream development
3. 3
Typical HA control plane in OpenStack
Pacemaker Cluster
Control Node 1 Node
DRBD
PostgreSQL
RabbitMQ
Keystone
Glance
Nova
Dashboard
Cinder
Neutron
Database Cluster
Node 1 Node 2
DRBD or shared storage
Database
Message Queue
Services Cluster
Node 1 Node 2 Node 3
Orchestration
Keystone
Glance
Nova
Dashboard
Cinder
Telemetry
Neutron
• Maximises cloud uptime
• Automatic restart of
OpenStack controller
services
• Active/Active API services
with load balancing
• DB + MQ either
Active/Active or
Active/Passive
4. 4
Under the covers
Services Cluster
Node 1 Node 2 Node 3
• Recommended by
official HA guide
• HAProxy distributes service
requests
• Pacemaker
‒ monitoring and control of nodes
and services
• Corosync
‒ cluster membership /
messaging / quorum /
leadership election
Corosync
Pacemaker
HAProxy
But what I really want to do is keep my workloads up!
5. 6
HA Cluster
Control node
OS
Message queue
Database
Identity
Images
Block storage
Networking
Dashboard
Compute
OS
Compute node
nova-compute
libvirt
HA only on control plane
OS
Compute node
nova-compute
libvirt
OS
Compute node
nova-compute
libvirt
6. 7
HA Cluster
Control node
OS
Message queue
Database
Identity
Images
Block storage
Networking
Dashboard
Compute
OS
Compute node
nova-compute
libvirt
Can we simply extend the cluster?
OS
Compute node
nova-compute
libvirt
OS
Compute node
nova-compute
libvirt
8. 9
Scaling up
• Corosync requires <= 32 nodes
• But we want lots of compute nodes!
• The obvious workarounds are ugly
‒ Multiple compute clusters
‒ introduces unwanted artificial boundaries
‒ Clusters inside / between guest VM instances
‒ requires cloud users to modify guest images (installing & configuring cluster
software)
‒ cluster stacks are not OS-agnostic
‒ cloud is supposed to make things easier not harder!
9. 10
pacemaker_remote to the rescue!
• New(-ish) Pacemaker feature
• Allows arbitrary scalability of an existing
Pacemaker cluster
11. 12
Capabilities
• Increases availability of compute nodes
‒ Detects failed compute services
‒ Automatic recovery of compute services where possible
• “Quarantines” failing compute nodes
‒ STONITH (fencing) extends to remote nodes
• Coordinates with control plane
‒ VMs on dead compute nodes are resurrected elsewhere
‒ In nova, this is described as “evacuation”
15. 16
Public Health Warning
• nova evacuate does not do evacuation
• nova evacuate does resurrection
• In Vancouver, nova developers considered a rename
‒ Hasn't happened yet
‒ Due to impact, seems unlikely to happen any time soon
‒ Whenever you see “evacuate” in a nova-related context,
pretend you saw “resurrect”
16. 17
Existing solutions
• NovaCompute / NovaEvacuate custom OCF RAs
‒ used by Red Hat / SUSE / Intel
‒ works with known limitations
• EvacuationD
‒ PoC to address above limitations
‒ decouples resurrection workflow from Pacemaker
• Masakari (NTT)
‒ similar architecture, different code
‒ monitoring at 3 layers (node, process, hypervisor)
• Approach of AWcloud / ChinaMobile
‒ very different; uses consul / raft / gossip
17. 18
Proposed solutions
• Use Mistral to orchestrate resurrection workflow
• Intel currently working on prototype
• Possibly the most promising approach
‒ Mistral considered pretty solid
‒ This is exactly the kind of thing it was designed for
• However, Mistral currently a SPoF … oops
‒ Don't worry, should be fixed in mitaka cycle
• Feasibility of convergence with Masakari will probably
be analysed within next week or two
18. 19
Community developments
• openstack-resource-agents project now on
stackforge
‒ maintained by me
• New #openstack-ha IRC channel on FreeNode
‒ automatic notifications for activity on HA repositories
• New topic category on openstack-dev@ mailing list
Subject: [HA] i can haz pets in my cloud?
• Weekly IRC meetings at Monday 9am UTC
• HA guide currently undergoing a revamp
• Everyone welcome to get involved!
20. Unpublished Work of SUSE LLC. All Rights Reserved.
This work is an unpublished work and contains confidential, proprietary and trade secret information of SUSE LLC.
Access to this work is restricted to SUSE employees who have a need to know to perform tasks within the scope of
their assignments. No part of this work may be practiced, performed, copied, distributed, revised, modified, translated,
abridged, condensed, expanded, collected, or adapted without the prior written consent of SUSE.
Any use or exploitation of this work without authorization could subject the perpetrator to criminal and civil liability.
General Disclaimer
This document is not to be construed as a promise by any participating company to develop, deliver, or market a
product. It is not a commitment to deliver any material, code, or functionality, and should not be relied upon in making
purchasing decisions. SUSE makes no representations or warranties with respect to the contents of this document,
and specifically disclaims any express or implied warranties of merchantability or fitness for any particular purpose. The
development, release, and timing of features or functionality described for SUSE products remains at the sole
discretion of SUSE. Further, SUSE reserves the right to revise this document and to make changes to its content, at
any time, without obligation to notify any person or entity of such revisions or changes. All SUSE marks referenced in
this presentation are trademarks or registered trademarks of Novell, Inc. in the United States and other countries. All
third-party trademarks are the property of their respective owners.