My talk at http://www.barcampsaigon.com (2016) about how did we architect and configure the OpenStack-based private cloud for running in production at FimPlus.vn
Performance optimization for all flash based on aarch64 v2.0Ceph Community
This document discusses performance optimization techniques for All Flash storage systems based on ARM architecture processors. It provides details on:
- The processor used, which is the Kunpeng920 ARM-based CPU with 32-64 cores at 2.6-3.0GHz, along with its memory and I/O controllers.
- Optimizing performance through both software and hardware techniques, including improving CPU usage, I/O performance, and network performance.
- Specific optimization techniques like data placement to reduce cross-NUMA access, multi-port NIC deployment, using multiple DDR channels, adjusting messaging throttling, and optimizing queue wait times in the object storage daemon (OSD).
- Other
Ceph is an open source distributed storage system designed for scalability and reliability. Ceph's block device, RADOS block device (RBD), is widely used to store virtual machines, and is the most popular block storage used with OpenStack.
In this session, you'll learn how RBD works, including how it:
* Uses RADOS classes to make access easier from user space and within the Linux kernel.
* Implements thin provisioning.
* Builds on RADOS self-managed snapshots for cloning and differential backups.
* Increases performance with caching of various kinds.
* Uses watch/notify RADOS primitives to handle online management operations.
* Integrates with QEMU, libvirt, and OpenStack.
Ceph is an open-source distributed storage platform that provides file, block, and object storage in a single unified system. It uses a distributed storage component called RADOS that provides reliable and scalable storage through data replication and erasure coding across commodity hardware. Higher-level services like RBD provide virtual block devices, RGW provides S3-compatible object storage, and CephFS provides a distributed file system.
DigitalOcean uses Ceph for block and object storage backing for their cloud services. They operate 37 production Ceph clusters running Nautilus and one on Luminous, storing over 54 PB of data across 21,500 OSDs. They deploy and manage Ceph clusters using Ansible playbooks and containerized Ceph packages, and monitor cluster health using Prometheus and Grafana dashboards. Upgrades can be challenging due to potential issues uncovered and slow performance on HDD backends.
Boosting I/O Performance with KVM io_uringShapeBlue
Storage performance is becoming much more important. KVM io_uring attempts to bring the I/O performance of a virtual machine on almost the same level of bare metal. Apache CloudStack has support for io_uring since version 4.16. Wido will show the difference in performance io_uring brings to the table.
Wido den Hollander is the CTO of CLouDinfra, an infrastructure company offering total Webhosting solutions. CLDIN provides datacenter, IP and virtualization services for the companies within TWS. Wido den Hollander is a PMC member of the Apache CloudStack Project and a Ceph expert. He started with CloudStack 9 years ago. What attracted his attention is the simplicity of CloudStack and the fact that it is an open-source solution. During the years Wido became a contributor, a PMC member and he was a VP of the project for a year. He is one of our most active members, who puts a lot of efforts to keep the project active and transform it into a turnkey solution for cloud builders.
-----------------------------------------
The CloudStack European User Group 2022 took place on 7th April. The day saw a virtual get together for the European CloudStack Community, hosting 265 attendees from 25 countries. The event hosted 10 sessions with from leading CloudStack experts, users and skilful engineers from the open-source world, which included: technical talks, user stories, new features and integrations presentations and more.
------------------------------------------
About CloudStack: https://cloudstack.apache.org/
This document discusses integrating the open source GlusterFS distributed file system with QEMU virtualization and the oVirt virtualization management platform. It provides an overview of GlusterFS and how it can be used as a virtual machine image store. It describes integrating GlusterFS with QEMU using different methods and shows performance results. It then outlines the architecture and features for managing GlusterFS volumes from within oVirt, including a REST API and upcoming integration plans. The goal is to provide an open alternative for large scale centralized management of server and desktop virtualization.
teuto.net Netzdienste GmbH is an 18 employee Linux systems house and web development company located in Bielefeld, Germany. They have been offering an OpenStack Ceph storage service in a closed beta since September 2013, running on 20 compute nodes and 5 Ceph storage nodes. Key factors in their decision to use one Ceph cluster to meet all of OpenStack's storage needs included no single point of failure, seamless scalability, and commercial support from Inktank.
Performance optimization for all flash based on aarch64 v2.0Ceph Community
This document discusses performance optimization techniques for All Flash storage systems based on ARM architecture processors. It provides details on:
- The processor used, which is the Kunpeng920 ARM-based CPU with 32-64 cores at 2.6-3.0GHz, along with its memory and I/O controllers.
- Optimizing performance through both software and hardware techniques, including improving CPU usage, I/O performance, and network performance.
- Specific optimization techniques like data placement to reduce cross-NUMA access, multi-port NIC deployment, using multiple DDR channels, adjusting messaging throttling, and optimizing queue wait times in the object storage daemon (OSD).
- Other
Ceph is an open source distributed storage system designed for scalability and reliability. Ceph's block device, RADOS block device (RBD), is widely used to store virtual machines, and is the most popular block storage used with OpenStack.
In this session, you'll learn how RBD works, including how it:
* Uses RADOS classes to make access easier from user space and within the Linux kernel.
* Implements thin provisioning.
* Builds on RADOS self-managed snapshots for cloning and differential backups.
* Increases performance with caching of various kinds.
* Uses watch/notify RADOS primitives to handle online management operations.
* Integrates with QEMU, libvirt, and OpenStack.
Ceph is an open-source distributed storage platform that provides file, block, and object storage in a single unified system. It uses a distributed storage component called RADOS that provides reliable and scalable storage through data replication and erasure coding across commodity hardware. Higher-level services like RBD provide virtual block devices, RGW provides S3-compatible object storage, and CephFS provides a distributed file system.
DigitalOcean uses Ceph for block and object storage backing for their cloud services. They operate 37 production Ceph clusters running Nautilus and one on Luminous, storing over 54 PB of data across 21,500 OSDs. They deploy and manage Ceph clusters using Ansible playbooks and containerized Ceph packages, and monitor cluster health using Prometheus and Grafana dashboards. Upgrades can be challenging due to potential issues uncovered and slow performance on HDD backends.
Boosting I/O Performance with KVM io_uringShapeBlue
Storage performance is becoming much more important. KVM io_uring attempts to bring the I/O performance of a virtual machine on almost the same level of bare metal. Apache CloudStack has support for io_uring since version 4.16. Wido will show the difference in performance io_uring brings to the table.
Wido den Hollander is the CTO of CLouDinfra, an infrastructure company offering total Webhosting solutions. CLDIN provides datacenter, IP and virtualization services for the companies within TWS. Wido den Hollander is a PMC member of the Apache CloudStack Project and a Ceph expert. He started with CloudStack 9 years ago. What attracted his attention is the simplicity of CloudStack and the fact that it is an open-source solution. During the years Wido became a contributor, a PMC member and he was a VP of the project for a year. He is one of our most active members, who puts a lot of efforts to keep the project active and transform it into a turnkey solution for cloud builders.
-----------------------------------------
The CloudStack European User Group 2022 took place on 7th April. The day saw a virtual get together for the European CloudStack Community, hosting 265 attendees from 25 countries. The event hosted 10 sessions with from leading CloudStack experts, users and skilful engineers from the open-source world, which included: technical talks, user stories, new features and integrations presentations and more.
------------------------------------------
About CloudStack: https://cloudstack.apache.org/
This document discusses integrating the open source GlusterFS distributed file system with QEMU virtualization and the oVirt virtualization management platform. It provides an overview of GlusterFS and how it can be used as a virtual machine image store. It describes integrating GlusterFS with QEMU using different methods and shows performance results. It then outlines the architecture and features for managing GlusterFS volumes from within oVirt, including a REST API and upcoming integration plans. The goal is to provide an open alternative for large scale centralized management of server and desktop virtualization.
teuto.net Netzdienste GmbH is an 18 employee Linux systems house and web development company located in Bielefeld, Germany. They have been offering an OpenStack Ceph storage service in a closed beta since September 2013, running on 20 compute nodes and 5 Ceph storage nodes. Key factors in their decision to use one Ceph cluster to meet all of OpenStack's storage needs included no single point of failure, seamless scalability, and commercial support from Inktank.
This document summarizes the updates in Ceph Pacific and previews the updates coming in Quincy. Some of the key updates in Pacific include improved usability through more hands-off defaults, distributed tracing in OSDs, and canceling ongoing scrubs. Performance improvements include more efficient PG deletion and msgr2 wire format. Telemetry features were added to collect anonymized crash reports and device health data. For Quincy, some highlights mentioned are using the mclock scheduler by default, new PG autoscaling profiles, and further BlueStore optimizations.
The document discusses research into data management, file systems, and storage systems being conducted at UC Santa Cruz. Specific projects mentioned include using Ceph as a prototyping platform, the SIRIUS project studying challenges of heterogeneous, multi-tiered storage for exascale systems, the Programmable Storage project developing the Malacology and Mantle systems, and the Skyhook project to build an elastic database system that leverages programmable storage interfaces. The research aims to address issues like data placement, predictable performance at scale, and allowing databases to better utilize storage resources.
Rook.io is an open source cloud-native storage orchestrator for Kubernetes that supports various storage providers like Ceph and allows provisioning of block storage, object storage, and shared file systems using CRDs; it provides a framework to deploy and manage different storage systems on Kubernetes including Ceph which is a highly scalable distributed storage solution for block storage, object storage, and file systems that can be run using Kubernetes primitives. The presenter demonstrates configuring and deploying Rook operator for Ceph which manages the Ceph cluster through CRDs to provide storage for applications like MySQL.
Rook.io is an open source cloud-native storage orchestrator for Kubernetes that supports various storage providers like Ceph and allows provisioning of block storage, object storage, and shared file systems using CRDs; it provides a framework to deploy and manage different storage systems on Kubernetes including Ceph which is a highly scalable distributed storage solution for production environments. The presenter demonstrates configuring and deploying Rook operator for Ceph which manages the Ceph components and allows dynamically growing or replacing the underlying storage infrastructure.
Gluster can provide block storage using LIO/TCMU. It was demonstrated providing an iSCSI block device from a Gluster volume, including block snapshots. Performance numbers were shown and it can integrate with containers by providing a persistent block device. Kubernetes was demonstrated using Gluster block storage by having nodes initiate iSCSI sessions to access a target device mounted in pods. Future work may include more testing, Heketi integration for provisioning, and hyper-convergence.
Making distributed storage easy: usability in Ceph Luminous and beyondSage Weil
Distributed storage is complicated, and historically Ceph hasn't spent a lot of time trying to hide that complexity, instead focusing on correctness, features, and flexibility. There has been a recent shift in focus to simplifying and streamlining the user/operator experience so that the information that is actually important is available without the noise of irrelevant details. Recent feature work has also focused on simplifying configurations that were previously possible but required tedious configuration steps to manage.
This talk will cover the key new efforts in Ceph Luminous that aim to simplify and automate cluster management, as well as the plans for upcoming releases to address longstanding Cephisms that make it "hard" (e.g., choosing PG counts).
This document provides an introduction to accessing Gluster volumes through various methods like traditional filesystem mounts, NFS/CIFS, and the Unified File and Object (UFO) feature. It discusses the benefits of using UFO to access data simultaneously via both REST API and filesystem. It also includes examples of common commands for authenticating and interacting with objects and containers when using UFO with Gluster.
This document summarizes what's new in Ceph. Key updates include improved management and usability features like simplified configuration, hands-off operation, and device health tracking. It also covers new orchestrator capabilities for Kubernetes and container platforms, continued performance optimizations, and multi-cloud capabilities like object storage federation across data centers and clouds.
BlueStore, A New Storage Backend for Ceph, One Year InSage Weil
BlueStore is a new storage backend for Ceph OSDs that consumes block devices directly, bypassing the local XFS file system that is currently used today. It's design is motivated by everything we've learned about OSD workloads and interface requirements over the last decade, and everything that has worked well and not so well when storing objects as files in local files systems like XFS, btrfs, or ext4. BlueStore has been under development for a bit more than a year now, and has reached a state where it is becoming usable in production. This talk will cover the BlueStore design, how it has evolved over the last year, and what challenges remain before it can become the new default storage backend.
GlusterFS is a POSIX-compliant distributed file system that aggregates various storage bricks across commodity servers into a single global namespace. It has no single point of failure or performance bottleneck. Red Hat Storage is an enterprise implementation of GlusterFS. It uses a elastic hashing algorithm to distribute files across bricks without a centralized metadata server. Various translators and volumes types provide features like replication, distribution, striping, and geo-replication. Administration involves adding peers, creating and managing distributed volumes, and manipulating bricks.
Ceph BlueStore - новый тип хранилища в Ceph / Максим Воронцов, (Redsys)Ontico
- Что такое SDS (общие места для (почти) всех решений — масштабирование, абстрагирование от аппаратных ресурсов, управление с помощью политик, кластерные ФС);
- Почему мы решили использовать SDS (нужно было объектное хранилище);
- Почему решили использовать именно Ceph, а не другие открытые (GlusterFS, Swift...) или проприетарные (IBM Elastic Storage, Huawei OceanStor) решения;
- Что еще умеет Ceph, кроме object storage (RBD, CephFS);
- Как работает Ceph (со стороны сервера);
- Что нового дает BlueStore по сравнению с классическим (поверх ФС);
- Сравнение производительности (метрики тестов);
- BlueStore — все еще tech preview;
- Заключение. Ссылки, литература.
This document discusses disaster recovery for oVirt virtualization environments. It describes how oVirt has evolved to support disaster recovery through exporting and importing storage domains. However, this process was manual. oVirt 4.2 added more attributes to OVF files to support automatic registration of VMs and templates between sites. Ansible roles are now available to automate failover and failback between primary and secondary oVirt sites using mapping files. The roles generate mappings, import storage domains, register VMs and templates, and start VMs to recover the secondary site or restore the primary site.
The document summarizes the state of the Gluster community and GlusterFS distributed file system. It discusses that GlusterFS is a scale-out NAS platform that provides a unified, distributed storage system without single points of failure. It also outlines recent updates to GlusterFS 3.3 including improved granular locking, proactive self-healing, and easier rebalancing. The document concludes by previewing upcoming work including better support for virtual machine images, libgfapi client API improvements, and quorum enforcement to prevent split-brain issues.
Ceph is an open source distributed storage system that provides scalable object, block, and file interfaces on a commodity hardware. Luminous, the latest stable release of Ceph, was just released in August. This talk will cover all that is new in Luminous (there is a lot!) and provide a sneak peak at the roadmap for Mimic, which is due out in the Spring.
Gluster d thread_synchronization_using_urcu_lca2016Gluster.org
This document discusses using user space RCU (URCU) for thread synchronization in GlusterD, the management daemon for the Gluster distributed file system. It begins with an introduction to Gluster and GlusterD, noting that GlusterD was initially single-threaded but was changed to be multi-threaded. This introduced the need for thread synchronization, which was initially implemented with a "big lock" but had issues. The document then provides an overview of RCU and how it can provide advantages over read-write locks for thread synchronization. It covers the key mechanisms of RCU, various URCU flavors, URCU APIs, and examples of when URCU would be useful.
OSBConf 2015 | Scale out backups with bareos and gluster by niels de vosNETWAYS
During this talk, Niels will explain the basics of Gluster and show how Bareos integrates with it. Gluster provides a Software Defined Storage environment that can scale-out when the backup storage needs to grow. With a live demonstration Niels shows how simple it is to setup a small Gluster environment and configure Bareos to use the native Gluster protocol.
Even the best system administrator cannot always avoids any and
every disaster that may plague his data center, but he should have a
contingency plan to recover from one - and an administrator that
manages his virtual data centers with oVirt is of course no different.
This session will cover the new features introduced in oVirt 3.5.0 to
handle such scenarios and will showcase how stringing together a set
of building blocks can produce a well rounded solution for disaster
scenarios.
Ceph Pacific is a major release of the Ceph distributed storage system scheduled for March 2021. It focuses on five key themes: usability, performance, ecosystem integration, multi-site capabilities, and quality. New features in Pacific include automated upgrades, improved dashboard functionality, snapshot-based CephFS mirroring, per-bucket replication in RGW, and expanded telemetry collection. Looking ahead, the Quincy release will focus on continued improvements in these areas such as resource-aware scheduling in cephadm and multi-site monitoring capabilities.
This document discusses various approaches to implementing high availability (HA) in OpenStack including active/active and active/passive configurations. It provides an overview of HA techniques used at Deutsche Telekom and eBay/PayPal including load balancing APIs and databases, replicating RabbitMQ and MySQL, and configuring Pacemaker/Corosync for OpenStack services. It also discusses lessons learned around testing failures, placing services across availability zones, and having backups for HA infrastructures.
This document summarizes the updates in Ceph Pacific and previews the updates coming in Quincy. Some of the key updates in Pacific include improved usability through more hands-off defaults, distributed tracing in OSDs, and canceling ongoing scrubs. Performance improvements include more efficient PG deletion and msgr2 wire format. Telemetry features were added to collect anonymized crash reports and device health data. For Quincy, some highlights mentioned are using the mclock scheduler by default, new PG autoscaling profiles, and further BlueStore optimizations.
The document discusses research into data management, file systems, and storage systems being conducted at UC Santa Cruz. Specific projects mentioned include using Ceph as a prototyping platform, the SIRIUS project studying challenges of heterogeneous, multi-tiered storage for exascale systems, the Programmable Storage project developing the Malacology and Mantle systems, and the Skyhook project to build an elastic database system that leverages programmable storage interfaces. The research aims to address issues like data placement, predictable performance at scale, and allowing databases to better utilize storage resources.
Rook.io is an open source cloud-native storage orchestrator for Kubernetes that supports various storage providers like Ceph and allows provisioning of block storage, object storage, and shared file systems using CRDs; it provides a framework to deploy and manage different storage systems on Kubernetes including Ceph which is a highly scalable distributed storage solution for block storage, object storage, and file systems that can be run using Kubernetes primitives. The presenter demonstrates configuring and deploying Rook operator for Ceph which manages the Ceph cluster through CRDs to provide storage for applications like MySQL.
Rook.io is an open source cloud-native storage orchestrator for Kubernetes that supports various storage providers like Ceph and allows provisioning of block storage, object storage, and shared file systems using CRDs; it provides a framework to deploy and manage different storage systems on Kubernetes including Ceph which is a highly scalable distributed storage solution for production environments. The presenter demonstrates configuring and deploying Rook operator for Ceph which manages the Ceph components and allows dynamically growing or replacing the underlying storage infrastructure.
Gluster can provide block storage using LIO/TCMU. It was demonstrated providing an iSCSI block device from a Gluster volume, including block snapshots. Performance numbers were shown and it can integrate with containers by providing a persistent block device. Kubernetes was demonstrated using Gluster block storage by having nodes initiate iSCSI sessions to access a target device mounted in pods. Future work may include more testing, Heketi integration for provisioning, and hyper-convergence.
Making distributed storage easy: usability in Ceph Luminous and beyondSage Weil
Distributed storage is complicated, and historically Ceph hasn't spent a lot of time trying to hide that complexity, instead focusing on correctness, features, and flexibility. There has been a recent shift in focus to simplifying and streamlining the user/operator experience so that the information that is actually important is available without the noise of irrelevant details. Recent feature work has also focused on simplifying configurations that were previously possible but required tedious configuration steps to manage.
This talk will cover the key new efforts in Ceph Luminous that aim to simplify and automate cluster management, as well as the plans for upcoming releases to address longstanding Cephisms that make it "hard" (e.g., choosing PG counts).
This document provides an introduction to accessing Gluster volumes through various methods like traditional filesystem mounts, NFS/CIFS, and the Unified File and Object (UFO) feature. It discusses the benefits of using UFO to access data simultaneously via both REST API and filesystem. It also includes examples of common commands for authenticating and interacting with objects and containers when using UFO with Gluster.
This document summarizes what's new in Ceph. Key updates include improved management and usability features like simplified configuration, hands-off operation, and device health tracking. It also covers new orchestrator capabilities for Kubernetes and container platforms, continued performance optimizations, and multi-cloud capabilities like object storage federation across data centers and clouds.
BlueStore, A New Storage Backend for Ceph, One Year InSage Weil
BlueStore is a new storage backend for Ceph OSDs that consumes block devices directly, bypassing the local XFS file system that is currently used today. It's design is motivated by everything we've learned about OSD workloads and interface requirements over the last decade, and everything that has worked well and not so well when storing objects as files in local files systems like XFS, btrfs, or ext4. BlueStore has been under development for a bit more than a year now, and has reached a state where it is becoming usable in production. This talk will cover the BlueStore design, how it has evolved over the last year, and what challenges remain before it can become the new default storage backend.
GlusterFS is a POSIX-compliant distributed file system that aggregates various storage bricks across commodity servers into a single global namespace. It has no single point of failure or performance bottleneck. Red Hat Storage is an enterprise implementation of GlusterFS. It uses a elastic hashing algorithm to distribute files across bricks without a centralized metadata server. Various translators and volumes types provide features like replication, distribution, striping, and geo-replication. Administration involves adding peers, creating and managing distributed volumes, and manipulating bricks.
Ceph BlueStore - новый тип хранилища в Ceph / Максим Воронцов, (Redsys)Ontico
- Что такое SDS (общие места для (почти) всех решений — масштабирование, абстрагирование от аппаратных ресурсов, управление с помощью политик, кластерные ФС);
- Почему мы решили использовать SDS (нужно было объектное хранилище);
- Почему решили использовать именно Ceph, а не другие открытые (GlusterFS, Swift...) или проприетарные (IBM Elastic Storage, Huawei OceanStor) решения;
- Что еще умеет Ceph, кроме object storage (RBD, CephFS);
- Как работает Ceph (со стороны сервера);
- Что нового дает BlueStore по сравнению с классическим (поверх ФС);
- Сравнение производительности (метрики тестов);
- BlueStore — все еще tech preview;
- Заключение. Ссылки, литература.
This document discusses disaster recovery for oVirt virtualization environments. It describes how oVirt has evolved to support disaster recovery through exporting and importing storage domains. However, this process was manual. oVirt 4.2 added more attributes to OVF files to support automatic registration of VMs and templates between sites. Ansible roles are now available to automate failover and failback between primary and secondary oVirt sites using mapping files. The roles generate mappings, import storage domains, register VMs and templates, and start VMs to recover the secondary site or restore the primary site.
The document summarizes the state of the Gluster community and GlusterFS distributed file system. It discusses that GlusterFS is a scale-out NAS platform that provides a unified, distributed storage system without single points of failure. It also outlines recent updates to GlusterFS 3.3 including improved granular locking, proactive self-healing, and easier rebalancing. The document concludes by previewing upcoming work including better support for virtual machine images, libgfapi client API improvements, and quorum enforcement to prevent split-brain issues.
Ceph is an open source distributed storage system that provides scalable object, block, and file interfaces on a commodity hardware. Luminous, the latest stable release of Ceph, was just released in August. This talk will cover all that is new in Luminous (there is a lot!) and provide a sneak peak at the roadmap for Mimic, which is due out in the Spring.
Gluster d thread_synchronization_using_urcu_lca2016Gluster.org
This document discusses using user space RCU (URCU) for thread synchronization in GlusterD, the management daemon for the Gluster distributed file system. It begins with an introduction to Gluster and GlusterD, noting that GlusterD was initially single-threaded but was changed to be multi-threaded. This introduced the need for thread synchronization, which was initially implemented with a "big lock" but had issues. The document then provides an overview of RCU and how it can provide advantages over read-write locks for thread synchronization. It covers the key mechanisms of RCU, various URCU flavors, URCU APIs, and examples of when URCU would be useful.
OSBConf 2015 | Scale out backups with bareos and gluster by niels de vosNETWAYS
During this talk, Niels will explain the basics of Gluster and show how Bareos integrates with it. Gluster provides a Software Defined Storage environment that can scale-out when the backup storage needs to grow. With a live demonstration Niels shows how simple it is to setup a small Gluster environment and configure Bareos to use the native Gluster protocol.
Even the best system administrator cannot always avoids any and
every disaster that may plague his data center, but he should have a
contingency plan to recover from one - and an administrator that
manages his virtual data centers with oVirt is of course no different.
This session will cover the new features introduced in oVirt 3.5.0 to
handle such scenarios and will showcase how stringing together a set
of building blocks can produce a well rounded solution for disaster
scenarios.
Ceph Pacific is a major release of the Ceph distributed storage system scheduled for March 2021. It focuses on five key themes: usability, performance, ecosystem integration, multi-site capabilities, and quality. New features in Pacific include automated upgrades, improved dashboard functionality, snapshot-based CephFS mirroring, per-bucket replication in RGW, and expanded telemetry collection. Looking ahead, the Quincy release will focus on continued improvements in these areas such as resource-aware scheduling in cephadm and multi-site monitoring capabilities.
This document discusses various approaches to implementing high availability (HA) in OpenStack including active/active and active/passive configurations. It provides an overview of HA techniques used at Deutsche Telekom and eBay/PayPal including load balancing APIs and databases, replicating RabbitMQ and MySQL, and configuring Pacemaker/Corosync for OpenStack services. It also discusses lessons learned around testing failures, placing services across availability zones, and having backups for HA infrastructures.
This document provides an overview of how Bloomberg uses Ceph and OpenStack in its cloud infrastructure. Some key points:
- Bloomberg uses Ceph for object storage with RGW and block storage with RBD. It uses OpenStack for compute functions.
- Initially Bloomberg had a fully converged architecture with Ceph and OpenStack on the same nodes, but this caused performance issues.
- Bloomberg now uses a semi-converged "POD" architecture with dedicated Ceph and OpenStack nodes in separate clusters for better scalability and performance.
- Ephemeral storage provides faster performance than Ceph but lacks data integrity protections. Ceph offers replication and reliability at the cost of some latency.
- Automation with Chef
How to deliver High Performance OpenStack Cloud: Christoph Dwertmann, Vault S...OpenStack
Securing Openstack in Line with the Government ISM and PSPF controls and how to deliver High Performance OpenStack Cloud to address Government Legacy Systems
Audience: Intermediate/Advanced
Topic: Security, Infrastructure, Performance
Abstract: As the CTO of Vault Systems, Christoph will take us through the challenges of implementing ASD’s ISM controls within Vault’s OpenStack cloud to create a Protected Certified OpenStack Platform and give a technical account of some of the optimizations he has done around Ceph on NVMe Storage to deliver High Performance Storage.
Speaker Bio: Christoph Dwertmann, Vault Systems
Christoph is a full stack engineer with four years of experience in deploying and securing Openstack. Fully automated software deployment and self-healing microservice containers are amongst his current interests. As the CTO of Vault Systems he recently deployed the world’s first pure NVMe Ceph cluster into production. From his previous work in network research for the National Science Foundation (NSF) he gathered in-depth knowledge spanning software-defined networks across continents.
OpenStack Australia Day Government - Canberra 2016
https://events.aptira.com/openstack-australia-day-canberra-2016/
99.999% Available OpenStack Cloud - A Builder's GuideDanny Al-Gaaf
This document discusses achieving 99.999% availability for OpenStack cloud services running on Ceph storage. It describes Deutsche Telekom's motivation to build highly available NFV clouds across multiple data centers. Various failure scenarios are considered, such as power, network, hardware failures, and disasters. Setting up OpenStack and Ceph for high availability requires redundant components and careful planning. Ensuring quorum across Ceph monitors and OpenStack services is critical. Achieving five nines availability requires distributing applications across multiple regions to tolerate data center or regional failures.
Community Update at OpenStack Summit BostonSage Weil
This community update from Sage Weil at Red Hat provides information on the current and upcoming releases of Ceph. The document summarizes that Luminous is the current stable release, BlueStore is now stable and default, and there have been significant performance improvements for hardware like HDDs. It also outlines many new features and improvements planned or in development for Ceph components like RBD, RGW, CephFS, erasure coding, and more in upcoming releases like Mimic.
Ceph is an open-source distributed storage system that provides object, block, and file storage on commodity hardware. It uses a pseudo-random placement algorithm called CRUSH to distribute data across a cluster in a fault-tolerant manner without single points of failure. Ceph has various applications including a RADOS Gateway for S3/Swift compatibility, RADOS Block Device for virtual machine images, and a CephFS for a POSIX-compliant distributed file system.
This document discusses using OpenNebula and StorPool to build powerful clouds. StorPool is a software-defined storage system used by managed service providers, cloud providers, and for private clouds. It integrates deeply with OpenNebula, OpenStack, and other platforms. When used with hyper-converged infrastructure and KVM virtualization, StorPool and OpenNebula can provide a scalable, high-performance solution for private or public clouds. StorPool uses a small percentage of server resources but provides high IOPS performance suitable for demanding workloads.
Many companies build new-age KVM clouds, only to find out that their applications & workloads do not perform well. In this talk we’ll show you how to get the most out of your KVM cloud and how to optimize it for performance: You’ll understand why performance matters and how to measure it properly. We’ll teach you how to optimize CPU and memory for ultimate performance and how to tune the storage layer for performance. You’ll find out what are the main components of an efficient new-age cloud and which network components work best. In addition, you’ll learn how to select the right hardware to achieve unmatched performance for your new-age cloud and applications.
Venko Moyankov is an experienced system administrator and solutions architect at StorPool storage. He has experience with managing large virtualizations, working in telcos, designing and supporting the infrastructure of large enterprises. In the last year, his focus has been in helping companies globally to build the best storage solution according to their needs and projects.
OpenNebulaConf2018 - Is Hyperconverged Infrastructure what you need? - Boyan ...OpenNebula Project
This document discusses hyperconverged infrastructure and when it may be a suitable solution. It defines hyperconverged infrastructure as storage using Ethernet networks with block storage running on hypervisor hosts, delivered and supported as a single SKU. The document outlines some scenarios when hyperconverged infrastructure could work well, such as for small availability zones or greenfield deployments that require efficient storage. It also discusses scenarios when a non-hyperconverged approach may be better, such as when independence of infrastructure is important. Examples of potential hyperconverged solutions are provided for shared hosting, small public clouds, and MSPs.
The document discusses the Ceph distributed storage system. It provides an overview of what Ceph is, how it works, and its key features. Ceph is an open-source unified storage system that provides object storage, block storage, and a file system. It uses a distributed system of monitors, storage nodes, and metadata servers to store and retrieve data reliably across multiple machines.
This document introduces the Ceph distributed storage system. Ceph provides object storage, block storage, and a distributed file system. It uses a CRUSH algorithm to distribute data across nodes and provides replication for fault tolerance. Ceph is open source and can scale to large capacities by running on commodity hardware.
Netflix Open Source Meetup Season 4 Episode 2aspyker
In this episode, we will take a close look at 2 different approaches to high-throughput/low-latency data stores, developed by Netflix.
The first, EVCache, is a battle-tested distributed memcached-backed data store, optimized for the cloud. You will also hear about the road ahead for EVCache it evolves into an L1/L2 cache over RAM and SSDs.
The second, Dynomite, is a framework to make any non-distributed data-store, distributed. Netflix's first implementation of Dynomite is based on Redis.
Come learn about the products' features and hear from Thomson and Reuters, Diego Pacheco from Ilegra and other third party speakers, internal and external to Netflix, on how these products fit in their stack and roadmap.
Achieving the Ultimate Performance with KVMDevOps.com
Building and managing a cloud is not an easy task. It needs solid knowledge, proper planning and extensive experience in selecting the proper components and putting them together.
Many companies build new-age KVM clouds, only to find out that their applications & workloads do not perform well. Join this webinar to learn how to get the most out of your KVM cloud and how to optimize it for performance.
Join this webinar and learn:
Why performance matters and how to measure it properly?
What are the main components of an efficient new-age cloud?
How to select the right hardware?
How to optimize CPU and memory for ultimate performance?
Which network components work best?
How to tune the storage layer for performance?
OpenStackTage Cologne - OpenStack at 99.999% availability with CephDanny Al-Gaaf
High availability is a very important and frequently discussed topic for clouds at the infrastructure level. There are several concepts to provide a HA-ready OpenStack and also software defined storage like Ceph is highly available with no single point of failure.
But what about HA if you bring OpenStack and Ceph together? What are the dependencies between them and how do they influence the availability of your cloud instances from the tenant or application point of view?
How does the design of your classic high-available data center, e.g. with two fire compartments, power backup, and redundant power and network lines impact your cluster setup? There are many different scenarios of potential failures. What does this mean regarding building and managing failure zones, especially in case of technologies like Ceph which need to be able to build a quorum to keep up running?
Como creamos QuestDB Cloud, un SaaS basado en Kubernetes alrededor de QuestDB...javier ramirez
QuestDB es una base de datos open source de alto rendimiento. Mucha gente nos comentaba que les gustaría usarla como servicio, sin tener que gestionar las máquinas. Así que nos pusimos manos a la obra para desarrollar una solución que nos permitiese lanzar instancias de QuestDB con provisionado, monitorización, seguridad o actualizaciones totalmente gestionadas.
Unos cuantos clusters de Kubernetes más tarde, conseguimos lanzar nuestra oferta de QuestDB Cloud. Esta charla es la historia de cómo llegamos ahí. Hablaré de herramientas como Calico, Karpenter, CoreDNS, Telegraf, Prometheus, Loki o Grafana, pero también de retos como autenticación, facturación, multi-nube, o de a qué tienes que decir que no para poder sobrevivir en la nube.
The document discusses HP Moonshot, a new server architecture from Hewlett-Packard. It is described as the world's first software-defined server, which uses dense cartridge-based servers in a single chassis for increased efficiency. The Moonshot system is said to provide up to 77% lower costs, 80% less space, 97% less complexity, and 89% less energy compared to traditional servers. It allows for flexible deployment of workloads on optimized cartridges. HP Cloud OS software is also introduced to simplify cloud service delivery on the Moonshot platform.
Similar to Running OpenStack in Production - Barcamp Saigon 2016 (20)
OpenStack 101: Introduction & Technical OverviewThang Man
This document provides an introduction and technical overview of OpenStack presented by Mẫn Thắng. It discusses the motivation for OpenStack, what OpenStack is, key features of OpenStack Compute and Storage, use cases, and resources for learning more and getting involved in the OpenStack community. The presentation encourages participants to try OpenStack locally or on public clouds and provides tips on how to participate in the OpenStack community.
The document discusses firewall fundamentals, including:
- Firewalls control network traffic flow between networks with different security levels. They authenticate access, manage traffic, and protect resources.
- Firewalls can be software, appliances, or integrated into devices. They operate at OSI layers 2-7 and deny or allow access based on predefined rules.
- Common firewall architectures include single-box, screened host, screened subnet, DMZ, and dual firewall setups. Firewalls have limitations and cannot protect against all internal/external threats.
Salesforce Integration for Bonterra Impact Management (fka Social Solutions A...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on integration of Salesforce with Bonterra Impact Management.
Interested in deploying an integration with Salesforce for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Dive into the realm of operating systems (OS) with Pravash Chandra Das, a seasoned Digital Forensic Analyst, as your guide. 🚀 This comprehensive presentation illuminates the core concepts, types, and evolution of OS, essential for understanding modern computing landscapes.
Beginning with the foundational definition, Das clarifies the pivotal role of OS as system software orchestrating hardware resources, software applications, and user interactions. Through succinct descriptions, he delineates the diverse types of OS, from single-user, single-task environments like early MS-DOS iterations, to multi-user, multi-tasking systems exemplified by modern Linux distributions.
Crucial components like the kernel and shell are dissected, highlighting their indispensable functions in resource management and user interface interaction. Das elucidates how the kernel acts as the central nervous system, orchestrating process scheduling, memory allocation, and device management. Meanwhile, the shell serves as the gateway for user commands, bridging the gap between human input and machine execution. 💻
The narrative then shifts to a captivating exploration of prominent desktop OSs, Windows, macOS, and Linux. Windows, with its globally ubiquitous presence and user-friendly interface, emerges as a cornerstone in personal computing history. macOS, lauded for its sleek design and seamless integration with Apple's ecosystem, stands as a beacon of stability and creativity. Linux, an open-source marvel, offers unparalleled flexibility and security, revolutionizing the computing landscape. 🖥️
Moving to the realm of mobile devices, Das unravels the dominance of Android and iOS. Android's open-source ethos fosters a vibrant ecosystem of customization and innovation, while iOS boasts a seamless user experience and robust security infrastructure. Meanwhile, discontinued platforms like Symbian and Palm OS evoke nostalgia for their pioneering roles in the smartphone revolution.
The journey concludes with a reflection on the ever-evolving landscape of OS, underscored by the emergence of real-time operating systems (RTOS) and the persistent quest for innovation and efficiency. As technology continues to shape our world, understanding the foundations and evolution of operating systems remains paramount. Join Pravash Chandra Das on this illuminating journey through the heart of computing. 🌟
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
9. Neutron vs. Nova-network
Legacy networking (Nova-network) OpenStack networking (Neutron)
Simple, single agent Complex, multiple agents
More matured, established Newer, maturing
Flat or VLAN Flat, VLAN, Overlays, L2-L3, SDN
No plug-in support Plug-in support for 3rd parties
Scales well Scaling requires 3rd party plug-ins
No multi-tier topologies Multi-tier topologies
9
10. CEPH Storage: Network Model
● Cluster should have more 2X-4X capacity than Public network 10
12. Rules of Thumb Sizing
● Ceph OSD
○ Disks
■ 8-10 SAS HDDs per 1x 10Gbps
■ ~12 SATA HDDs per 1x 10Gbps
■ 1x SSD for write journal per 4-6 OSD drives
■ JBOD (no RAID configuration)
○ RAM
■ 1GB of RAM per 1 TB of OSD storage space
○ CPU
■ 0.5 CPU core/1 Ghz of a core per OSD disk (1-2 CPU cores for SSD drives)
● Ceph Monitor (min 3 per cluster)
○ 1 GB of RAM per daemon
○ 1 ceph-mon node per 15-20 OSD nodes
12
13. Configuration
● Ceph Storage Cluster
○ Multi-pools:
SSD, SAS/SATA
○ Replicas of objects:
min 2
○ Data protection:
Replicated vs. EC
○ Memory allocator:
TCMalloc vs. JEMalloc
13
● Ceph Block Device (RBD)
○ rbd_cache
● Ceph Object Gateway (RGW)
○ rgw_cache_lru_size
○ rgw_thread_pool_size
○ rgw_override_bucket_index_max_shards
○ LB: LVS (direct routing) vs. HAProxy
○ Web server: Civetweb vs. Nginx vs.
Apache
14. Compute
● Specialized hardwares
○ GPU
○ SSD locality
○ Encryption keys in USB devices
● Scheduling on specific hosts (e.g. KVM, XenServer), hardwares
○ Host Aggregates
● Separating critical workloads from others
○ Availability Zones
14
15. Compute (cont.)
● High performance computing for VMs (*)
○ No overcommit (RAM, CPU)
○ CPU Pinning
○ NUMA Topology awareness
● High availability for VMs
○ Live/Cold Migration
○ Evacuation
○ resume_guests_state_on_host_boot=true
● Scalability for VMs
○ Resize
(*) http://www.slideshare.net/vietstack/vietopenstack-meetup-7th-high-performance-vm
15