At Red Hat Storage Day Minneapolis on 4/12/16, Intel's Dan Ferber presented on Intel storage components, benchmarks, and contributions as they relate to Ceph.
Storage tiering and erasure coding in Ceph (SCaLE13x)Sage Weil
Ceph is designed around the assumption that all components of the system (disks, hosts, networks) can fail, and has traditionally leveraged replication to provide data durability and reliability. The CRUSH placement algorithm is used to allow failure domains to be defined across hosts, racks, rows, or datacenters, depending on the deployment scale and requirements.
Recent releases have added support for erasure coding, which can provide much higher data durability and lower storage overheads. However, in practice erasure codes have different performance characteristics than traditional replication and, under some workloads, come at some expense. At the same time, we have introduced a storage tiering infrastructure and cache pools that allow alternate hardware backends (like high-end flash) to be leveraged for active data sets while cold data are transparently migrated to slower backends. The combination of these two features enables a surprisingly broad range of new applications and deployment configurations.
This talk will cover a few Ceph fundamentals, discuss the new tiering and erasure coding features, and then discuss a variety of ways that the new capabilities can be leveraged.
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...Odinot Stanislas
Après la petite intro sur le stockage distribué et la description de Ceph, Jian Zhang réalise dans cette présentation quelques benchmarks intéressants : tests séquentiels, tests random et surtout comparaison des résultats avant et après optimisations. Les paramètres de configuration touchés et optimisations (Large page numbers, Omap data sur un disque séparé, ...) apportent au minimum 2x de perf en plus.
The document discusses strategies for optimizing Ceph performance at scale. It describes the presenters' typical node configurations, including storage nodes with 72 HDDs and NVME journals, and monitor/RGW nodes. Various techniques are discussed like ensuring proper NUMA alignment of processes, IRQs, and mount points. General tuning tips include using latest drivers, OS tuning, and addressing network issues. The document stresses that monitors can become overloaded during large rebalances and deleting large pools, so more than one monitor is needed for large clusters.
Red Hat Ceph Storage Acceleration Utilizing Flash Technology Red_Hat_Storage
Red Hat Ceph Storage can utilize flash technology to accelerate applications in three ways: 1) use all flash storage for highest performance, 2) use a hybrid configuration with performance critical data on flash tier and colder data on HDD tier, or 3) utilize host caching of critical data on flash. Benchmark results showed that using NVMe SSDs in Ceph provided much higher performance than SATA SSDs, with speed increases of up to 8x for some workloads. However, testing also showed that Ceph may not be well-suited for OLTP MySQL workloads due to small random reads/writes, as local SSD storage outperformed the Ceph cluster. Proper Linux tuning is also needed to maximize SSD performance within
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsColleen Corrice
At Red Hat Storage Day Minneapolis on 4/12/16, Intel's Dan Ferber presented on Intel storage components, benchmarks, and contributions as they relate to Ceph.
Red Hat's storage roadmap aims to offer a unified, open software-defined storage portfolio for next generation workloads. Their portfolio includes Ceph and Gluster storage, which provide data services like object storage, block storage, and file sharing on standard hardware. Red Hat is working to improve management, performance, security, and integration of these products. They are also exploring new workloads like running MySQL on Ceph block storage and using Gluster Storage for hyperconverged Red Hat Enterprise Virtualization environments.
Storage tiering and erasure coding in Ceph (SCaLE13x)Sage Weil
Ceph is designed around the assumption that all components of the system (disks, hosts, networks) can fail, and has traditionally leveraged replication to provide data durability and reliability. The CRUSH placement algorithm is used to allow failure domains to be defined across hosts, racks, rows, or datacenters, depending on the deployment scale and requirements.
Recent releases have added support for erasure coding, which can provide much higher data durability and lower storage overheads. However, in practice erasure codes have different performance characteristics than traditional replication and, under some workloads, come at some expense. At the same time, we have introduced a storage tiering infrastructure and cache pools that allow alternate hardware backends (like high-end flash) to be leveraged for active data sets while cold data are transparently migrated to slower backends. The combination of these two features enables a surprisingly broad range of new applications and deployment configurations.
This talk will cover a few Ceph fundamentals, discuss the new tiering and erasure coding features, and then discuss a variety of ways that the new capabilities can be leveraged.
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...Odinot Stanislas
Après la petite intro sur le stockage distribué et la description de Ceph, Jian Zhang réalise dans cette présentation quelques benchmarks intéressants : tests séquentiels, tests random et surtout comparaison des résultats avant et après optimisations. Les paramètres de configuration touchés et optimisations (Large page numbers, Omap data sur un disque séparé, ...) apportent au minimum 2x de perf en plus.
The document discusses strategies for optimizing Ceph performance at scale. It describes the presenters' typical node configurations, including storage nodes with 72 HDDs and NVME journals, and monitor/RGW nodes. Various techniques are discussed like ensuring proper NUMA alignment of processes, IRQs, and mount points. General tuning tips include using latest drivers, OS tuning, and addressing network issues. The document stresses that monitors can become overloaded during large rebalances and deleting large pools, so more than one monitor is needed for large clusters.
Red Hat Ceph Storage Acceleration Utilizing Flash Technology Red_Hat_Storage
Red Hat Ceph Storage can utilize flash technology to accelerate applications in three ways: 1) use all flash storage for highest performance, 2) use a hybrid configuration with performance critical data on flash tier and colder data on HDD tier, or 3) utilize host caching of critical data on flash. Benchmark results showed that using NVMe SSDs in Ceph provided much higher performance than SATA SSDs, with speed increases of up to 8x for some workloads. However, testing also showed that Ceph may not be well-suited for OLTP MySQL workloads due to small random reads/writes, as local SSD storage outperformed the Ceph cluster. Proper Linux tuning is also needed to maximize SSD performance within
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsColleen Corrice
At Red Hat Storage Day Minneapolis on 4/12/16, Intel's Dan Ferber presented on Intel storage components, benchmarks, and contributions as they relate to Ceph.
Red Hat's storage roadmap aims to offer a unified, open software-defined storage portfolio for next generation workloads. Their portfolio includes Ceph and Gluster storage, which provide data services like object storage, block storage, and file sharing on standard hardware. Red Hat is working to improve management, performance, security, and integration of these products. They are also exploring new workloads like running MySQL on Ceph block storage and using Gluster Storage for hyperconverged Red Hat Enterprise Virtualization environments.
This document outlines an agenda for a session on running MySQL on Ceph storage. The first part will discuss using MySQL on Ceph versus AWS and include a performance head-to-head. The second part will cover Ceph architecture including components like RADOS, pools, and CRUSH algorithm for data placement. The final part will discuss tuning MySQL and Ceph together for optimal performance including adjusting buffer pool size, transaction flushing, and creating specialized pools for IOPS workloads. An accompanying lab will compare MySQL performance on Ceph versus other cloud platforms.
HKG15-401: Ceph and Software Defined Storage on ARM serversLinaro
HKG15-401: Ceph and Software Defined Storage on ARM servers
---------------------------------------------------
Speaker: Yazen Ghannam Steve Capper
Date: February 12, 2015
---------------------------------------------------
★ Session Summary ★
Running Ceph in the colocation, ongoing optimizations
--------------------------------------------------
★ Resources ★
Pathable: https://hkg15.pathable.com/meetings/250828
Video: https://www.youtube.com/watch?v=RdZojLL7ttk
Etherpad: http://pad.linaro.org/p/hkg15-401
---------------------------------------------------
★ Event Details ★
Linaro Connect Hong Kong 2015 - #HKG15
February 9-13th, 2015
Regal Airport Hotel Hong Kong Airport
---------------------------------------------------
http://www.linaro.org
http://connect.linaro.org
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA ArchitectureDanielle Womboldt
This document discusses an all-flash Ceph array design from QCT based on NUMA architecture. It provides an agenda that covers all-flash Ceph and use cases, QCT's all-flash Ceph solution for IOPS, an overview of QCT's lab environment and detailed architecture, and the importance of NUMA. It also includes sections on why all-flash storage is used, different all-flash Ceph use cases, QCT's IOPS-optimized all-flash Ceph solution, benefits of using NVMe storage, QCT's lab test environment, Ceph tuning recommendations, and benefits of using multi-partitioned NVMe SSDs for Ceph OSDs.
The document provides an overview and update on CephFS, the distributed file system implemented in Ceph. Key points include new capabilities in the Jewel release like improved scrub/repair functionality to handle metadata damage, fine-grained authorization using new MDS auth caps, and experimental support for multiple CephFS filesystems backed by a single RADOS storage cluster. It also discusses ongoing work integrating CephFS with OpenStack Manila for shared file system provisioning and new tools like CephFSVolumeClient.
Build an High-Performance and High-Durable Block Storage Service Based on CephRongze Zhu
This document discusses building a high-performance and durable block storage service using Ceph. It describes the architecture, including a minimum deployment of 12 OSD nodes and 3 monitor nodes. It outlines optimizations made to Ceph, Qemu, and the operating system configuration to achieve high performance, including 6000 IOPS and 170MB/s throughput. It also discusses how the CRUSH map can be optimized to reduce recovery times and number of copysets to improve durability to 99.99999999%.
QCT is a global provider of hyperscale datacenter solutions including servers, storage, networking equipment, and integrated rack systems. It aims to deliver the efficiency, scalability, and reliability of hyperscale designs to all datacenter customers using standard open hardware. QCT is a subsidiary of Quanta Computer, a Fortune 500 company, allowing it to leverage over 14 years of experience in datacenter engineering and manufacturing.
Red Hat Storage Day New York - What's New in Red Hat Ceph StorageRed_Hat_Storage
Red Hat Ceph Storage 2.0 will focus on usability improvements like the Red Hat Storage Console management tool, improved object storage features such as global clusters and authentication support, and performance enhancements from the new BlueStore backend. Future releases will add capabilities like block volume mirroring, CephFS support, and integration with OpenStack services. Upstream development will continue expanding Ceph's object, block, and file functionality.
Ceph is an open source distributed storage system designed for scalability and reliability. Ceph's block device, RADOS block device (RBD), is widely used to store virtual machines, and is the most popular block storage used with OpenStack.
In this session, you'll learn how RBD works, including how it:
* Uses RADOS classes to make access easier from user space and within the Linux kernel.
* Implements thin provisioning.
* Builds on RADOS self-managed snapshots for cloning and differential backups.
* Increases performance with caching of various kinds.
* Uses watch/notify RADOS primitives to handle online management operations.
* Integrates with QEMU, libvirt, and OpenStack.
QCT Ceph Solution - Design Consideration and Reference ArchitecturePatrick McGarry
This document discusses QCT's Ceph storage solutions, including an overview of Ceph architecture, QCT hardware platforms, Red Hat Ceph software, workload considerations, reference architectures, test results and a QCT/Red Hat whitepaper. It provides technical details on QCT's throughput-optimized and capacity-optimized solutions and shows how they address different storage needs through workload-driven design. Hands-on testing and a test drive lab are offered to explore Ceph features and configurations.
Ceph at Work in Bloomberg: Object Store, RBD and OpenStackRed_Hat_Storage
Bloomberg's Chris Jones and Chris Morgan joined Red Hat Storage Day New York on 1/19/16 to explain how Red Hat Ceph Storage helps the financial giant tackle its data storage challenges.
This document summarizes a presentation about tuning MySQL performance on Ceph block storage. The presentation covers Ceph architecture, tuning Ceph block devices, and tuning QEMU block virtualization. It then shows benchmarks comparing different configurations for reads, writes, and a 70/30 read/write mix using Sysbench OLTP workloads. Configurations tested include QEMU backends, caching modes, I/O threading, virtio types and queues, and containers versus metal. The goal is to understand how to optimize MySQL on Ceph block devices.
- Librados is a C/C++ programming interface that provides applications access to the Ceph distributed object store (RADOS) and hides the complexity of networking, data distribution, replication and failure recovery.
- It can be used by Ceph components like RADOS Gateway and tools like rados, as well as third party applications that want to use Ceph for storage. Examples include providing storage for mail systems, Hadoop, and building custom applications.
- The interface handles configuration, connections, I/O operations on objects, extended attributes and more so applications can easily integrate scalable reliable storage via Ceph without having to implement these functions themselves.
BlueStore: a new, faster storage backend for CephSage Weil
BlueStore is a new storage backend for Ceph that provides faster performance compared to the existing FileStore backend. BlueStore stores metadata in RocksDB and data directly on block devices, avoiding double writes and improving transaction performance. It supports multiple storage tiers by allowing different components like the RocksDB WAL, database and object data to be placed on SSDs, HDDs or NVRAM as appropriate.
Ceph - High Performance Without High CostsJonathan Long
Ceph is a high-performance storage platform that provides storage without high costs. The presentation discusses BlueStore, a redesign of Ceph's object store to improve performance and efficiency. BlueStore preserves wire compatibility but uses an incompatible storage format. It aims to double write performance and match or exceed read performance of the previous FileStore design. BlueStore simplifies the architecture and uses algorithms tailored for different hardware like flash. It was in a tech preview in the Jewel release and aims to be default in the Luminous release next year.
Ceph scale testing with 10 Billion ObjectsKaran Singh
In this performance testing, we ingested 10 Billion objects into the Ceph Object Storage system and measured its performance. We have observed deterministic performance, check out this presentation to know the details.
CephFS performance testing was conducted on a Jewel deployment. Key findings include:
- Single MDS performance is limited by its single-threaded design; operations reached CPU limits
- Improper client behavior can cause MDS OOM issues by exceeding inode caching limits
- Metadata operations like create, open, update showed similar performance, reaching 4-5k ops/sec maximum
- Caching had a large impact on performance when the working set exceeded cache size
Ceph Intro and Architectural Overview by Ross Turkbuildacloud
This document provides an overview of Ceph architecture and introduces its key components. Ceph is an open-source distributed storage system that provides object, block, and file storage in a single unified storage cluster. It utilizes several unique technologies, including CRUSH for distributed object mapping, thin provisioning for efficient storage of block devices, and dynamic subtree partitioning for clustered metadata management. The document outlines Ceph's core components like RADOS, Librados, RBD, CephFS, and RADOS Gateway and how they work together to provide a scalable, reliable and distributed storage platform.
This document discusses userspace storage systems as an alternative to kernel-based storage for petascale workloads. It outlines several userspace filesystems, block storage systems, and object storage systems used in practice. Common languages used include C, Python, Java, and Golang. Interfaces to the kernel include FUSE, UIO, DPDK and libvma. Challenges include balancing performance, scalability, and complexity across unified, self-managing systems. Specific examples covered are NFS-Ganesha, GlusterFS, HDFS, NBD, tgt, and caching systems like Tachyon and Redis.
Red hat Storage Day LA - Designing Ceph Clusters Using Intel-Based HardwareRed_Hat_Storage
This document discusses how data growth driven by mobile, social media, IoT, and big data/cloud is requiring a fundamental shift in storage cost structures from scale-up to scale-out architectures. It provides an overview of key storage technologies and workloads driving public cloud storage, and how Ceph can help deliver on the promise of the cloud by providing next generation storage architectures with flash to enable new capabilities in small footprints. It also illustrates the wide performance range Ceph can provide for different workloads and hardware configurations.
Ceph, Now and Later: Our Plan for Open Unified Cloud StorageSage Weil
Ceph is a highly scalable open source distributed storage system that provides object, block, and file interfaces on a single platform. Although Ceph RBD block storage has dominated OpenStack deployments for several years, maturing object (S3, Swift, and librados) interfaces and stable CephFS (file) interfaces now make Ceph the only fully open source unified storage platform.
This talk will cover Ceph's architectural vision and project mission and how our approach differs from alternative approaches to storage in the OpenStack ecosystem. In particular, we will look at how our open development model dovetails well with OpenStack, how major contributors are advancing Ceph capabilities and performance at a rapid pace to adapt to new hardware types and deployment models, and what major features we are priotizing for the next few years to meet the needs of expanding cloud workloads.
This document outlines an agenda for a session on running MySQL on Ceph storage. The first part will discuss using MySQL on Ceph versus AWS and include a performance head-to-head. The second part will cover Ceph architecture including components like RADOS, pools, and CRUSH algorithm for data placement. The final part will discuss tuning MySQL and Ceph together for optimal performance including adjusting buffer pool size, transaction flushing, and creating specialized pools for IOPS workloads. An accompanying lab will compare MySQL performance on Ceph versus other cloud platforms.
HKG15-401: Ceph and Software Defined Storage on ARM serversLinaro
HKG15-401: Ceph and Software Defined Storage on ARM servers
---------------------------------------------------
Speaker: Yazen Ghannam Steve Capper
Date: February 12, 2015
---------------------------------------------------
★ Session Summary ★
Running Ceph in the colocation, ongoing optimizations
--------------------------------------------------
★ Resources ★
Pathable: https://hkg15.pathable.com/meetings/250828
Video: https://www.youtube.com/watch?v=RdZojLL7ttk
Etherpad: http://pad.linaro.org/p/hkg15-401
---------------------------------------------------
★ Event Details ★
Linaro Connect Hong Kong 2015 - #HKG15
February 9-13th, 2015
Regal Airport Hotel Hong Kong Airport
---------------------------------------------------
http://www.linaro.org
http://connect.linaro.org
Ceph Day Beijing - Ceph All-Flash Array Design Based on NUMA ArchitectureDanielle Womboldt
This document discusses an all-flash Ceph array design from QCT based on NUMA architecture. It provides an agenda that covers all-flash Ceph and use cases, QCT's all-flash Ceph solution for IOPS, an overview of QCT's lab environment and detailed architecture, and the importance of NUMA. It also includes sections on why all-flash storage is used, different all-flash Ceph use cases, QCT's IOPS-optimized all-flash Ceph solution, benefits of using NVMe storage, QCT's lab test environment, Ceph tuning recommendations, and benefits of using multi-partitioned NVMe SSDs for Ceph OSDs.
The document provides an overview and update on CephFS, the distributed file system implemented in Ceph. Key points include new capabilities in the Jewel release like improved scrub/repair functionality to handle metadata damage, fine-grained authorization using new MDS auth caps, and experimental support for multiple CephFS filesystems backed by a single RADOS storage cluster. It also discusses ongoing work integrating CephFS with OpenStack Manila for shared file system provisioning and new tools like CephFSVolumeClient.
Build an High-Performance and High-Durable Block Storage Service Based on CephRongze Zhu
This document discusses building a high-performance and durable block storage service using Ceph. It describes the architecture, including a minimum deployment of 12 OSD nodes and 3 monitor nodes. It outlines optimizations made to Ceph, Qemu, and the operating system configuration to achieve high performance, including 6000 IOPS and 170MB/s throughput. It also discusses how the CRUSH map can be optimized to reduce recovery times and number of copysets to improve durability to 99.99999999%.
QCT is a global provider of hyperscale datacenter solutions including servers, storage, networking equipment, and integrated rack systems. It aims to deliver the efficiency, scalability, and reliability of hyperscale designs to all datacenter customers using standard open hardware. QCT is a subsidiary of Quanta Computer, a Fortune 500 company, allowing it to leverage over 14 years of experience in datacenter engineering and manufacturing.
Red Hat Storage Day New York - What's New in Red Hat Ceph StorageRed_Hat_Storage
Red Hat Ceph Storage 2.0 will focus on usability improvements like the Red Hat Storage Console management tool, improved object storage features such as global clusters and authentication support, and performance enhancements from the new BlueStore backend. Future releases will add capabilities like block volume mirroring, CephFS support, and integration with OpenStack services. Upstream development will continue expanding Ceph's object, block, and file functionality.
Ceph is an open source distributed storage system designed for scalability and reliability. Ceph's block device, RADOS block device (RBD), is widely used to store virtual machines, and is the most popular block storage used with OpenStack.
In this session, you'll learn how RBD works, including how it:
* Uses RADOS classes to make access easier from user space and within the Linux kernel.
* Implements thin provisioning.
* Builds on RADOS self-managed snapshots for cloning and differential backups.
* Increases performance with caching of various kinds.
* Uses watch/notify RADOS primitives to handle online management operations.
* Integrates with QEMU, libvirt, and OpenStack.
QCT Ceph Solution - Design Consideration and Reference ArchitecturePatrick McGarry
This document discusses QCT's Ceph storage solutions, including an overview of Ceph architecture, QCT hardware platforms, Red Hat Ceph software, workload considerations, reference architectures, test results and a QCT/Red Hat whitepaper. It provides technical details on QCT's throughput-optimized and capacity-optimized solutions and shows how they address different storage needs through workload-driven design. Hands-on testing and a test drive lab are offered to explore Ceph features and configurations.
Ceph at Work in Bloomberg: Object Store, RBD and OpenStackRed_Hat_Storage
Bloomberg's Chris Jones and Chris Morgan joined Red Hat Storage Day New York on 1/19/16 to explain how Red Hat Ceph Storage helps the financial giant tackle its data storage challenges.
This document summarizes a presentation about tuning MySQL performance on Ceph block storage. The presentation covers Ceph architecture, tuning Ceph block devices, and tuning QEMU block virtualization. It then shows benchmarks comparing different configurations for reads, writes, and a 70/30 read/write mix using Sysbench OLTP workloads. Configurations tested include QEMU backends, caching modes, I/O threading, virtio types and queues, and containers versus metal. The goal is to understand how to optimize MySQL on Ceph block devices.
- Librados is a C/C++ programming interface that provides applications access to the Ceph distributed object store (RADOS) and hides the complexity of networking, data distribution, replication and failure recovery.
- It can be used by Ceph components like RADOS Gateway and tools like rados, as well as third party applications that want to use Ceph for storage. Examples include providing storage for mail systems, Hadoop, and building custom applications.
- The interface handles configuration, connections, I/O operations on objects, extended attributes and more so applications can easily integrate scalable reliable storage via Ceph without having to implement these functions themselves.
BlueStore: a new, faster storage backend for CephSage Weil
BlueStore is a new storage backend for Ceph that provides faster performance compared to the existing FileStore backend. BlueStore stores metadata in RocksDB and data directly on block devices, avoiding double writes and improving transaction performance. It supports multiple storage tiers by allowing different components like the RocksDB WAL, database and object data to be placed on SSDs, HDDs or NVRAM as appropriate.
Ceph - High Performance Without High CostsJonathan Long
Ceph is a high-performance storage platform that provides storage without high costs. The presentation discusses BlueStore, a redesign of Ceph's object store to improve performance and efficiency. BlueStore preserves wire compatibility but uses an incompatible storage format. It aims to double write performance and match or exceed read performance of the previous FileStore design. BlueStore simplifies the architecture and uses algorithms tailored for different hardware like flash. It was in a tech preview in the Jewel release and aims to be default in the Luminous release next year.
Ceph scale testing with 10 Billion ObjectsKaran Singh
In this performance testing, we ingested 10 Billion objects into the Ceph Object Storage system and measured its performance. We have observed deterministic performance, check out this presentation to know the details.
CephFS performance testing was conducted on a Jewel deployment. Key findings include:
- Single MDS performance is limited by its single-threaded design; operations reached CPU limits
- Improper client behavior can cause MDS OOM issues by exceeding inode caching limits
- Metadata operations like create, open, update showed similar performance, reaching 4-5k ops/sec maximum
- Caching had a large impact on performance when the working set exceeded cache size
Ceph Intro and Architectural Overview by Ross Turkbuildacloud
This document provides an overview of Ceph architecture and introduces its key components. Ceph is an open-source distributed storage system that provides object, block, and file storage in a single unified storage cluster. It utilizes several unique technologies, including CRUSH for distributed object mapping, thin provisioning for efficient storage of block devices, and dynamic subtree partitioning for clustered metadata management. The document outlines Ceph's core components like RADOS, Librados, RBD, CephFS, and RADOS Gateway and how they work together to provide a scalable, reliable and distributed storage platform.
This document discusses userspace storage systems as an alternative to kernel-based storage for petascale workloads. It outlines several userspace filesystems, block storage systems, and object storage systems used in practice. Common languages used include C, Python, Java, and Golang. Interfaces to the kernel include FUSE, UIO, DPDK and libvma. Challenges include balancing performance, scalability, and complexity across unified, self-managing systems. Specific examples covered are NFS-Ganesha, GlusterFS, HDFS, NBD, tgt, and caching systems like Tachyon and Redis.
Red hat Storage Day LA - Designing Ceph Clusters Using Intel-Based HardwareRed_Hat_Storage
This document discusses how data growth driven by mobile, social media, IoT, and big data/cloud is requiring a fundamental shift in storage cost structures from scale-up to scale-out architectures. It provides an overview of key storage technologies and workloads driving public cloud storage, and how Ceph can help deliver on the promise of the cloud by providing next generation storage architectures with flash to enable new capabilities in small footprints. It also illustrates the wide performance range Ceph can provide for different workloads and hardware configurations.
Ceph, Now and Later: Our Plan for Open Unified Cloud StorageSage Weil
Ceph is a highly scalable open source distributed storage system that provides object, block, and file interfaces on a single platform. Although Ceph RBD block storage has dominated OpenStack deployments for several years, maturing object (S3, Swift, and librados) interfaces and stable CephFS (file) interfaces now make Ceph the only fully open source unified storage platform.
This talk will cover Ceph's architectural vision and project mission and how our approach differs from alternative approaches to storage in the OpenStack ecosystem. In particular, we will look at how our open development model dovetails well with OpenStack, how major contributors are advancing Ceph capabilities and performance at a rapid pace to adapt to new hardware types and deployment models, and what major features we are priotizing for the next few years to meet the needs of expanding cloud workloads.
BlueStore: a new, faster storage backend for CephSage Weil
Traditionally Ceph has made use of local file systems like XFS or btrfs to store its data. However, the mismatch between the OSD's requirements and the POSIX interface provided by kernel file systems has a huge performance cost and requires a lot of complexity. BlueStore, an entirely new OSD storage backend, utilizes block devices directly, doubling performance for most workloads. This talk will cover the motivation a new backend, the design and implementation, the improved performance on HDDs, SSDs, and NVMe, and discuss some of the thornier issues we had to overcome when replacing tried and true kernel file systems with entirely new code running in userspace.
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...Patrick McGarry
This document discusses using recently published Ceph reference architectures to select a Ceph configuration. It provides an inventory of existing reference architectures from Red Hat and SUSE. It previews highlights from an upcoming Intel and Red Hat Ceph reference architecture paper, including recommended configurations and hardware. It also describes an Intel all-NVMe Ceph benchmark configuration for MySQL workloads. In summary, reference architectures provide guidelines for building optimized Ceph solutions based on specific workloads and use cases.
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSCeph Community
This document summarizes the performance of an all-NVMe Ceph cluster using Intel P3700 NVMe SSDs. Key results include achieving over 1.35 million 4K random read IOPS and 171K 4K random write IOPS with sub-millisecond latency. Partitioning the NVMe drives into multiple OSDs improved performance and CPU utilization compared to a single OSD per drive. The cluster also demonstrated over 5GB/s of sequential bandwidth.
This document summarizes new features and upcoming releases for Ceph. In the Jewel release in April 2016, CephFS became more stable with improvements to repair and disaster recovery tools. The BlueStore backend was introduced experimentally to replace Filestore. Future releases Kraken and Luminous will include multi-active MDS support for CephFS, erasure code overwrites for RBD, management tools, and continued optimizations for performance and scalability.
Red Hat Storage Day Boston - OpenStack + Ceph StorageRed_Hat_Storage
- Red Hat OpenStack Platform delivers an integrated and production-ready OpenStack cloud platform that combines Red Hat's hardened OpenStack infrastructure which is co-engineered with Red Hat Enterprise Linux.
- Ceph is an open-source, massively scalable software-defined storage that provides a single, efficient, and unified storage platform on clustered commodity hardware. Ceph is flexible and can provide block, object, and file-level storage for OpenStack.
- Architectures using OpenStack and Ceph include hyperconverged infrastructure which co-locates compute and storage on the same machines, and multi-site configurations with replicated Ceph storage across sites for disaster recovery.
Red Hat Storage Day Dallas - Defiance of the Appliance Red_Hat_Storage
The document discusses the challenges with traditional enterprise storage and the benefits of software-defined storage using Red Hat Gluster Storage and Ceph. It highlights how software-defined storage provides near linear performance scaling, lower total cost of ownership, open source innovation, container-native storage, and freedom from vendor lock-in compared to traditional proprietary storage systems.
Red Hat Storage Day New York - Intel Unlocking Big Data Infrastructure Effici...Red_Hat_Storage
This document discusses using Ceph storage with Apache Hadoop to provide a scalable and efficient storage solution for big data workloads. It outlines the challenges of scaling Hadoop storage independently from compute resources using the native Hadoop Distributed File System. The solution presented is to use the open source Ceph storage system instead of direct-attached storage. This allows Hadoop compute and storage resources to scale independently and provides a centralized storage platform for all enterprise data workloads. Performance tests showed the Ceph and Hadoop configuration providing up to a 60% improvement in I/O performance when using Intel caching software and SSDs.
The document provides an introduction to NVMe over Fabrics, including:
- What NVMe over Fabrics is and its advantages like end-to-end NVMe semantics and low latency remote storage.
- How NVMe is being expanded to support message-based operations over various fabrics like RDMA, Fibre Channel, and Ethernet.
- Examples of how NVMe over Fabrics is being implemented in data center architectures and storage solutions.
Red Hat Storage Day Dallas - Storage for OpenShift Containers Red_Hat_Storage
This document discusses using Red Hat Gluster Storage for persistent storage of OpenShift containers. It describes how containers improve software development and management. Containers provide more efficient use of resources than virtual machines. Red Hat Gluster Storage provides scalable, distributed storage optimized for container environments. It can be deployed on-premises or in the cloud and integrated with OpenShift to offer storage as a service for containerized applications.
Ceph Day Taipei - Delivering cost-effective, high performance, Ceph cluster Ceph Community
An all-NVMe Ceph cluster was configured with 5 storage nodes, each containing 20 Intel P3700 SSDs providing a total of 80 object storage daemons (OSDs). Benchmarking showed over 1.4 million 4K random read IOPS at an average latency of 1ms and 220K 4K random write IOPS at 5ms latency. For a 70/30% read/write mix, over 560K random IOPS were achieved at 3ms latency. Sysbench MySQL testing on the cluster showed linear scaling from 2 to 8 queries per second with an average I/O size of 16KB.
When compared to the no-cache solution results, the Intel solution increased database performance in our tests by up to 65 percent using a real-world database workload. Additionally, the Intel solution provided 7 percent greater database performance when compared to the competing solution. This means that with the Intel solution, administrators can give their databases an extra boost to improve performance for end users, especially when RAM upgrades and additional storage are no longer feasible options.
Covers the problems of achieving scalability in server farm environments and how distributed data grids provide in-memory storage and boost performance. Includes summary of ScaleOut Software product offerings including ScaleOut State Server and Grid Computing Edition.
DPDK Summit 2015 - RIFT.io - Tim MortsolfJim St. Leger
DPDK Summit 2015 in San Francisco.
Presentation by RIFT.io's CTO Tim Mortsolf.
For additional details and the video recording please visit www.dpdksummit.com.
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...Red_Hat_Storage
This document discusses the need for storage modernization driven by trends like mobile, social media, IoT and big data. It outlines how scale-out architectures using open source Ceph software can help meet this need more cost effectively than traditional scale-up storage. Specific optimizations for IOPS, throughput and capacity are described. Intel is presented as helping advance the industry through open source contributions and optimized platforms, software and SSD technologies. Real-world examples are given showing the wide performance range Ceph can provide.
The document compares the performance of Ceph storage cluster using TCP and RDMA (XIO) as the transport mechanisms. It finds that XIO provides around 30-50% higher IOPS and bandwidth compared to TCP with the same hardware setup. However, TCP performance is improving and catching up to XIO as the number of OSDs increases. While XIO provides better CPU utilization, it requires over 2x more memory usage than TCP. Scaling out to multiple nodes shows TCP scaling better than XIO. XIO performance is also unstable and connection startup times are longer compared to TCP.
Ceph Day Beijing - Storage Modernization with Intel and CephDanielle Womboldt
The document discusses trends in data growth and storage technologies that are driving the need for storage modernization. It outlines Intel's role in advancing the storage industry through open source technologies and standards. A significant portion of the document focuses on Intel's work optimizing Ceph for Intel platforms, including profiling and benchmarking Ceph performance on Intel SSDs, 3D XPoint, and Optane drives.
Ceph Day Beijing - Storage Modernization with Intel & Ceph Ceph Community
The document discusses trends in data growth and storage technologies that are driving the need for storage modernization. It outlines Intel's role in advancing the storage industry through open source technologies and standards. Specifically, it focuses on Intel's work optimizing Ceph for Intel platforms, including performance profiling, enabling Intel optimized solutions, and end customer proofs-of-concept using Ceph with Intel SSDs, Optane, and platforms.
In-Memory and TimeSeries Technology to Accelerate NoSQL Analyticssandor szabo
The ability of Informix to combine the in
-
memor
y
performance of Informix Warehouse Accelerator
and the flexibility of TimeSeries and NoSQL
analytics positions it to be ready for the IoT era.
1. The document introduces the Intel Xeon Scalable platform, which provides the foundation for data center innovation with a 1.65x average performance boost over previous generations.
2. It highlights key advantages of the platform including scalable performance, agility in rapid service delivery, and hardware-enhanced security with near-zero performance overhead.
3. Various workload-optimized solutions are discussed that leverage the platform's performance to accelerate insights from analytics, deploy cloud infrastructure more quickly, and transform networks.
The document discusses accelerating Ceph storage performance using SPDK. SPDK introduces optimizations like asynchronous APIs, userspace I/O stacks, and polling mode drivers to reduce software overhead and better utilize fast storage devices. This allows Ceph to better support high performance networks and storage like NVMe SSDs. The document provides an example where SPDK helped XSKY's BlueStore object store achieve significant performance gains over the standard Ceph implementation.
Fujitsu World Tour 2017 - Compute Platform For The Digital WorldFujitsu India
Fujitsu has decades of experience designing and manufacturing servers. Their PRIMERGY servers are known for best-in-class quality that ensures continuous operation with almost no unplanned downtimes. This is achieved through rigorous testing and manufacturing processes in their state-of-the-art factories in Germany. Fujitsu's demand-driven manufacturing approach allows them to produce servers flexibly based on current orders, enabling fast response times and fulfilling individual customer requests.
Join us for an exciting and informative preview of the broadest range of next-generation systems optimized for tomorrow’s data center workloads, Powered by 4th Gen Intel® Xeon® Scalable Processors (formerly codenamed Sapphire Rapids).
Experts from Supermicro and Intel will discuss how the upcoming Supermicro X13 systems will enable new performance levels utilizing state-of-the-art technology, including DDR5, PCIe 5.0, Compute Express Link™ 1.1, and Intel® Advanced Matrix Extensions (Intel AMX).
DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergenceinside-BigData.com
In this deck, Johann Lombardi from Intel presents: DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergence.
"Intel has been building an entirely open source software ecosystem for data-centric computing, fully optimized for Intel® architecture and non-volatile memory (NVM) technologies, including Intel Optane DC persistent memory and Intel Optane DC SSDs. Distributed Asynchronous Object Storage (DAOS) is the foundation of the Intel exascale storage stack. DAOS is an open source software-defined scale-out object store that provides high bandwidth, low latency, and high I/O operations per second (IOPS) storage containers to HPC applications. It enables next-generation data-centric workflows that combine simulation, data analytics, and AI."
Unlike traditional storage stacks that were primarily designed for rotating media, DAOS is architected from the ground up to make use of new NVM technologies, and it is extremely lightweight because it operates end-to-end in user space with full operating system bypass. DAOS offers a shift away from an I/O model designed for block-based, high-latency storage to one that inherently supports fine- grained data access and unlocks the performance of next- generation storage technologies.
Watch the video: https://youtu.be/wnGBW31yhLM
Learn more: https://www.intel.com/content/www/us/en/high-performance-computing/daos-high-performance-storage-brief.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Webinář: Dell VRTX - datacentrum vše-v-jednom za skvělou cenu / 7.10.2013Jaroslav Prodelal
Dokážete si představit, že byste provozovali své datacentrum v prostředí kanceláře? Ano, je to možné. Společnost Dell uvedla na trh novinku v podobě tzv. datacenter-in-a-box (vše-v-jednom), které je optimalizované (odhlučnění, napájení) pro provoz i v kanceláři, samozřejmě jej můžete dát i do samostatné místnosti.
Dell VRTX kombinuje v jediném 5U šasí výpočetní výkon (až 4 2-CPU servery), diskové úložiště (až 24 HDD) a síť.
Ve webináři vás seznámíme s touto cenově velmi zajímavou novinkou a ukážeme rozdíl mezi tímto řešením a případnými alternativami v době samostaných serverů, diskového pole a síťových switchů.
Agenda:
* co je Dell VRTX?
* segment zákazníků pro VRTX
* co VRTX nabízí
* řešení provozované na VRTX
* technické specifikace
* možná použití
* cena
* aktuální nabídky a promo akce
The document discusses reimagining the datacenter through software defined infrastructure. This allows datacenters to become more dynamic, automated and efficient by treating compute, storage and networking resources as composable blocks that can be allocated on demand. This approach breaks down traditional silos and allows simpler deployment and maintenance while improving agility, automation and efficiency. The software defined approach is compared to the traditional rigid infrastructure model and examples are given of how it can improve provisioning times, utilization rates and flexibility.
Accelerating Mission Critical Transformation at Red Hat Summit 2011Pauline Nist
This document discusses accelerating mission critical workloads by migrating them from legacy and proprietary systems to open standard x86 platforms based on Intel Xeon processors. It provides an overview of how Intel is enabling this transition through improved performance, reliability, and security features in recent Xeon generations, as well as growing ecosystem support. Analyst reports and customer quotes are presented showing the migration of mission critical workloads from RISC/UNIX platforms to Xeon, driven by lower costs and comparable capabilities.
Intel's Data Center & Connected Systems Group and Diane Bryant shares the latest news on the latest Intel Xeon E5v2 family of processors and technologies like Intel Network Builders to enable the re-architecture of the Data Center.
Noile solutii Intel pentru afaceri eficiente-tm-20mai2010Agora Group
This document discusses Intel technologies for efficient IT infrastructures and provides the following key points:
1. Intel has fabrication facilities around the world producing microchips on advanced technology nodes.
2. Intel's Core processor family delivers scalable performance across devices from netbooks to servers through Intel architecture.
3. Intel's new 2010 Core processors feature technologies like Turbo Boost for intelligent performance and integrated graphics.
Noile soluţii Intel pentru afaceri eficiente-23apr2010Agora Group
This document discusses Intel technologies for efficient IT infrastructures and provides the following key points:
1. Intel has fabrication facilities around the world producing microchips on advanced technology nodes.
2. Intel's Core processor family delivers scalable performance across devices from netbooks to servers through Intel architecture.
3. Intel's new 2010 Core processors feature technologies like Turbo Boost for intelligent performance and integrated graphics.
Noile soluţii Intel pentru afaceri eficiente-20apr2010Agora Group
This document discusses Intel technologies for efficient IT infrastructures and provides the following key points:
1. Intel has fabrication facilities around the world producing microchips on advanced technology nodes.
2. Intel's Core processor family delivers scalable performance across devices from netbooks to servers through Intel architecture.
3. Intel's new 2010 Core processor family features improved performance and energy efficiency through technologies like Turbo Boost.
Noile tehnologii INTEL pentru infrastructuri IT eficiente-19mar2010Agora Group
This document discusses Intel technologies for efficient IT infrastructures and provides the following key points:
1. Intel has fabrication facilities around the world producing microchips on advanced technology nodes.
2. Intel's Core processor family delivers scalable performance across devices from netbooks to servers through Intel architecture.
3. Intel's new 2010 Core processors feature technologies like Turbo Boost for intelligent performance and integrated graphics.
Ceph Day Shanghai - SSD/NVM Technology Boosting Ceph Performance Ceph Community
This document discusses using SSDs and emerging non-volatile memory technologies like 3D XPoint to boost performance of Ceph storage clusters. It outlines how SSDs can be used as journals and caches to significantly increase throughput and reduce latency compared to HDD-only clusters. A case study from Yahoo showed that using Intel NVMe SSDs with caching software delivered over 2x throughput and half the latency with only 5% of data cached. Future technologies like 3D NAND and 3D XPoint will allow building higher performance, higher capacity SSDs that could extend the use of Ceph.
Similar to Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions (20)
Red Hat Storage Day Dallas - Red Hat Ceph Storage Acceleration Utilizing Flas...Red_Hat_Storage
Red Hat Ceph Storage can utilize flash technology to accelerate applications in three ways: 1) utilize flash caching to accelerate critical data writes and reads, 2) utilize storage tiering to place performance critical data on flash and less critical data on HDDs, and 3) utilize all-flash storage to accelerate performance when all data is critical or caching/tiering cannot be used. The document then discusses best practices for leveraging NVMe SSDs versus SATA SSDs in Ceph configurations and optimizing Linux settings.
Red Hat Storage Day Dallas - Gluster Storage in Containerized Application Red_Hat_Storage
The document discusses using Gluster Storage to provide storage for containerized applications in a Kubernetes cluster. It outlines the challenges of replatforming an ecommerce site to use open source technologies, applying RAS(S) principles, and having a scalable and fault-tolerant solution. The plan is to use Docker containers, Kubernetes for orchestration, and GlusterFS storage. GlusterFS provides highly available, replicated storage across all Kubernetes nodes to support the storage needs of containerized applications.
Red Hat Storage Day Dallas - Why Software-defined Storage MattersRed_Hat_Storage
This document discusses the evolution of storage from traditional appliances to software-defined storage. It notes that many IT decision makers find current storage capabilities inadequate and unable to handle emerging workloads. Traditional appliances face issues like vendor lock-in, lack of flexibility, and high costs. Public cloud storage is more scalable but still has complexity and limitations. The document then introduces software-defined storage as an open solution with standardized platforms that addresses these issues through increased cost efficiency, provisioning speed, and deployment options with less vendor lock-in and skill requirements. It describes Red Hat's portfolio of Ceph and Gluster open source software-defined storage solutions and their target use cases.
Red Hat Storage Day Boston - Why Software-defined Storage MattersRed_Hat_Storage
Software-defined storage is an approach to data storage that uses software to control physical storage infrastructure and manages it as a unified pool of storage. This provides several advantages over traditional proprietary storage, including using standard hardware, centralized management, scale-out architectures, and open source software. Red Hat offers Red Hat Ceph Storage and Red Hat Gluster Storage, which provide software-defined storage solutions that are more flexible, cost-effective, and scalable than traditional storage appliances.
Red Hat Storage Day Boston - Supermicro Super StorageRed_Hat_Storage
The document discusses Supermicro's evolution from server and storage innovation to total solution innovation. It provides examples of their all-flash storage servers and Red Hat Ceph reference architectures using Supermicro hardware. The document also discusses optimizing hardware configurations for different workloads and summarizes Supermicro's portfolio of Ceph-ready nodes and turnkey storage solutions.
Red Hat Storage Day Boston - Persistent Storage for Containers Red_Hat_Storage
Persistent storage is important for containerized applications. Red Hat provides container-ready storage using Red Hat Gluster Storage which provides scalable, distributed file storage for containers. It allows storage and containers to coexist on the same hardware, improving utilization and lowering costs. Red Hat Gluster Storage is optimized to provide container-native storage on OpenShift for workloads like databases to get the benefits of containers while ensuring persistent storage.
Red Hat Storage Day Boston - Red Hat Gluster Storage vs. Traditional Storage ...Red_Hat_Storage
Red Hat Gluster Storage provides a software-defined storage solution that is more cost efficient and flexible than traditional storage appliances. It leverages standard x86 hardware and has open source architecture with no vendor lock-in. A comparison shows Gluster Storage outperforms EMC Isilon on factors like cost, scalability, data protection methods, access protocols, and management capabilities. Gluster Storage is positioned to go beyond traditional storage by supporting containers, disaster recovery in cloud environments, and its roadmap includes additional advanced features.
Red Hat Storage Day New York - Red Hat Gluster Storage: Historical Tick Data ...Red_Hat_Storage
Red Hat Gluster Storage is a software-defined, distributed, scale-out file storage solution that is cost-efficient, high performing at scale, and easy to deploy, manage and scale in public, private and hybrid cloud environments. It offers mature NFS, SMB and HDFS interfaces for enterprise applications such as analytics, media streaming, active archives and enterprise virtualization. The document discusses using Red Hat Gluster Storage for historical tick data repositories, including its architecture, benefits over traditional storage solutions, and analytics workflows.
Red Hat Storage Day New York - QCT: Avoid the mess, deploy with a validated s...Red_Hat_Storage
This document provides an overview of QCT's validated Red Hat Ceph and Gluster storage solutions. QCT offers pre-configured and optimized storage appliances built with Red Hat Ceph and Gluster storage software. Their QxStor solutions include different configurations optimized for throughput, capacity, or IOPS. The document discusses QCT's testing results showing the performance and scalability of their Ceph and Gluster solutions. It also describes QCT's center of excellence where they collaborate with partners to test and develop new solutions.
Red Hat Storage Day - When the Ceph Hits the FanRed_Hat_Storage
This document discusses common issues that can cause a Ceph cluster to fail or experience performance problems ("hitting the fan"). It outlines seven common trouble areas: using unsupported upstream bits or features in production, unsupported configurations, poor cluster growth management, lack of skills/practices, risky configuration choices, poor network configuration, and failure to plan implementations carefully. The document provides recommendations to avoid problems, such as using supported releases, training staff properly, consulting experts for design/planning, and performing regular health checks. It promotes engaging Red Hat support and services to assist with design, implementation and issue resolution.
Red Hat Storage Day New York - Penguin Computing Spotlight: Delivering Open S...Red_Hat_Storage
This document discusses Penguin Computing's open solutions utilizing Red Hat Storage. It describes Penguin Computing as providing compute, storage, and networking solutions using open technologies. It then discusses various Penguin Computing solutions like the Tundra Extreme Scale open compute platform, Arc5ca Ethernet switches, and FrostByte HS storage appliances. The document also summarizes Red Hat Gluster Storage benefits for financial data analytics like deeper analysis, lower costs, and better performance compared to traditional storage solutions.
Red Hat Storage Day New York - New Reference ArchitecturesRed_Hat_Storage
The document provides an overview and summary of Red Hat's reference architecture work including MySQL and Hadoop, software-defined NAS, and digital media repositories. It discusses trends toward disaggregating Hadoop compute and storage and various data flow options. It also summarizes performance testing Red Hat conducted comparing AWS EBS and Ceph for MySQL workloads, and analyzing factors like IOPS/GB ratios, core-to-flash ratios, and pricing. Server categories and vendor examples are defined. Comparisons of throughput and costs at scale between software-defined scale-out storage and traditional enterprise NAS solutions are also presented.
Red Hat Storage Day New York - Persistent Storage for ContainersRed_Hat_Storage
Red Hat Gluster Storage provides persistent container storage for OpenShift. It has evolved from container-ready (running outside containers) to container-native (running inside containers). The current and upcoming versions provide dynamic storage provisioning without admin intervention, improved usability, and support for database workloads through non-shared storage. A demo shows deploying Gluster Storage containers in OpenShift and creating a new persistent volume claim for an application.
Red Hat Storage Day New York -Performance Intensive Workloads with Samsung NV...Red_Hat_Storage
This document discusses using Samsung NVMe SSDs and Red Hat Ceph storage to create a high performance storage tier for OpenStack environments. It presents a reference architecture using a 3-node Ceph cluster with Samsung NVMe SSDs that achieved over 28GB/s for sequential reads. This architecture provides scalable, open source storage optimized for performance-intensive workloads like databases, analytics, and networking. Future work is discussed to develop a similar architecture using GlusterFS storage.
Red Hat Storage Day New York - Welcome Remarks Red_Hat_Storage
Red Hat's presentation discusses changes in the storage industry and Red Hat's storage portfolio. Specifically:
- The storage needs of organizations are outpacing the capabilities of traditional storage solutions, with many IT decision makers reporting that their current storage cannot handle emerging workloads.
- The datacenter is changing with new development models like Agile and DevOps, application architectures like microservices, and deployment methods like containers and hybrid cloud. This is disrupting the storage industry.
- Red Hat's storage portfolio includes the open source Gluster and Ceph storage systems, which provide a scale-out, self-managing architecture supported by Red Hat across physical, virtual, private cloud, container and public cloud environments.
Red Hat Storage Day Seattle: Why Software-Defined Storage MattersRed_Hat_Storage
The document discusses the benefits of software-defined storage over traditional storage approaches. It argues that software-defined storage uses standard hardware and open source software, providing flexibility, scalability, and lower costs compared to proprietary appliances or public cloud storage. It also describes Red Hat's portfolio of software-defined storage solutions, including Ceph and Gluster, which leverage open source technologies to power a variety of enterprise workloads.
Red Hat Storage Day Seattle: Supermicro Solutions for Red Hat Ceph and Red Ha...Red_Hat_Storage
This document discusses Supermicro's evolution from server and storage innovation to total solutions innovation. It provides examples of their all-flash storage servers and Red Hat Ceph testing results. Finally, it outlines their approach to providing optimized, turnkey storage solutions based on workload requirements and best practices learned from customer deployments and testing.
Red Hat Storage Day Seattle: Persistent Storage for Containerized ApplicationsRed_Hat_Storage
This document discusses persistent storage solutions for containerized applications. It describes how containers provide benefits like faster development and deployment cycles compared to virtual machines. However, most applications still require persistent storage for data. The document outlines requirements for container storage solutions, such as scalability, resilience, flexibility and being software-defined. It presents Red Hat Storage as a solution, highlighting features like replication, erasure coding and snapshots. Red Hat Storage can provide persistent storage to containers using technologies like Ceph, Amazon EBS, NFS and GlusterFS.
Red Hat Storage Day Seattle: Stretching A Gluster Cluster for Resilient Messa...Red_Hat_Storage
One of four Canadian universities ranked in the top 100 worldwide, McMaster University began a project in 2012 to replace its legacy systems with Oracle PeopleSoft for financials, human resources, and other functions. The new Mosaic system runs on Red Hat Enterprise Linux using Red Hat Gluster Storage for shared file storage and messaging where needed. The physical infrastructure is designed for high availability even if one data center room fails, using load balancing, database replication, and clustered resources.
Red Hat Storage Day Seattle: Stabilizing Petabyte Ceph Cluster in OpenStack C...Red_Hat_Storage
Cisco uses Ceph for storage in its OpenStack cloud platform. The initial Ceph cluster design used HDDs which caused stability issues as the cluster grew to petabytes in size. Improvements included throttling client IO, upgrading Ceph versions, moving MON metadata to SSDs, and retrofitting journals to NVMe SSDs. These steps stabilized performance and reduced recovery times. Lessons included having clear stability goals and automating testing to prevent technical debt from shortcuts.
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
A Comprehensive Guide to DeFi Development Services in 2024Intelisync
DeFi represents a paradigm shift in the financial industry. Instead of relying on traditional, centralized institutions like banks, DeFi leverages blockchain technology to create a decentralized network of financial services. This means that financial transactions can occur directly between parties, without intermediaries, using smart contracts on platforms like Ethereum.
In 2024, we are witnessing an explosion of new DeFi projects and protocols, each pushing the boundaries of what’s possible in finance.
In summary, DeFi in 2024 is not just a trend; it’s a revolution that democratizes finance, enhances security and transparency, and fosters continuous innovation. As we proceed through this presentation, we'll explore the various components and services of DeFi in detail, shedding light on how they are transforming the financial landscape.
At Intelisync, we specialize in providing comprehensive DeFi development services tailored to meet the unique needs of our clients. From smart contract development to dApp creation and security audits, we ensure that your DeFi project is built with innovation, security, and scalability in mind. Trust Intelisync to guide you through the intricate landscape of decentralized finance and unlock the full potential of blockchain technology.
Ready to take your DeFi project to the next level? Partner with Intelisync for expert DeFi development services today!
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Dive into the realm of operating systems (OS) with Pravash Chandra Das, a seasoned Digital Forensic Analyst, as your guide. 🚀 This comprehensive presentation illuminates the core concepts, types, and evolution of OS, essential for understanding modern computing landscapes.
Beginning with the foundational definition, Das clarifies the pivotal role of OS as system software orchestrating hardware resources, software applications, and user interactions. Through succinct descriptions, he delineates the diverse types of OS, from single-user, single-task environments like early MS-DOS iterations, to multi-user, multi-tasking systems exemplified by modern Linux distributions.
Crucial components like the kernel and shell are dissected, highlighting their indispensable functions in resource management and user interface interaction. Das elucidates how the kernel acts as the central nervous system, orchestrating process scheduling, memory allocation, and device management. Meanwhile, the shell serves as the gateway for user commands, bridging the gap between human input and machine execution. 💻
The narrative then shifts to a captivating exploration of prominent desktop OSs, Windows, macOS, and Linux. Windows, with its globally ubiquitous presence and user-friendly interface, emerges as a cornerstone in personal computing history. macOS, lauded for its sleek design and seamless integration with Apple's ecosystem, stands as a beacon of stability and creativity. Linux, an open-source marvel, offers unparalleled flexibility and security, revolutionizing the computing landscape. 🖥️
Moving to the realm of mobile devices, Das unravels the dominance of Android and iOS. Android's open-source ethos fosters a vibrant ecosystem of customization and innovation, while iOS boasts a seamless user experience and robust security infrastructure. Meanwhile, discontinued platforms like Symbian and Palm OS evoke nostalgia for their pioneering roles in the smartphone revolution.
The journey concludes with a reflection on the ever-evolving landscape of OS, underscored by the emergence of real-time operating systems (RTOS) and the persistent quest for innovation and efficiency. As technology continues to shape our world, understanding the foundations and evolution of operating systems remains paramount. Join Pravash Chandra Das on this illuminating journey through the heart of computing. 🌟
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
3. is Changing
Source: IDC – The Digital Universe of Opportunities: Rich Data and the Increasing Value of the Internet of Things - April 2014
Fromnow until 2020, the size ofthe digital
universe will aboutdouble everytwoyears
Information Growth*
2X
What we do with data ischanging, traditional
storage infrastructuredoesnot solve
tomorrow’sproblems
Complexity
ShiftingofIT servicesto cloud computing
and next-generationplatforms
Cloud
Emergence offlash storage and
software-defined environments
New Technologies
The World
4. Explosion
EVERY MINUTE EVERY DAY*
300 HOURS
VIDEO UPLOADED
TO YOUTUBE
51,000
APPS DOWNLOADED
204
MILLION E-MAILS
Source: TechSpartan.co.uk - 2013 vs 2015 In an Internet minute
48 HOURS
VIDEO UPLOADED
TO YOUTUBE
47,000
APPS DOWNLOADED
200
MILLIONE-MAILS
2013 2015
Information
5. TheImpactof
theCloud
Empowerment of the end-user
through cloud services
Emergence of new technologies
and architectures
Shifting the role of information
technology professionals
11. WhereDataisCreated
0
10
20
30
40
50
2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020
(ZB)
Unstructured Data Structured Data
Sources: IDC, 2011 Worldwide Enterprise Storage Systems 2011-2015 Forecast update
IDC, The Digital Universe Study In 2020 Forecasts
4.4ZB
44ZB
90%
10%
0%
20%
40%
60%
80%
100%
2012 2013 2014 2015 2016 2017 2018 2019 2020
% of Total Digital Universe
Emerging Markets Mature Markets
Sources: IDC,’s Digital Universe Study, 2014
USA, Canada, Western Europe,
Australia, NZ, and Japan
China, India, Mexico, Brazil,
and Russia
By 2020, about 90% of all data will be
unstructured, driven by Consumer Images,
Voice, and the Web
Emerging Markets will Surpass Mature
Markets before 2017 regarding data creation.
Data Creation by Type (ZB)
11
16. Latency: ~100x
Size of Data: ~1,000x 1000x faster than NAND
1000x higher endurance of NAND
10x denser than DRAM
Technology claims are based on comparisons of latency, density, and write cycling metrics amongst memory technologies recorded on published specifications of in-market memory
products against internal Intel specifications.
3DXpOINT™
TECHNOLOGY
New Class of
Non-Volatile Memory
17. 17
Intel®SolidStateDrives
“The only SSDs that never ever
gave me any issues like timeouts,
task aborts… are Intel DC S3700s”
From a post on ceph-devel*
Source:
http://ceph.com/r esources/maili ng-‐‑list-‐‑irc
18. 18
Intel’sroleinstorage
Advance theIndustry
OpenSource&Standards
BuildanOpenEcosystem
Intel®StorageBuilders
Endusersolutions
Cloud,Enterprise
IntelTechnologyLeadership
Storage Optimized CPU’s
Intel® Xeon® E5v4 2600 Platform
Intel® Xeon® Processor D-1500Platform
Storage Optimized Software
Intel® Intelligent AccelerationLibrary
Intel® Storage Performance Development Kit
Non-Volatile Memory
3D Xpoint™
Intel® Solid StateDrives for Datacenter
>7 Cloud storage solutions
architectures
70+ partners
Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such asSYSmark and MobileMark, are measured using specific
computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist
you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products.
helpingcustomerstoenableNextgen storage
Next gen solutions
architectures>26
>10 Enterprise storage solution
architectures
19.
20. 20
Intel Ceph Contribution Timeline
2014 2015 2016
* Right Edge of box indicates approximate release date
New Key/Value Store
Backend (rocksdb)
Giant* Hammer Infernalis Jewel
CRUSH Placement
Algorithm improvements
(straw2 bucket type)
BluestoreBackend
Optimizations for NVM
Bluestore SPDK
Optimizations
RADOS I/O Hinting
(35% better EC Write erformance)
Cache-tieringwith SSDs
(Write support)
PMStore
(NVM-optimized backend
based on libpmem)
RGW, Bluestore
Compression,Encryption
(w/ ISA-L, QAT backend)
VirtualStorageManager
(VSM)Open Sourced
CeTune
Open Sourced
ErasureCoding
support with ISA-L
Cache-tieringwith SSDs
(Read support)
Client-sideBlock Cache
(librbd)
21. Ceph@Intel–2016CephFocusAreas
21
Optimize for Intel® platforms, flash and networking
• Compression, Encryption hardware offloads (QAT & SOCs)
• PMStore (for 3D XPoint DIMMs)
• RBD caching and Cache tiering with NVM
• IA optimized storage libraries to reduce latency (ISA-L, SPDK)
Performance profiling, analysis and community contributions
• All flash workload profiling and latency analysis
• Streaming,Database and Analytics workload drivenoptimizations
Ceph enterprise usages and hardening
• Manageability (Virtual Storage Manager)
• Multi Data Center clustering (e.g., async mirroring)
End Customer POCs with focus on broad industry influence
• CDN,Cloud DVR,Video Surveillance, Ceph Cloud Services, AnalyticsPOCs
Ready to use IA, Intel NVM optimized systems & solutions from OEMs & ISVs
• Ready to use IA, Intel NVM optimized systems & solutions from OEMs & ISVs
• Intel system configurations, white papers,case studies
• Industry events coverage
Go to
Market
Intel® Storage
AccelerationLibrary
(Intel® ISA-L)
Intel® StoragePerformance
DevelopmentKit (SPDK)
Intel® Cache
Acceleration
Software (Intel® CAS)
Virtual StorageManager
Ce-Tune Ceph
Profiler
23. 23
4KRandomRead&WritePerformanceSummaryFirst Ceph cluster to break 1 Million 4K random IOPS
Software
and
workloads
used
in
performance
tests
may
have
been
optimized
for
performance
only
on
Intel
microprocessors.
Any
difference
in
system
hardware
or
software
design
or
configuration
may
affect
actual
performance.
See
configuration
slides
in
backup
for
details
on
software
configuration
and
test
benchmark
parameters.
Workload Pattern Max IOPS
4K 100% Random Reads (2TB Dataset)
1.35Million
4K 100% Random Reads (4.8TB Dataset)
1.15Million
4K 100% Random Writes (4.8TB Dataset)
200K
4K 70%/30% Read/Write OLTP Mix
(4.8TB Dataset) 452K
Source: Openstack Summit 2015: Accelerating Cassandraworkloadson ceph with all flashpcie ssds
24. Red Hat Ceph Reference Architecture Documents
24
25. MetaFormulaforCephDeployments
• Have a general understanding of the use cases you want to support with Ceph
• Understand the kind of performance or cost/performance you want to deliver
• Refer to a reference architecture resource to match your use case(s) with
known and measured reference architectures:
• http://www.redhat.com/en/resources/performance-and-sizing-guide-red-hat-
ceph-storage-qct-servers
• https://www.redhat.com/en/files/resources/en-rhst-cephstorage-supermicro-
INC0270868_v2_0715.pdf
• These documents have Ceph config, tuning and best practices guidance
• Additional help is available from Red Hat, including support and quick start
25