This document provides an overview and summary of the Gluster distributed file system for system administrators. It begins with introductions and defines key Gluster concepts and components. It then discusses how Gluster provides scaling and redundancy through various volume types like distributed, replicated, and erasure coded volumes. The document outlines how data can be accessed using native Gluster clients or NFS/SMB. It also covers general Gluster administration tasks and capabilities like profiling and geo-replication.
In this session, we'll discuss new volume types in Red Hat Gluster Storage. We will talk about erasure codes and storage tiers, and how they can work together. Future directions will also be touched on, including rule based classifiers and data transformations.
You will learn about:
How erasure codes lower the cost of storage.
How to configure and manage an erasure coded volume.
How to tune Gluster and Linux to optimize erasure code performance.
Using erasure codes for archival workloads.
How to utilize an SSD inexpensively as a storage tier.
Gluster's erasure code and storage tiering design.
This session will cover performance-related developments in Red Hat Gluster Storage 3 and share best practices for testing, sizing, configuration, and tuning.
Join us to learn about:
Current features in Red Hat Gluster Storage, including 3-way replication, JBOD support, and thin-provisioning.
Features that are in development, including network file system (NFS) support with Ganesha, erasure coding, and cache tiering.
New performance enhancements related to the area of remote directory memory access (RDMA), small-file performance, FUSE caching, and solid state disks (SSD) readiness.
Red Hat Storage Server Administration Deep DiveRed_Hat_Storage
"In this session for administrators of all skill levels, you’ll get a deep technical dive into Red Hat Storage Server and GlusterFS administration.
We’ll start with the basics of what scale-out storage is, and learn about the unique implementation of Red Hat Storage Server and its advantages over legacy and competing technologies. From the basic knowledge and design principles, we’ll move to a live start-to-finish demonstration. Your experience will include:
Building a cluster.
Allocating resources.
Creating and modifying volumes of different types.
Accessing data via multiple client protocols.
A resiliency demonstration.
Expanding and contracting volumes.
Implementing directory quotas.
Recovering from and preventing split-brain.
Asynchronous parallel geo-replication.
Behind-the-curtain views of configuration files and logs.
Extended attributes used by GlusterFS.
Performance tuning basics.
New and upcoming feature demonstrations.
Those new to the scale-out product will leave this session with the knowledge and confidence to set up their first Red Hat Storage Server environment. Experienced administrators will sharpen their skills and gain insights into the newest features. IT executives and managers will gain a valuable overview to help fuel the drive for next-generation infrastructures."
In this session, we'll discuss new volume types in Red Hat Gluster Storage. We will talk about erasure codes and storage tiers, and how they can work together. Future directions will also be touched on, including rule based classifiers and data transformations.
You will learn about:
How erasure codes lower the cost of storage.
How to configure and manage an erasure coded volume.
How to tune Gluster and Linux to optimize erasure code performance.
Using erasure codes for archival workloads.
How to utilize an SSD inexpensively as a storage tier.
Gluster's erasure code and storage tiering design.
This session will cover performance-related developments in Red Hat Gluster Storage 3 and share best practices for testing, sizing, configuration, and tuning.
Join us to learn about:
Current features in Red Hat Gluster Storage, including 3-way replication, JBOD support, and thin-provisioning.
Features that are in development, including network file system (NFS) support with Ganesha, erasure coding, and cache tiering.
New performance enhancements related to the area of remote directory memory access (RDMA), small-file performance, FUSE caching, and solid state disks (SSD) readiness.
Red Hat Storage Server Administration Deep DiveRed_Hat_Storage
"In this session for administrators of all skill levels, you’ll get a deep technical dive into Red Hat Storage Server and GlusterFS administration.
We’ll start with the basics of what scale-out storage is, and learn about the unique implementation of Red Hat Storage Server and its advantages over legacy and competing technologies. From the basic knowledge and design principles, we’ll move to a live start-to-finish demonstration. Your experience will include:
Building a cluster.
Allocating resources.
Creating and modifying volumes of different types.
Accessing data via multiple client protocols.
A resiliency demonstration.
Expanding and contracting volumes.
Implementing directory quotas.
Recovering from and preventing split-brain.
Asynchronous parallel geo-replication.
Behind-the-curtain views of configuration files and logs.
Extended attributes used by GlusterFS.
Performance tuning basics.
New and upcoming feature demonstrations.
Those new to the scale-out product will leave this session with the knowledge and confidence to set up their first Red Hat Storage Server environment. Experienced administrators will sharpen their skills and gain insights into the newest features. IT executives and managers will gain a valuable overview to help fuel the drive for next-generation infrastructures."
Red Hat Gluster Storage, Container Storage and CephFS PlansRed_Hat_Storage
At Red Hat Storage Day New York on 1/19/16, Red Hat's Sayan Saha took attendees through an overview of Red Hat Gluster Storage that included future plans for the product, Red Hat's plans for container storage, and the company's plans for CephFS.
Gluster for Geeks: Performance Tuning Tips & TricksGlusterFS
In this Gluster for Geeks technical webinar, Jacob Shucart, Senior Systems Engineer, will provide useful tips and tricks to make a Gluster cluster meet your performance requirements. He will review considerations for all different phases including planning, configuration, implementation, tuning, and benchmarking.
Topics covered will include:
• Protocols (CIFS, NFS, GlusterFS)
• Hardware configuration
• Tuning parameters
• Performance benchmarks
Red Hat Gluster Storage - Direction, Roadmap and Use-CasesRed_Hat_Storage
Red Hat Gluster Storage is open, software-defined storage that helps you manage big, unstructured, and semistructured data. This product is based on the open source project GlusterFS, a distributed scale-out file system technology, and focuses on file sharing, analytics, and hyper-converged use cases.
In this session, you will:
See real-life case studies about Red Hat Gluster Storage’s usage in production environments, including ideal workloads.
Learn about the Red Hat Gluster Storage roadmap, including innovations from the GlusterFS community pipeline.
Gain insights into how the product will be integrated with Red Hat Enterprise Virtualization (including hyperconvergence), Red Hat Satellite, and Red Hat Enterprise Linux OpenStack Platform.
Ceph is an open source distributed storage system designed for scalability and reliability. Ceph's block device, RADOS block device (RBD), is widely used to store virtual machines, and is the most popular block storage used with OpenStack.
In this session, you'll learn how RBD works, including how it:
* Uses RADOS classes to make access easier from user space and within the Linux kernel.
* Implements thin provisioning.
* Builds on RADOS self-managed snapshots for cloning and differential backups.
* Increases performance with caching of various kinds.
* Uses watch/notify RADOS primitives to handle online management operations.
* Integrates with QEMU, libvirt, and OpenStack.
Red Hat Enterprise Linux: Open, hyperconverged infrastructureRed_Hat_Storage
The next generation of IT will be built around flexible infrastructures and operational efficiencies, lowering costs and increasing overall business value in the organization.
A hyperconverged infrastructure that's built on Red Hat supported technologies--including Linux, Gluster storage, and oVirt virtualization manager--will run on commodity x86 servers using the performance of local storage, to deliver a cost-effective, modular, highly scalable, and secure hyperconverged solution.
Red Hat Gluster Storage, Container Storage and CephFS PlansRed_Hat_Storage
At Red Hat Storage Day New York on 1/19/16, Red Hat's Sayan Saha took attendees through an overview of Red Hat Gluster Storage that included future plans for the product, Red Hat's plans for container storage, and the company's plans for CephFS.
Gluster for Geeks: Performance Tuning Tips & TricksGlusterFS
In this Gluster for Geeks technical webinar, Jacob Shucart, Senior Systems Engineer, will provide useful tips and tricks to make a Gluster cluster meet your performance requirements. He will review considerations for all different phases including planning, configuration, implementation, tuning, and benchmarking.
Topics covered will include:
• Protocols (CIFS, NFS, GlusterFS)
• Hardware configuration
• Tuning parameters
• Performance benchmarks
Red Hat Gluster Storage - Direction, Roadmap and Use-CasesRed_Hat_Storage
Red Hat Gluster Storage is open, software-defined storage that helps you manage big, unstructured, and semistructured data. This product is based on the open source project GlusterFS, a distributed scale-out file system technology, and focuses on file sharing, analytics, and hyper-converged use cases.
In this session, you will:
See real-life case studies about Red Hat Gluster Storage’s usage in production environments, including ideal workloads.
Learn about the Red Hat Gluster Storage roadmap, including innovations from the GlusterFS community pipeline.
Gain insights into how the product will be integrated with Red Hat Enterprise Virtualization (including hyperconvergence), Red Hat Satellite, and Red Hat Enterprise Linux OpenStack Platform.
Ceph is an open source distributed storage system designed for scalability and reliability. Ceph's block device, RADOS block device (RBD), is widely used to store virtual machines, and is the most popular block storage used with OpenStack.
In this session, you'll learn how RBD works, including how it:
* Uses RADOS classes to make access easier from user space and within the Linux kernel.
* Implements thin provisioning.
* Builds on RADOS self-managed snapshots for cloning and differential backups.
* Increases performance with caching of various kinds.
* Uses watch/notify RADOS primitives to handle online management operations.
* Integrates with QEMU, libvirt, and OpenStack.
Red Hat Enterprise Linux: Open, hyperconverged infrastructureRed_Hat_Storage
The next generation of IT will be built around flexible infrastructures and operational efficiencies, lowering costs and increasing overall business value in the organization.
A hyperconverged infrastructure that's built on Red Hat supported technologies--including Linux, Gluster storage, and oVirt virtualization manager--will run on commodity x86 servers using the performance of local storage, to deliver a cost-effective, modular, highly scalable, and secure hyperconverged solution.
This is to introduce the related components in SUSE Linux Enterprise High Availability Extension product to build High Available Storage (ha-lvm/drbd/iscsi/nfs, clvm, ocfs2, cluster-raid1).
In this video from SC15, Larry Jones from Seagate provides an overview of the company's revamped HPC storage product line. At SC15, Seagate announced a major expansion of its HPC product portfolio including the ClusterStor HPC hard disk drive designed for Big Data applications.
Learn more: http://www.seagate.com/products/enterprise-servers-storage/enterprise-storage-systems/clustered-file-systems/
Watch the video presentation: http://wp.me/p3RLHQ-eMC
Sign up for our insideHPC Newsletter
Part 3 - Local Name Resolution in Linux, FreeBSD and macOS/iOSMen and Mice
The focus of this webinar will be to take a deeper look into this local name-resolution system and the implementations for other Unix systems like Linux and FreeBSD. Linux’s new über-Daemon “systemd” supports both mDNS and the Windows LLMNR (Link-Local-Multicast-Name-Resolution). We will also show how well a Systemd-Linux behaves in heterogenous networks running both Windows and macOS.
What CloudStackers Need To Know About LINSTOR/DRBDShapeBlue
Philipp explains the best performing Open Source software-defined storage software available to Apache CloudStack today. It consists of two well-concerted components. LINSTOR and DRBD. Each of them also has its independent use cases, where it is deployed alone. In this presentation, the combination of these two is examined. They form the control plane and the data plane of the SDS. We will touch on: Performance, scalability, hyper-convergence (data-locality for high IO performance), resiliency through data replication (synchronous within a site, 2-way, 3-way, or more), snapshots, backup (to S3), encryption at rest, deduplication, compression, placement policies (regarding failure domains), management CLI and webGUI, monitoring interface, self-healing (restoring redundancy after device/node failure), the federation of multiple sites (async mirroring and repeatedly snapshot difference shipping), QoS control (noisy neighbors limitation) and of course: complete integration with CloudStack for KVM guests. It is Open Source software following the Unix philosophy. Each component solves one task, made for maximal re-usability. The solution leverages the Linux kernel, LVM and/or ZFS, and many Open Source software libraries. Building on these giant Open Source foundations, not only saves LINBIT from re-inventing the wheels, it also empowers your day 2 operation teams since they are already familiar with these technologies.
Philipp Reisner is one of the founders and CEO of LINBIT in Vienna/Austria. He holds a Dipl.-Ing. (comparable to MSc) degree in computer science from Technical University in Vienna. His professional career has been dominated by developing DRBD, a storage replication software for Linux. While in the early years (2001) this was writing kernel code, today he leads a company of 30 employees with locations in Austria and the USA. LINBIT is an Open Source company offering enterprise-level support subscriptions for its Open Source technologies.
-----------------------------------------
CloudStack Collaboration Conference 2022 took place on 14th-16th November in Sofia, Bulgaria and virtually. The day saw a hybrid get-together of the global CloudStack community hosting 370 attendees. The event hosted 43 sessions from leading CloudStack experts, users and skilful engineers from the open-source world, which included: technical talks, user stories, new features and integrations presentations and more.
Big Data in Container; Hadoop Spark in Docker and MesosHeiko Loewe
3 examples for Big Data analytics containerized:
1. The installation with Docker and Weave for small and medium,
2. Hadoop on Mesos w/ Appache Myriad
3. Spark on Mesos
In this day and age, data grows so fast it’s not uncommon for those of us using a relational database to reach the limits of its capacity. In this session, Kwangbock Lee explains how Samsung uses ClustrixDB to handle fast-growing data without manual database sharding. He highlights lessons learned, including a few hiccups along the way, and shares Samsung's experience migrating to ClustrixDB.
Snapshots In Red Hat Storage Server Overview & QuickstartRed_Hat_Storage
"The snapshots feature in Red Hat Storage Server provides an integrated solution for point-in-time recovery of GlusterFS volumes. Join us to check it out. And while you’ll be exposed to architectural details of how snapshots are implemented,this session won’t be a deep dive into snapshot internals.
In this session, you’ll learn:
How to configure and use snapshots in Red Hat Storage Server.
How to do point-in-time recovery using snapshots.
The architectural overview of snapshots in Red Hat Storage Server.
The key features and offerings in Red Hat Storage Server snapshots.
Best practice recommendations for configuring Red Hat Storage Server volumes for optimal use with snapshots.
The limitations in the snapshots feature release.
Some important issues to consider when migrating from older Red Hat Storage Server releases to a snapshot-supported one.
The product roadmap and extensions to snapshots.
"
Hadoop 3.0 has been years in the making, and now it's finally arriving. Andrew Wang and Daniel Templeton offer an overview of new features, including HDFS erasure coding, YARN Timeline Service v2, YARN federation, and much more, and discuss current release management status and community testing efforts dedicated to making Hadoop 3.0 the best Hadoop major release yet.
Gluster fs tutorial part 2 gluster and big data- gluster for devs and sys admins presentation
1. DUSTIN L. BLACK, RHCA1
Dustin L. Black, RHCA
Sr. Technical Account Manager, Red Hat
2012-07-17
Demystifying Gluster
GlusterFS for the SysAdmin
2. DUSTIN L. BLACK, RHCA2
#whoami
● Systems and Infrastructure Geek
● Decade+ of Linux, UNIX, networking
● <notacoder/>
● Believe in Open Source Everything
● Sr. Technical Account Manager, Red
Hat GSS
● dustin@redhat.com
3. DUSTIN L. BLACK, RHCA3
#whatis TAM
● Premium named-resource support
● Proactive and early access
● Regular calls and on-site engagements
● Customer advocate within Red Hat and upstream
● Multi-vendor support coordinator
● High-touch access to engineering
● Influence for software enhancements
● NOT Hands-on or consulting
4. DUSTIN L. BLACK, RHCA4
Agenda
● Technology Overview
● A Peek at Gluster Logic
● Scaling Up and Out
● Redundancy and Fault Tolerance
● Data Access
● General Administration
● Use Cases
5. DUSTIN L. BLACK, RHCA5
Technology Overview
Demystifying Gluster
GlusterFS for the SysAdmin
6. DUSTIN L. BLACK, RHCA6
What is GlusterFS?
● POSIX-Compliant Distributed File System
● No Metadata Server
● Network Attached Storage (NAS)
● Heterogeneous Commodity Hardware
● Aggregated Storage and Memory
● Standards-Based – Clients, Applications, Networks
● Flexible and Agile Scaling
● Capacity – Petabytes and beyond
● Performance – Thousands of Clients
● Single Global Namespace
7. DUSTIN L. BLACK, RHCA7
What is Red Hat Storage?
● Enterprise Implementation of GlusterFS
● Software Appliance
● Bare Metal Installation
● Built on RHEL + XFS
● Subscription Model
● Storage Software Appliance
● Datacenter and Private Cloud Deployments
● Virtual Storage Appliance
● Amazon Web Services Public Cloud Deployments
8. DUSTIN L. BLACK, RHCA8
GlusterFS vs. Traditional Solutions
● A basic NAS has limited scalability and redundancy
● Other distributed filesystems limited by metadata
● SAN is costly & complicated but high performance &
scalable
● GlusterFS
● Linear Scaling
● Minimal Overhead
● High Redundancy
● Simple and Inexpensive Deployment
9. DUSTIN L. BLACK, RHCA9
Technology Stack
Demystifying Gluster
GlusterFS for the SysAdmin
10. DUSTIN L. BLACK, RHCA10
Terminology
● Brick
● A filesystem mountpoint
● A unit of storage used as a GlusterFS building block
● Translator
● Logic between the bits and the Global Namespace
● Layered to provide GlusterFS functionality
● Volume
● Bricks combined and passed through translators
● Node
● Server running the gluster daemon and sharing volumes
11. DUSTIN L. BLACK, RHCA11
Foundation Components
● Private Cloud (Datacenter)
● Common Commodity x86_64 Servers
● RHS: Hardware Compatibility List (HCL)
● Public Cloud
● Amazon Web Services (AWS)
● EC2 + EBS
12. DUSTIN L. BLACK, RHCA12
Disk, LVM, and Filesystems
● Direct-Attached Storage (DAS)
● Just a Bunch Of Disks (JBOD)
● Hardware RAID
● RHS: RAID 6 required
● Logical Volume Management (LVM)
● XFS, EXT3/4, BTRFS
● Extended attributes support required
13. DUSTIN L. BLACK, RHCA13
Data Access Overview
● Gluster Native Client
● Filesystem in Userspace (FUSE)
● NFS
● Built-in Service
● SMB/CIFS
● Samba server required
14. DUSTIN L. BLACK, RHCA14
Gluster Components
● glusterd
● Elastic volume management daemon
● Runs on all export servers
● Interfaced through gluster CLI
● glusterfsd
● GlusterFS brick daemon
● One process for each brick
● Managed by glusterd
17. DUSTIN L. BLACK, RHCA17
Scaling
Demystifying Gluster
GlusterFS for the SysAdmin
18. DUSTIN L. BLACK, RHCA18
Scaling Up
● Add disks and filesystems to a node
● Expand a gluster volume by adding bricks
XFS
19. DUSTIN L. BLACK, RHCA19
Scaling Out
● Add gluster nodes to trusted pool
● Add filesystems as new bricks
20. DUSTIN L. BLACK, RHCA20
Under the Hood
Demystifying Gluster
GlusterFS for the SysAdmin
21. DUSTIN L. BLACK, RHCA21
Elastic Hash Algorithm
● No central metadata
● No Performance Bottleneck
● Eliminates risk scenarios
● Location hashed intelligently on path and filename
● Unique identifiers, similar to md5sum
● The “Elastic” Part
● Files assigned to virtual volumes
● Virtual volumes assigned to multiple bricks
● Volumes easily reassigned on the fly
23. DUSTIN L. BLACK, RHCA23
Distribution and Replication
Demystifying Gluster
GlusterFS for the SysAdmin
24. DUSTIN L. BLACK, RHCA24
Distributed Volume
● Files “evenly” spread across bricks
● File-level RAID 0
● Server/Disk failure could be catastrophic
25. DUSTIN L. BLACK, RHCA25
Replicated Volume
● Copies files to multiple bricks
● File-level RAID 1
26. DUSTIN L. BLACK, RHCA26
Distributed Replicated Volume
● Distributes files across replicated bricks
● RAID 1 plus improved read performance
27. DUSTIN L. BLACK, RHCA27
Geo Replication
● Asynchronous across LAN, WAN, or Internet
● Master-Slave model -- Cascading possible
● Continuous and incremental
● Time should be synchronized on all master nodes
28. DUSTIN L. BLACK, RHCA28
Replicated Volumes vs Geo-replication
Replicated Volumes Geo-replication
Mirrors data across clusters Mirrors data across geographically
distributed clusters
Provides high-availability Ensures backing up of data for disaster
recovery
Synchronous replication (each and every
file operation is sent across all the bricks)
Asynchronous replication (checks for the
changes in files periodically and syncs
them on detecting differences)
29. DUSTIN L. BLACK, RHCA29
Layered Functionality
Demystifying Gluster
GlusterFS for the SysAdmin
30. DUSTIN L. BLACK, RHCA30
Striped Volumes
● Individual files split among bricks
● Similar to RAID 0
● Limited Use Cases – HPC Pre/Post Processing
31. DUSTIN L. BLACK, RHCA31
Distributed Striped Volume
● Files striped across two or more nodes
● Striping plus scalability
32. DUSTIN L. BLACK, RHCA32
Striped Replicated Volume
● RHS 2.0 / GlusterFS 3.3+
● Similar to RAID 10 (1+0)
Replicated Volume
Striped Striped
33. DUSTIN L. BLACK, RHCA33
Distributed Striped Replicated Volume
● GlusterFS 3.3+
● Limited Use Cases – Map Reduce
34. DUSTIN L. BLACK, RHCA34
Data Access
Demystifying Gluster
GlusterFS for the SysAdmin
35. DUSTIN L. BLACK, RHCA35
Gluster Native Client (FUSE)
● FUSE kernel module allows the filesystem to be built
and operated entirely in userspace
● Specify mount to any GlusterFS node
● Native Client fetches volfile from mount server, then
communicates directly with all nodes to access data
● Recommended for high concurrency and high write
performance
36. DUSTIN L. BLACK, RHCA36
NFS
● Standard NFS v3 clients
● Mount with vers=3 option
● Standard automounter is supported
● Mount to any node, or use a load balancer
● Gluster NFS server includes Network Lock Manager
(NLM) to synchronize locks across clients
● Better performance for reading many small files
37. DUSTIN L. BLACK, RHCA37
SMB/CIFS
● GlusterFS volume is first redundantly mounted with the
Native Client on localhost
● Native mount point is then shared via Samba
● Must be setup on each node you wish to connect to via
CIFS
38. DUSTIN L. BLACK, RHCA38
General Administration
Demystifying Gluster
GlusterFS for the SysAdmin
40. DUSTIN L. BLACK, RHCA40
Adding Nodes (peers) and Volumes
gluster> peer probe server3
gluster> peer status
Number of Peers: 2
Hostname: server2
Uuid: 5e987bda-16dd-43c2-835b-08b7d55e94e5
State: Peer in Cluster (Connected)
Hostname: server3
Uuid: 1e0ca3aa-9ef7-4f66-8f15-cbc348f29ff7
State: Peer in Cluster (Connected)
gluster> volume create my-dist-vol server2:/brick2 server3:/brick3
gluster> volume info my-dist-vol
Volume Name: my-dist-vol
Type: Distribute
Status: Created
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: server2:/brick2
Brick2: server3:/brick3
gluster> volume start my-dist-vol
Distributed Volume
41. DUSTIN L. BLACK, RHCA41
Distributed Striped Replicated Volume
gluster> volume create test-volume replica 2 stripe 2 transport tcp
server1:/exp1 server1:/exp2 server2:/exp3 server2:/exp4
server3:/exp5 server3:/exp6 server4:/exp7 server4:/exp8
Multiple bricks of a replicate volume are present on the same server. This setup is not
optimal.
Do you still want to continue creating the volume? (y/n) y
Creation of volume test-volume has been successful. Please start the volume to access
data.
<- test-volume<- test-volume
<- distributed files -><- distributed files ->
stripe 2stripe 2
replica 2replica 2
42. DUSTIN L. BLACK, RHCA42
Distributed Striped Replicated Volume
gluster> volume info test-volume
Volume Name: test-volume
Type: Distributed-Striped-Replicate
Volume ID: 8f8b8b59-d1a1-42fe-ae05-abe2537d0e2d
Status: Created
Number of Bricks: 2 x 2 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: server1:/exp1
Brick2: server2:/exp3
Brick3: server1:/exp2
Brick4: server2:/exp4
Brick5: server3:/exp5
Brick6: server4:/exp7
Brick7: server3:/exp6
Brick8: server4:/exp8
gluster> volume create test-volume stripe 2 replica 2 transport tcp
server1:/exp1 server2:/exp3 server1:/exp2 server2:/exp4
server3:/exp5 server4:/exp7 server3:/exp6 server4:/exp8
Creation of volume test-volume has been successful. Please start the volume to access
data.
44. DUSTIN L. BLACK, RHCA44
Migrating Data / Replacing Bricks
gluster> volume replace-brick my-dist-vol server3:/brick3 server5:/brick5 start
gluster> volume replace-brick my-dist-vol server3:/brick3 server5:/brick5 status
Current File = /usr/src/linux-headers-2.6.31-14/block/Makefile
Number of files migrated = 10567
Migration complete
gluster> volume replace-brick my-dist-vol server3:/brick3 server5:/brick5 commit
45. DUSTIN L. BLACK, RHCA45
Volume Options
gluster> volume set my-dist-vol auth.allow 192.168.1.*
gluster> volume set my-dist-vol auth.reject 10.*
gluster> volume set my-dist-vol nfs.volume-access read-only
gluster> volume set my-dist-vol nfs.disable on
gluster> volume set my-dist-vol features.read-only on
gluster> volume set my-dist-vol performance.cache-size 67108864
gluster> volume set my-dist-vol auth.allow 192.168.1.*
gluster> volume set my-dist-vol auth.reject 10.*
NFS
Auth
Other advanced options
46. DUSTIN L. BLACK, RHCA46
Volume Top Command
gluster> volume set my-dist-vol auth.allow 192.168.1.*
gluster> volume set my-dist-vol auth.reject 10.*
gluster> volume top my-dist-vol read brick server3:/brick3 list-cnt 3
Brick: server:/export/dir1
==========Read file stats========
read filename
call count
116 /clients/client0/~dmtmp/SEED/LARGE.FIL
64 /clients/client0/~dmtmp/SEED/MEDIUM.FIL
54 /clients/client2/~dmtmp/SEED/LARGE.FIL
● Many top commands are available for analysis of
files, directories, and bricks
● Read and write performance test commands available
● Perform active dd tests and measure throughput
48. DUSTIN L. BLACK, RHCA48
Geo-Replication
gluster> volume set my-dist-vol auth.allow 192.168.1.*
gluster> volume set my-dist-vol auth.reject 10.*
gluster> volume geo-replication my-dist-vol slavehost1:my-dist-repl start
Starting geo-replication session between my-dist-vol & slavehost1:my-dist-repl has been
successful
gluster> volume geo-replication my-dist-vol status
MASTER SLAVE STATUS
--------------------------------------------------------------------------------
my-dist-vol gluster://slavehost1:my-dist-repl OK
# ssh-keygen -f /var/lib/glusterd/geo-replication/secret.pem
# ssh-copy-id -i /var/lib/glusterd/geo-replication/secret.pem repluser@slavehost1
gluster> volume geo-replication my-dist-vol repluser@slavehost1:/repl_dir start
Starting geo-replication session between my-dist-vol & slavehost1:/repl_dir has been
successful
gluster> volume geo-replication my-dist-vol status
MASTER SLAVE STATUS
--------------------------------------------------------------------------------
my-dist-vol ssh://repluser@slavehost1:/repl_dir OK
Remote GlusterFS Volume
Remote SSH
gluster> volume info my-dist-vol
...
Options Reconfigured:
geo-replication.indexing: on
49. DUSTIN L. BLACK, RHCA49
Use Cases
Demystifying Gluster
GlusterFS for the SysAdmin
50. DUSTIN L. BLACK, RHCA50
Common Solutions
● Media / Content Distribution Network (CDN)
● Backup / Archive / Disaster Recovery (DR)
● Large Scale File Server
● Home directories
● High Performance Computing (HPC)
● Infrastructure as a Service (IaaS) storage layer
51. DUSTIN L. BLACK, RHCA51
Hadoop – Map Reduce
● Access data within and outside of Hadoop
● No HDFS name node single point of failure / bottleneck
● Seamless replacement for HDFS
● Scales with the massive growth of big data
52. DUSTIN L. BLACK, RHCA52
CIC Electronic Signature Solutions
● Challenge
● Must leverage economics of the cloud
● Storage performance in the cloud too slow
● Need to meet demanding client SLA’s
● Solution
● Red Hat Storage Software Appliance
● Amazon EC2 and Elastic Block Storage (EBS)
● Benefits
● Faster development and delivery of new products
● SLA’s met with headroom to spare
● Accelerated cloud migration
● Scale-out for rapid and simple expansion
● Data is highly available for 24/7 client access
Hybrid Cloud: Electronic Signature Solutions
● Reduced time-to-
market for new
products
● Meeting all client SLAs
● Accelerating move to
the cloud
53. DUSTIN L. BLACK, RHCA53
Pandora Internet Radio
● Challenge
● Explosive user & title growth
● As many as 12 file formats for each song
● ‘Hot’ content and long tail
● Solution
● Three data centers, each with a six-node GlusterFS
cluster
● Replication for high availability
● 250+ TB total capacity
● Benefits
● Easily scale capacity
● Centralized management; one administrator to manage
day-to-day operations
● No changes to application
● Higher reliability
Private Cloud: Media Serving
● 1.2 PB of audio served
per week
● 13 million files
● Over 50 GB/sec peak
traffic
54. DUSTIN L. BLACK, RHCA54
Brightcove
• Over 1 PB currently in
Gluster
• Separate 4 PB project
in the works
Private Cloud: Media Serving
● Challenge
● Explosive customer & title growth
● Massive video in multiple locations
● Costs rising, esp. with HD formats
● Solution
● Complete scale-out based on commodity DAS/JBOD
and GlusterFS
● Replication for high availability
● 1PB total capacity
● Benefits
● Easily scale capacity
● Centralized management; one administrator to manage
day-to-day operations
● Higher reliability
● Path to multi-site
55. DUSTIN L. BLACK, RHCA55
Pattern Energy
• Rapid and advance
weather predictions
• Maximizing energy
assets
• Cost savings and
avoidance
High Performance Computing for Weather Prediction
●
Challenge
●
Need to deliver rapid advance weather predictions
●
Identify wind and solar abundance in advance
●
More effectively perform preventative maintenance and
repair
●
Solution
●
32 HP compute nodes
●
Red Hat SSA for high throughput and availability
●
20TB+ total capacity
●
Benefits
●
Predicts solar and wind patterns 3 to 5 days in advance
●
Maximize energy production and repair times
●
Avoid costs of outsourcing weather predictions
●
Solution has paid for itself many times over
56. Thank You!
● dustin@redhat.com
● storage-sales@redhat.com
● RHS:
www.redhat.com/storage/
● GlusterFS:
www.gluster.org
● TAM:
access.redhat.com/support/offerings/tam/
@Glusterorg
@RedHatStorage
Gluster
Red Hat Storage
Demystifying Gluster
GlusterFS for the SysAdmin