customer customer has found that the performance levels are acceptable, but increase capacity by If a has found that the performance levels are acceptable, but wants to wants to increase cld addcould add another 4, 1 TB each server, and will not generally generally experience performan another 4, 1 TB drives to drives to each server, and will not experience performance degrad
than 12). Note that in this case, they are adding 2 more low-price servers, and can simdrives. (See Config. C, above)If they want to both quadruple performance and quadruple capacity, they could distributeeach server would have 12,1 TB drives). (See Config. D, below)Note that by the time a solution has approximately 10 drives, the performance bottleneckmoved to the network. (See Config. D, above)
Ethernet network. Note that performance in this example is more than 25x that which we sais evidenced by an increase in performance from 200 MB/s in the baseline configuratioConfig. E, below)As you will note, the power of the scale-out model is that both capacity and performancmeet requirements. It is not necessary to know what performance levels will be needed 2,configurations can be easily adjusted as the need demands.
➜ BETA_LINK="http://download.gluster.com/pub/gluster/glusterfs/qa-releases/3.3-beta-2/glusterfs-3.3beta2.tar.gz"➜ wget $BETA_LINK➜ tar zxvf glusterfs-3.3beta2.tar.gz➜ cd glusterfs-3.3beta2➜ ./configure && make➜ sudo make install
# Start Gluster management daemon for each server➜ sudo /etc/init.d/glusterd start# Adding Servers to Trusted Storage Pool➜ for HOST in host1 host2 host3; do gluster peer probe $HOST; done#=> Probe successful Probe successful Probe successful
➜ sudo gluster peer status#=> Number of Peers: 3 Hostname: host1 Uuid: 81982001-ba0d-455a-bae8-cb93679dbddd State: Peer in Cluster (Connected) Hostname: host2 Uuid: 03945cd4-7487-4b2c-9384-f006a76dfee5 State: Peer in Cluster (Connected)...
# Create a distribute Volume named ‘log’➜ SERVER_LOG_PATH=”/mnt/glusterfs/server/log”➜ sudo gluster volume create log transport tcp host1: $SERVER_LOG_PATH host2: $SERVER_LOG_PATH host3: $SERVER_LOG_PATH
Multi-site cascading Geo-replication Geo-replication over LAN You can configure GlusterFS Geo-replication to mirror data over a Local Area Network. Geo-replication over WAN You can configure GlusterFS Geo-replication to replicate data over a Wide Area Network. Geo-replication over WAN You can configure GlusterFS Geo-replication to replicate data over a Wide Area Network. Geo-replication over Internet You ds can configure GlusterFS Geo-replication to mirror data over the Internet. Geo-replication over InternetGluster File system Administration Guide_3.2_02_B Pg No. 47 You can configure GlusterFS Geo-replication to mirror data over the Internet.
Figure 5, below, illustrates a typical distributed metadata server implementation. It can be seen that this approalso results in considerable overhead processing for file access, and by design has built-in exposurecorruption scenarios. Here again we see a legacy approach to scale-out storage not congruent withrequirement of the modern data center or with the burgeoning migration to virtualization and cloud computing. !Figure!5!Decentralized!Metadata!Approach
any office that stores physical documents in folders in filing cabinets, that person should be able to f -Similarly, one could implement an algorithmic approach to data storage that used a similarlocate files. For example, in a ten system cluster, one isk 10, etc. Figure 6, below illustrates this concept. !Figure!6:!Understanding!EHA:!Algorithm
and run it through the hashing algorithm. Each pathname/filename results in a unique numerical rFor the sake of simplicity, one could imagine assigning all files whose hash ends in the number 1all which end in the number 2 to the second disk, etc. Figure 7, below, illustrates this concept. !Figure!7!Understanding!EHA:!Hashing
1. Setting up a very large number of virtual volumes 2. Using the hashing algorithm to assign files to virtual volumes 3. Using a separate process to assign virtual volumes to multiple physical devicesThus, when disks or nodes are added or deleted, the algorithm itself does not need to be changed. However,virtual volumes can be migrated or assigned to new physical locations as the need arises. Figure 8, below,illustrates the Glus !Figure!8!Understanding!EHA:!Elasticity