On Beyond Disk: Breaking the Flash Cost Barrier


Published on

Scott Dietzen, CEO, Pure Storage, discusses why Flash memory is replacing rotating disk storage in data centers, and how Pure Storage enables flash memory to deliver 10X the performance of typical disk arrays at a lower cost.

Published in: Technology
  • Be the first to comment

No Downloads
Total Views
On Slideshare
From Embeds
Number of Embeds
Embeds 0
No embeds

No notes for slide

On Beyond Disk: Breaking the Flash Cost Barrier

  1. 1. On Beyond Disk:Breaking the Flash Cost BarrierScott DietzenCEO, Pure Storage
  2. 2. Rotating Disk Challenges on Random I/O Deconstructing a Random Read Typical Resulting Duty Pattern: Spin disk 1 ~2ms Data Data Data Data Read Data 3 ~0.04ms ß S E E K à ß S E E K à ß S E E K à 2 Actuate arm Time à ~4ms Disks spend >95% of their time seeking and rotating, not delivering data! Slide #3
  3. 3. Moore s Law vs. Newton s Law Moore s Law: HDD Areal Density (Kryder’s Law): 58% CAGR 40-100% CAGR HDD Latency (Seek Time) 10 -3% CAGR ms 5 Sources: Intel, IDEMA, IBM Research 0 2006 2008 2011 2013 Slide #4
  4. 4. The Storage Performance Crisis From the perspective of a CPU doing random I/O, today’s hard drive is slower than tape was two decades ago! CPU demand for random I/O Performance Performance Gap IOPs/ Spindle … while IOPS/GB drop Time IOPs/ GB Slide #5
  5. 5. Flash Memory Enables a Quantum Leap inPotential Performance Rotating Disk Flash Memory DRAM Random Read 180 5-10,000s 500,000+ IOPs Latency (us) 5,000 - 20,000 200-500 5 Active/Standby Power (watts) 15 / 10 6 / 0.05 10 Performance • Random reads terrible • Random reads great • Non-persistent through Dynamics • Sequential reads good • Sequential reads good power loss Slide #6
  6. 6. Flash Architecture Approaches & $s Server-Attach Array with Array with All-Flash PCI Flash Flash Cache Flash Tier Appl or Array  PCI card in   Flash as a read and/ • Sub-LUN/FS tiering, • 100% flash appliance application server or write cache in where a LUN or FS array, or to cache is spread across flash • Currently ~$20/GB,  Host CPU FS metadata and disk $40/GB HA typically used for flash management • All data persisted to • Hot blocks/files • True arrays (HA, snapshots, et al.) spinning disk moved to flash, cold  100s of GB left on disk coming • Typically <1% - 5%  ~$20/GB, of total capacity • Typically 1% - 10% of • At what cost? $40/GB HA total capacity How about • >$40/GB for flash <$7/GB? • >$40/GB for flash Slide #7
  7. 7. The Storage Economics Curve The Flash “Economic Wall” •  $40-100/GB HA usable •  Self-selects only into niche workloads All-Flash Performance Disk / Flash Storage Performance Hybrid Needs Rising •  Virtualization Performance The Disk “Performance Wall” Disk •  Disk getting larger not faster •  Cloud computing Capacity •  IOPS/TB going down •  CPU growth Disk Cost Slide #8
  8. 8. What if the Flash Cost Barrier Were Broken? All-Flash All-Flash Performance Disk / Flash Hybrid Performance Disk Capacity Capacity Disk Disk Cost Cost Slide #9
  9. 9. The Pure Storage Unique Recipe:Breaking the Cost Barrier to Flash •  100Ks of IOPS Purpose-Built 1 All Flash Architecture •  <1ms latency •  No tiering •  100% MLC flash •  Inline deduplication 2 and compression Inline Data •  5-20x reduction Reduction •  Active/active Enterprise 3 high availability •  10s to 100s of Resiliency TBs scale and Scale •  Plug-compatible Slide #10
  10. 10. Excellent Data Reduction on Real World Data VMware VMware •  15-to-1 data reduction •  17-to-1 data reduction •  50% fewer servers •  Virtualized SQL & Exchange MS SQL Oracle •  9.2-to-1 data reduction •  7-to-1 data reduction •  3.5x query time improvement •  Enabled virtualized Oracle Slide #11
  11. 11. Pure Storage / Samsung VMworld Demo  1000 VMs running in 16U of server and storage  Two HP DL580 with Intel E7-4800 processors ……   4 processors/10 cores; 1 TB of Samsung Green Memory  Pure Storage FlashArray   2 Controllers and 2 “Disk” Shelves (Samsung SSDs)  Software: vSphere 5.0 (beta), Virtual Center 5.0  Server workload simulation, Linux/Windows Mix 64 TB Lun  100K IOPS at very low latency (< ½ ms avg.)  Single 64TB LUN  Containing 38TBs of VMs  Reduced to <2TB (> than 15:1 data reduction) Slide #12
  12. 12. The First All-Flash Enterprise Array10x Faster 10x Lower 5x Less Latency Power, Space 10x Less Power & Space 5-20X Data Reduction For LESS than the price of performance disk Slide #13
  13. 13. Pure Storage Beats Disk Storage in Every Dimension 10,000s of IOPS 100,000s of IOPS 5-10s ms latency < 1 ms latency 1000s of watts 100s of watts $5-10 / GB < $5-10 / GB Slide #14
  14. 14. Pure SimplicitySay Goodbye to…  Setting and managing RAID  Host / array block alignment  Troubleshooting performance bottlenecks   head contention   port contention   spindle contention   cache partitioning  Day-long RAID re-builds  Sizing / managing multiple tiers of storage  Over-provisioning and reclamation Slide #15
  15. 15. Summary   Datacenters are facing a storage performance crisis  Flash is the answer for random I/O intensive workloads  Flash plus data reduction can deliver very compelling economics (better than the price of disk-centric storage arrays)  Understanding your applications “I/O fingerprint” (IOPS, block size, locality, read/write, latency) is key to choosing the best flash strategy for your environment  Flash should be evaluated on several dimensions, it s not only about performance Power & Operations Performance Cost Integration Protection Size Simplicity Slide #16