Software Defined Storage
Reality or BS?
Defining Software Defined Storage
• We must resist software defined
washing
• My Criteria:
– Must run on x86 server hardwa...
Software Defined Storage is not New
• Novell NetWare turned PC AT into file server
– Software defined NAS
• I wrote “How t...
Software Defined Storage Taxonomy
• Storage SoftwareDefinus OldSchoolus
– Runs on standard server
– Publishes block or fil...
Old School SDS
• Software runs underWindows or Linux
• Publish storage as iSCSI or file
• Standard RAID
– Could require HW...
Selected Classic SDS
• FreeNAS/OpenNAS Etc.
– Open source ZFS based NAS/iSCSI
• Include SSD caching
• Open-E DSS
– Open so...
Wrap in Tin?
• Most next generation storage arrays are SDS
Nimble Storage Fusion-IO IOcontrol (Nexgen)
Tintri Tegile
• Why...
OrVirtualize The Servers
• Creating aVirtual Storage Appliance (VSA)
• VSAs great solution for ROBO, SMB
• Local storage, ...
Why Converge Storage and Compute?
• Makes corporate data center like cloud
– A good or bad thing
• Storage array slot and ...
Enter The ServerSAN
• Scale-out whereVSAs are
fail-over clusters
– Storage across n hypervisor
hosts to form one pool
– Ma...
ServerSAN architecture differentiators
• Data protection model
– Per node RAID?
– N-way replication
– Network RAID?
• Flas...
Hyperconvirged Systems
• Nutanix
– Derived from Google File System
– 4 nodes/block
– Multi-hypervisor
– Storage for cluste...
Vmware’sVSAN
• SSD as read/write cache
• N-way replication (no local RAID)
– Default 2 copies, requires 3 nodes
– 3 copies...
Software Only ServerSANs
• HP StoreVirtual (Lefthand)
– Sub-LUN tiering for SSDs
– iSCSI system scales to 10 nodes
– Data ...
More Software ServerSANs
• EMC ScaleIO
– Extreme scale-out to 100s of nodes
– Multi-hypervisor
• Kernel modules for KVM, X...
ServerSANs and Server Form Factors
• Mainstream
– 1u servers offer limited storage
– 2u servers the sweet spot
• 6-24 driv...
Challenges to SDS
• Purchasing politics and budgets
– Everyone likes to point at where their money
went, especially storag...
Operational Challenges
• Storage guys are paranoid for good reason
– Storage is persistent and so are storage mistakes
• S...
ServerSAN vs Dedicated Storage
• VMware’s benchmark config
– 32 nodes
– 400GB SSD and 7 1TB drives each
– VSAN cost ~$11,0...
Questions and Contact
• Contact info:
– Hmarks@deepstorage.net
– @DeepStoragenet onTwitter
Upcoming SlideShare
Loading in...5
×

Software defined storage real or bs-2014

1,763

Published on

Published in: Technology, Business
0 Comments
3 Likes
Statistics
Notes
  • Be the first to comment

No Downloads
Views
Total Views
1,763
On Slideshare
0
From Embeds
0
Number of Embeds
2
Actions
Shares
0
Downloads
116
Comments
0
Likes
3
Embeds 0
No embeds

No notes for slide

Software defined storage real or bs-2014

  1. 1. Software Defined Storage Reality or BS?
  2. 2. Defining Software Defined Storage • We must resist software defined washing • My Criteria: – Must run on x86 server hardware • No custom ASICs • NVRAM allowed – Can be sold as software or Appliance • “Wrapped in tin”
  3. 3. Software Defined Storage is not New • Novell NetWare turned PC AT into file server – Software defined NAS • I wrote “How to BuildYour Own iSCSI Array” in 2006 • Faster x86 processors are driving SDS – In performance – In scale – In functionality
  4. 4. Software Defined Storage Taxonomy • Storage SoftwareDefinus OldSchoolus – Runs on standard server – Publishes block or file • Storage SoftwareDefinusVirtualum – The classicVSA – Like OldSchoolus but runs in aVM • Storage SoftwareDefinusVirtucalis Scaleoutus – Also known as ServerSAN – Pools storage across hypervisor hosts
  5. 5. Old School SDS • Software runs underWindows or Linux • Publish storage as iSCSI or file • Standard RAID – Could require HW controller • Synchronous replication with failover
  6. 6. Selected Classic SDS • FreeNAS/OpenNAS Etc. – Open source ZFS based NAS/iSCSI • Include SSD caching • Open-E DSS – Open source assemblage w/support • NexentaStor – Commercial ZFS • Supports shared storage • StarWind – Windows based iSCSI target w/SSD caching
  7. 7. Wrap in Tin? • Most next generation storage arrays are SDS Nimble Storage Fusion-IO IOcontrol (Nexgen) Tintri Tegile • Why? – Qualification and support – Margins – Channel issues – Customer preferences
  8. 8. OrVirtualize The Servers • Creating aVirtual Storage Appliance (VSA) • VSAs great solution for ROBO, SMB • Local storage, may require RAID controller • Publish as iSCSI or NFS • Example: Stormagic – Basic iSCSIVSA 2 nodes $2500
  9. 9. Why Converge Storage and Compute? • Makes corporate data center like cloud – A good or bad thing • Storage array slot and SAN costs – Generally higher than the disk drive that plugs in • Server slots are already paid for • Political, management issues – Moves storage to server team
  10. 10. Enter The ServerSAN • Scale-out whereVSAs are fail-over clusters – Storage across n hypervisor hosts to form one pool – Maximum cluster sizes 8-32 nodes • Use SSD as cache or tier • Can be software or hyperconvirged servers
  11. 11. ServerSAN architecture differentiators • Data protection model – Per node RAID? – N-way replication – Network RAID? • Flash usage: – Write through or write back cache – SubLUN tiering • Prioritization/Storage QoS • Data locality • Data reduction • Snapshots and cloning
  12. 12. Hyperconvirged Systems • Nutanix – Derived from Google File System – 4 nodes/block – Multi-hypervisor – Storage for cluster only • Simplivity – Dedupe and backup to the cloud – Storage available to other servers – 2u Servers • Both have compute and storage heavy models • Pivot3 for VDI only • ScaleComputing KVM based for SMBs
  13. 13. Vmware’sVSAN • SSD as read/write cache • N-way replication (no local RAID) – Default 2 copies, requires 3 nodes – 3 copies requires 5 nodes (my recommendation) • Scales to 32 nodes • Runs directly in hypervisor kernel • Storage only available to cluster members • Relies on vSphere snapshots, replication • License $2495/CPU or $50/VDI image
  14. 14. Software Only ServerSANs • HP StoreVirtual (Lefthand) – Sub-LUN tiering for SSDs – iSCSI system scales to 10 nodes – Data Protection • Per Node RAID • 2-4 way replication or network RAID 5 or 6 • Maxta Storage Platform – Data deduplication and compression – Metadata based snapshots – Data integrity via hashes/scrubbing – Data locality
  15. 15. More Software ServerSANs • EMC ScaleIO – Extreme scale-out to 100s of nodes – Multi-hypervisor • Kernel modules for KVM, XEN, Hyper-V – Multiple storage pools – Some QoS – Metadata snaps and clones • Sanbolic Melio – Evolved from clustered file system – Perhaps the most mature
  16. 16. ServerSANs and Server Form Factors • Mainstream – 1u servers offer limited storage – 2u servers the sweet spot • 6-24 drive bays for both SSDs and HDDs • Slots for 10Gbps Ethernet and PCIe SSDs • Blades unsuitable – 1-2 SFF disk bays – Mezzanine PCIe SSDs generally >$8000 • High density servers can work
  17. 17. Challenges to SDS • Purchasing politics and budgets – Everyone likes to point at where their money went, especially storage guys – So who’s budget • Easier if savings big enough that storage+compute now costs ≤ storage or compute – VDI can be camel’s nose because it needs dedicated infrastructure anyway
  18. 18. Operational Challenges • Storage guys are paranoid for good reason – Storage is persistent and so are storage mistakes • Server guys are less paranoid • VSAN with default 2 copies – VMware admin takes 1 server offline – The 1 disk with data fails.
  19. 19. ServerSAN vs Dedicated Storage • VMware’s benchmark config – 32 nodes – 400GB SSD and 7 1TB drives each – VSAN cost ~$11,000/server for 2.3TB useable (73TB total) • Best of breed dedicated storage – Tintri T650 • 33.6TB usable $160,000 • ~100,000 real IOPS • PerVM snaps and replication
  20. 20. Questions and Contact • Contact info: – Hmarks@deepstorage.net – @DeepStoragenet onTwitter
  1. A particular slide catching your eye?

    Clipping is a handy way to collect important slides you want to go back to later.

×