4. 07.05.2015 4
Overview
Business Marketplace
• https://portal.telekomcloud.com/
• SaaS Applications from Software Partners (ISVs) and
DT offered to SME customers
• i.e.Saperion, Sage,PadCloud, Teamlike, Fastbill, Imeet,
Weclapp, SilverERP, Teamdisk ...
• Complements othercloud offerings from Deutsche Telekom (Enterprise cloud from T-Systems, Cisco
Intercloud, Mediencenter etc.)
• IaaS platform based only on Open Source technologies like OpenStack, CEPH and Linux
• Project started in 2012 with OS Essex, CEPH in production since 3/2013 (bobtail)
5. 07.05.2015–Strictlyconfidential,Confidential,Internal– Author/Presentationtitle 5
Overview
why opensource?Why ceph?
• no vendorlock in!
• easier to change and adapt newtechnologies / concepts - more independent from vendor priorities
• low cost of ownership and operation, utilizing commodity hardware and Open Source
• no license fees - but professional support
• modular and horizontally scalable platform
• automation and flexibility allow for faster deployment cycles, than in traditional hosting
• control overopen sourcecode - faster bug fixing and feature delivery
19. 07.05.2015 19
lessonslearned
operational experience
• Chose your hardware well !!
• I,e. RAID and hard disks -> enterprise gradedisks (desktop HDs aremissing important features like TLER/ERC)
• CPU/RAM planning: calculate 1GHz CPU powerand 2GB RAM persingle OSD
• pick nodes with lowstorage capacity density for smaller clusters
• At least 5 nodes for a 3 replica cluster (i.e.for PoC, testing and development purposes)
• Cluster configuration “adjustments”:
• increasing PG num> impact on cluster becauseofmassive data migration
• Rolling software updates / upgrades workedperfectly
• CEPH: has a character– buthighly reliable - neverlost data
20. 07.05.2015 20
lessonslearned
operational experience
• Failed / ”Slow”disks
• Inconsistent PGs
• Incomplete PGs
• RBD pool configured with min_size=2
• Blocks IO operations to the pool / cluster
• fixed in Hammer (allows PG replication while replica level below min_size pool/OSD)
24. 07.05.2015 24
Overview
SCALE OUT Project
+40%
Current overall capacity:
~60 storage nodes
5,4 PB Storage Gross
~0,5 PB S3 Storage Net
Planned Capacity for 2015:
~90 storage nodes
7,5 PB Storage Gross
~1,5 PB S3 Storage Net
25. 07.05.2015 25
Future setup
scale out project
• 2 physically separated rooms
• Data distributed accordingthe rule
• not more than 2 replicas in - oneroom not more than 1
replica in onerack
27. 07.05.2015 27
Future setup
dreams
• cachetiering
• make use of shiny newSSDs in a hot zone / cachepool
• SSD pools
• Openstack live migration for VMs(boot from rbd volume)