SlideShare a Scribd company logo
1 of 37
Download to read offline
WHAT’S NEW IN CEPH
Ceph Day Silicon Valley
at
The University of California Santa Cruz Silicon Valley
Campus
Neha Ojha, Patrick Donnelly
PRIORITIES
● Community
● Management and usability
● Performance
● Core Ceph
○ RADOS
○ RBD
○ RGW
○ CephFS
● Container platforms
● Dashboard
COMMUNITY
CEPHALOCON
● Cephalocon APAC
○ Mar 2018
○ Community organized (not us!)
○ ~1000 attendees
○ Overwhelming amount of technical content
○ Highlighted huge opportunity to build developer community in APAC
● Next Cephalocon
○ Europe in the spring!
○ 2 days, 3-4 tracks
○ Finalizing plans (venue and timing…)
CEPH DAYS
● One day, regional events
○ https://ceph.com/cephday
● Upcoming
○ Ceph Day Berlin - Nov 12 (day before OpenStack Summit)
MANAGEMENT AND USABILITY
AUTOMATION AND MANAGEMENT
● Focus on “hands off” operation
● Hidden/automated pg_num selection
○ Enable pg_num decreases as well as increases
○ Automated, hands-off management of pool pg_num based on utilization, workload, etc.
● Automated tuning
○ Manage cache sizes, configurables, etc based on user-provided memory envelope
○ Conditional defaults of performance-related functions based on device types
● Additional guard rails
○ ‘ceph osd safe-to-destroy’, ‘ok-to-stop’ checks
○ Include safe-to-destroy check in ‘ceph osd destroy/purge’
TELEMETRY AND INSIGHTS
● Phone-home via upstream telemetry or downstream Insights
● Centralized collection of crash reports
○ Alerting for transient failures (daemon crash + restart)
○ Phoned home to track failures and bugs in the wild, prioritize bugs, etc.
● Enablement for proactive/preemptive support
● Disk failure prediction
○ Preemptive evacuation of failing devices
○ Self-contained prediction or higher quality prediction via SaaS GRPC - thanks to Rick Chen
@ProphetStor
RADOS
NEW IN MIMIC
● Centralized Config
○ Stored on monitors - one place to update, validate, see history
○ Manageable by the Dashboard in future
● Ceph-volume
○ Replacement for ceph-disk - no udev, predictable, no longer race condition prone
● Asynchronous Recovery
○ No blocking I/O for recovery
○ Better client I/O performance during recovery
PERFORMANCE
HARDWARE IS CHANGING
● The world is moving to flash and NVMe
○ Substantially lower latencies and higher throughputs
○ Capacities rivaling HDDs
○ Massively lower $/IOPS
○ $/bit is still a few years away
● HDD-based storage is becoming niche space
○ WD just shut down a factory this week
● Storage software must adapt to survive
○ Not only top-line performance (max IOPS); also IOPS per CPU core
PROJECT CRIMSON
● Reimplementing Ceph OSD data path
○ Kefu, Casey (Red Hat)
○ Chunmei, Lisa, Yingxin (Intel)
● Seastar (from Scylla community)
○ run to completion model
○ explicit sharding of data and processing across CPU cores
● DPDK, SPDK bring network and storage drivers into userspace
● Current status
○ Working messenger, various infrastructure pieces (e.g., config mgmt, auth)
○ Shared caches and simplified data path coming next
○ Initial prototypes will be against MemStore (non-blocking)
● Kefu is presenting progress at Scylla Summit 2018 (Nov 6-7 in SF)
OTHER PERFORMANCE
● New logging infrastructure (lttng-based)
● Sample-based tracing
○ OpenTracing, Jaeger
● OpTracker improvements
● Mutex -> std::mutex
○ Compile out lockdep for production builds
● Auth signature check optimizations
● BlueStore allocator improvements
MSGR2
● New on-wire protocol
○ Improved protocol feature negotiation
○ Multiplexing (maybe, eventually)
○ IANA-assigned port number for mons (ce4h = 3300)
● Encryption over the wire
● Probably no signature support?
○ fast mode (with no cryptographic integrity checks) or
○ secure mode (full encryption)
● Dual-stack support (IPv4 + IPv6)
● Kerberos authentication
○ Use kerberos credentials to map to Ceph roles, issue Ceph CLI commands
MISC
● OSD memory stability
○ PG log length limited
○ OSD internally adjusts caches to stay within bounds
● QoS
○ min, max, and priority-based limits and reservations
● Improved introspection, utilization metrics (Nautilus)
● Future: ‘ceph top’
○ Sampling-based real-time view of client workload
RBD
BLOCK
● Improved orchestration/management of async mirroring
○ Point-in-time consistent DR with failover/failback, mirrored snapshots, etc.
● Security namespaces for RBD
○ Simple RBAC-style CephX cap profiles
○ Basic provisioned space quotas
● Client-side caching - PR in progress
● Transparent live image migration
● Simplified pool- and image-level configuration overrides
● Future: ‘rbd top’ for real-time per-image workload
RGW
RGW / OBJECT
● Ongoing performance and scalability improvements
● Security (Nautilus)
○ Cloud-based (STS, keystone) and enterprise (OPA, kerberos/AD)
● pub/sub API (Nautilus)
○ Recently prototyped for OpenWhisk project; targets Nautilus
● Sync from public cloud (Nautilus)
○ RGW sync to S3 added in Mimic; this adds other direction
● Tiering
○ Push individual objects to external cloud (Nautilus)
○ Push entire buckets to external cloud
MULTI- AND HYBRID CLOUD
● Next generation of applications will primarily consume object storage
○ Block is great for backing VMs and containers,
○ and file will serve legacy workloads and even scale to huge data sets,
○ but most cat pictures/videos/etc will land in objects
● Modern IT infrastructure spans multiple data centers, public/private clouds
● In the public cloud, it will be hard to beat native storage pricing (e.g., S3)
● RGW should expand to encompass “data services”
○ Data portability (especially paired with application portability)
○ Data placement (vs capacity, bandwidth, compliance/regulatory regimes)
○ Lifecycle management
○ Introspection (what am I storing and where?)
○ Policy and automation for all of the above
CEPHFS
MULTI-MDS ROBUSTNESS
● Multi-MDS stable and supported since Luminous
○ Mimic makes adding/removing MDS easier.
● Snapshots stable and supported since Mimic.
○ Requires v4.17+ kernel for kclient.
● Lots of work remaining to make dynamic load balancer more robust
○ Generate realistic workloads at scale (many MDS daemons, lots of RAM for each)
○ Balancer tuning
24
● Directory (subtree) snapshots anywhere in the
file system hierarchy.
● Note: for the kernel client, use the latest
kernel.
● Mimic will change the default allow_new_snaps
to true. Existing file systems must turn the flag
on after upgrade.
Credit to Zheng Yan (Red Hat).
$ cat foo/file1
Hello
$ mkdir /cephfs/foo/.snap/2017-10-06
$ echo “world!” >> foo/file1
$ cat foo/file1
Hello
World!
$ cat foo/.snap/2017-10-06/file1
Hello
bar
file2
/
foo
file1
SNAPSHOTS STABLE
25
● Byte limit:
$ setfattr -n ceph.quota.max_bytes -v 100000000 /some/dir
● File limit:
$ setfattr -n ceph.quota.max_files -v 10000 /some/dir
● Caveat: limits are enforced eventually.
● Kernel changes not yet merged upstream
Cooperatively fixed by Luis Henriques (SUSE) and Zheng Yan
(Red Hat).
dir2
file2
quota config:
max_files=20,max_bytes=1048576
/
dir1
file1
quota config:
max_files=200,max_bytes=10485760
(KERNEL) QUOTA SUPPORT
● Bring volume/subvolume management into ceph-mgr
○ Change ceph-volume-client.py to simply wrap new functions
○ “Formalize” volume/subvolume concepts
○ Modify Rook, Kubernetes provisioners, Manila to all consume/share same interface
● Scale-out NFS
○ Cluster-managed ganesha gateways with active/active
○ Robust cluster-coherent NFSv4 recovery
○ Protect Ceph from untrusted clients
VOLUME MANAGEMENT & NFS GATEWAYS
cephfs-shell
● New alternative client that doesn’t
requiring mounting CephFS
● Outreachy project Summer 2018 by
Pavani Rajula.
$ cephfs-shell
CephFS:~/>>> mkdir foo
CephFS:~/>>> cd foo
CephFS:~/foo>>> put /etc/hosts hosts
CephFS:~/foo>>> cat hosts
127.0.0.1 localhost.localdomain
localhost
...
...
PERFORMANCE
● Buffered create/unlink by clients
○ Use write/buffer capability on directory
○ Pass out allocated inodes to each session for creating files
○ Asynchronously flush create/unlink to MDS
● Sharding metadata within an MDS (“sub-ranks”) to scale across cores
○ Mostly lock-less is a design goal
○ Fast zero-copy metadata/cap export between sub-ranks
CONTAINER PLATFORMS
ROOK
● Native, robust operator for kubernetes and openshift
● Intelligent deployment of Ceph daemons
○ e.g., add/remove/move mon daemons while maintaining quorum
○ e.g., intelligently schedule RGW/iSCSI/NFS gateways across nodes
● Integration with SDN functions
○ e.g., schedule Ganesha NFS gateways and attach them to tenant Neutron networks via Kuryr
● Upgrade orchestration
○ Update Rook operator pod (triggered via CLI or dashboard?)
○ Rook updates Ceph daemons in prescribed order, gated with health, safety, availability checks
○ Rook manages any release-specific steps (like forcing scrubs, declaring upgrade “done”, etc.)
● Existing and happy user community, CNCF member project
K8S STRATEGY
● Align with Ceph with kubernetes community interest/adoption (Rook, CNCF)
● Enable early kubernetes/openshift adopters (IT users, not IT)
○ Provision Rook/Ceph clusters layered over existing infrastructure
● Displace of legacy storage on prem for kubernetes
○ Allow multiple kube clusters, Rook instances to share an external Ceph cluster
● Enable kubernetes as a service (e.g., provided by IT)
○ Enable multi-tenant workflows for Rook storage classes (e.g., Pools)
● Maximize consistency of experience on public cloud
● Expose underlying Ceph federation capabilities (especially object)
● Also: use kubernetes “under the hood” for standalone Ceph
CEPH DASHBOARD
DASHBOARD
● Converged community investment on built-in web dashboard
○ Hybrid of openATTIC and John’s original dashboard proof of concept
○ Self-hosted by ceph-mgr with easy, tight integration with other cluster management and
automation functions
● Currently mostly “logical” cluster functions
○ Management of Ceph services (pools, RBD images, file systems, configuration, etc.)
○ Subsuming ceph-metrics and openATTIC grafana metrics
● Orchestrator abstraction
○ Allow ceph-mgr and dashboard (or CLI, APIs) to drive ansible, Rook, DeepSea, etc
○ Provision or deprovision Ceph daemons, add/remove nodes, replace disks, etc.
○ Abstracts/hides choice of orchestration layer, enabling generalized automation, GUI, docs, UX
● Indirectly laying foundation for stable and versioned management API
Questions?
Neha Ojha - nojha@redhat.com
Patrick Donnelly - pdonnell@redhat.com
Appendix
CEPH-CSI
● Replace upstream Kubernetes and Rook flexvol with ceph-csi
● Development driven by CERN, Cisco, with help from Huamin (Red Hat)
● Stretch goal is to replace Rook’s flexvol with ceph-csi in next 0.9 release
MULTI-CLUSTER CEPHFS
● Geo-replication
○ Loosely-consistent and point-in-time consistent DR replication (Nautilus)
○ Active/active async replication, with associated consistency caveats?
● Sync and share (NextCloud) integration
○ Concurrent access via usual POSIX (kcephfs, NFS, etc) mounts and NextCloud to same files,
with revisions

More Related Content

What's hot

My personal journey through the World of Open Source! How What Was Old Beco...
My personal journey through  the World of Open Source!  How What Was Old Beco...My personal journey through  the World of Open Source!  How What Was Old Beco...
My personal journey through the World of Open Source! How What Was Old Beco...Ceph Community
 
Ceph Tech Talk: Ceph at DigitalOcean
Ceph Tech Talk: Ceph at DigitalOceanCeph Tech Talk: Ceph at DigitalOcean
Ceph Tech Talk: Ceph at DigitalOceanCeph Community
 
Disk health prediction for Ceph
Disk health prediction for CephDisk health prediction for Ceph
Disk health prediction for CephCeph Community
 
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...Ceph Community
 
Ceph, the future of Storage - Sage Weil
Ceph, the future of Storage - Sage WeilCeph, the future of Storage - Sage Weil
Ceph, the future of Storage - Sage WeilCeph Community
 
Red Hat Gluster Storage, Container Storage and CephFS Plans
Red Hat Gluster Storage, Container Storage and CephFS PlansRed Hat Gluster Storage, Container Storage and CephFS Plans
Red Hat Gluster Storage, Container Storage and CephFS PlansRed_Hat_Storage
 
Performance optimization for all flash based on aarch64 v2.0
Performance optimization for all flash based on aarch64 v2.0Performance optimization for all flash based on aarch64 v2.0
Performance optimization for all flash based on aarch64 v2.0Ceph Community
 
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red Hat
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red HatHyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red Hat
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red HatOpenStack
 
Red Hat Ceph Storage Roadmap: January 2016
Red Hat Ceph Storage Roadmap: January 2016Red Hat Ceph Storage Roadmap: January 2016
Red Hat Ceph Storage Roadmap: January 2016Red_Hat_Storage
 
Red Hat Enterprise Linux: Open, hyperconverged infrastructure
Red Hat Enterprise Linux: Open, hyperconverged infrastructureRed Hat Enterprise Linux: Open, hyperconverged infrastructure
Red Hat Enterprise Linux: Open, hyperconverged infrastructureRed_Hat_Storage
 
2021.06. Ceph Project Update
2021.06. Ceph Project Update2021.06. Ceph Project Update
2021.06. Ceph Project UpdateCeph Community
 
Ceph Month 2021: RADOS Update
Ceph Month 2021: RADOS UpdateCeph Month 2021: RADOS Update
Ceph Month 2021: RADOS UpdateCeph Community
 
Red Hat Storage Day New York - What's New in Red Hat Ceph Storage
Red Hat Storage Day New York - What's New in Red Hat Ceph StorageRed Hat Storage Day New York - What's New in Red Hat Ceph Storage
Red Hat Storage Day New York - What's New in Red Hat Ceph StorageRed_Hat_Storage
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storageAndrew Underwood
 
2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific DashboardCeph Community
 
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...Ceph Community
 
Red Hat Gluster Storage - Direction, Roadmap and Use-Cases
Red Hat Gluster Storage - Direction, Roadmap and Use-CasesRed Hat Gluster Storage - Direction, Roadmap and Use-Cases
Red Hat Gluster Storage - Direction, Roadmap and Use-CasesRed_Hat_Storage
 
2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph2019.06.27 Intro to Ceph
2019.06.27 Intro to CephCeph Community
 
Red Hat Storage for Mere Mortals
Red Hat Storage for Mere MortalsRed Hat Storage for Mere Mortals
Red Hat Storage for Mere MortalsRed_Hat_Storage
 

What's hot (20)

My personal journey through the World of Open Source! How What Was Old Beco...
My personal journey through  the World of Open Source!  How What Was Old Beco...My personal journey through  the World of Open Source!  How What Was Old Beco...
My personal journey through the World of Open Source! How What Was Old Beco...
 
Ceph Tech Talk: Ceph at DigitalOcean
Ceph Tech Talk: Ceph at DigitalOceanCeph Tech Talk: Ceph at DigitalOcean
Ceph Tech Talk: Ceph at DigitalOcean
 
Disk health prediction for Ceph
Disk health prediction for CephDisk health prediction for Ceph
Disk health prediction for Ceph
 
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
 
Ceph, the future of Storage - Sage Weil
Ceph, the future of Storage - Sage WeilCeph, the future of Storage - Sage Weil
Ceph, the future of Storage - Sage Weil
 
Red Hat Gluster Storage, Container Storage and CephFS Plans
Red Hat Gluster Storage, Container Storage and CephFS PlansRed Hat Gluster Storage, Container Storage and CephFS Plans
Red Hat Gluster Storage, Container Storage and CephFS Plans
 
Performance optimization for all flash based on aarch64 v2.0
Performance optimization for all flash based on aarch64 v2.0Performance optimization for all flash based on aarch64 v2.0
Performance optimization for all flash based on aarch64 v2.0
 
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red Hat
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red HatHyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red Hat
Hyperconverged Cloud, Not just a toy anymore - Andrew Hatfield, Red Hat
 
Red Hat Ceph Storage Roadmap: January 2016
Red Hat Ceph Storage Roadmap: January 2016Red Hat Ceph Storage Roadmap: January 2016
Red Hat Ceph Storage Roadmap: January 2016
 
Red Hat Enterprise Linux: Open, hyperconverged infrastructure
Red Hat Enterprise Linux: Open, hyperconverged infrastructureRed Hat Enterprise Linux: Open, hyperconverged infrastructure
Red Hat Enterprise Linux: Open, hyperconverged infrastructure
 
2021.06. Ceph Project Update
2021.06. Ceph Project Update2021.06. Ceph Project Update
2021.06. Ceph Project Update
 
Ceph Month 2021: RADOS Update
Ceph Month 2021: RADOS UpdateCeph Month 2021: RADOS Update
Ceph Month 2021: RADOS Update
 
Red Hat Storage Day New York - What's New in Red Hat Ceph Storage
Red Hat Storage Day New York - What's New in Red Hat Ceph StorageRed Hat Storage Day New York - What's New in Red Hat Ceph Storage
Red Hat Storage Day New York - What's New in Red Hat Ceph Storage
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storage
 
2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard2021.02 new in Ceph Pacific Dashboard
2021.02 new in Ceph Pacific Dashboard
 
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
 
Red Hat Gluster Storage - Direction, Roadmap and Use-Cases
Red Hat Gluster Storage - Direction, Roadmap and Use-CasesRed Hat Gluster Storage - Direction, Roadmap and Use-Cases
Red Hat Gluster Storage - Direction, Roadmap and Use-Cases
 
Red Hat Storage Roadmap
Red Hat Storage RoadmapRed Hat Storage Roadmap
Red Hat Storage Roadmap
 
2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph2019.06.27 Intro to Ceph
2019.06.27 Intro to Ceph
 
Red Hat Storage for Mere Mortals
Red Hat Storage for Mere MortalsRed Hat Storage for Mere Mortals
Red Hat Storage for Mere Mortals
 

Similar to What's New with Ceph - Ceph Day Silicon Valley

OpenEBS hangout #4
OpenEBS hangout #4OpenEBS hangout #4
OpenEBS hangout #4OpenEBS
 
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storage
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storageWebinar: OpenEBS - Still Free and now FASTEST Kubernetes storage
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storageMayaData Inc
 
London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph Ceph Community
 
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph Ceph Community
 
Ceph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's CephCeph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's CephCeph Community
 
OpenStack Best Practices and Considerations - terasky tech day
OpenStack Best Practices and Considerations  - terasky tech dayOpenStack Best Practices and Considerations  - terasky tech day
OpenStack Best Practices and Considerations - terasky tech dayArthur Berezin
 
Introduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackIntroduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackOpenStack_Online
 
Hadoop 3 @ Hadoop Summit San Jose 2017
Hadoop 3 @ Hadoop Summit San Jose 2017Hadoop 3 @ Hadoop Summit San Jose 2017
Hadoop 3 @ Hadoop Summit San Jose 2017Junping Du
 
Apache Hadoop 3.0 Community Update
Apache Hadoop 3.0 Community UpdateApache Hadoop 3.0 Community Update
Apache Hadoop 3.0 Community UpdateDataWorks Summit
 
Ceph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Community
 
Ceph: A decade in the making and still going strong
Ceph: A decade in the making and still going strongCeph: A decade in the making and still going strong
Ceph: A decade in the making and still going strongPatrick McGarry
 
Data Policies for the Kafka-API with WebAssembly | Alexander Gallego, Vectorized
Data Policies for the Kafka-API with WebAssembly | Alexander Gallego, VectorizedData Policies for the Kafka-API with WebAssembly | Alexander Gallego, Vectorized
Data Policies for the Kafka-API with WebAssembly | Alexander Gallego, VectorizedHostedbyConfluent
 
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and OutlookLinux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and OutlookDanny Al-Gaaf
 
[WSO2Con EU 2018] Architecting for a Container Native Environment
[WSO2Con EU 2018] Architecting for a Container Native Environment[WSO2Con EU 2018] Architecting for a Container Native Environment
[WSO2Con EU 2018] Architecting for a Container Native EnvironmentWSO2
 
The road to enterprise ready open stack storage as service
The road to enterprise ready open stack storage as serviceThe road to enterprise ready open stack storage as service
The road to enterprise ready open stack storage as serviceSean Cohen
 
Netflix Container Scheduling and Execution - QCon New York 2016
Netflix Container Scheduling and Execution - QCon New York 2016Netflix Container Scheduling and Execution - QCon New York 2016
Netflix Container Scheduling and Execution - QCon New York 2016aspyker
 
Scheduling a fuller house - Talk at QCon NY 2016
Scheduling a fuller house - Talk at QCon NY 2016Scheduling a fuller house - Talk at QCon NY 2016
Scheduling a fuller house - Talk at QCon NY 2016Sharma Podila
 

Similar to What's New with Ceph - Ceph Day Silicon Valley (20)

OpenEBS hangout #4
OpenEBS hangout #4OpenEBS hangout #4
OpenEBS hangout #4
 
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storage
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storageWebinar: OpenEBS - Still Free and now FASTEST Kubernetes storage
Webinar: OpenEBS - Still Free and now FASTEST Kubernetes storage
 
London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph London Ceph Day Keynote: Building Tomorrow's Ceph
London Ceph Day Keynote: Building Tomorrow's Ceph
 
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
Ceph Day Santa Clara: Keynote: Building Tomorrow's Ceph
 
Ceph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's CephCeph Day NYC: Building Tomorrow's Ceph
Ceph Day NYC: Building Tomorrow's Ceph
 
OpenStack Best Practices and Considerations - terasky tech day
OpenStack Best Practices and Considerations  - terasky tech dayOpenStack Best Practices and Considerations  - terasky tech day
OpenStack Best Practices and Considerations - terasky tech day
 
Introduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackIntroduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStack
 
Hadoop 3 @ Hadoop Summit San Jose 2017
Hadoop 3 @ Hadoop Summit San Jose 2017Hadoop 3 @ Hadoop Summit San Jose 2017
Hadoop 3 @ Hadoop Summit San Jose 2017
 
Apache Hadoop 3.0 Community Update
Apache Hadoop 3.0 Community UpdateApache Hadoop 3.0 Community Update
Apache Hadoop 3.0 Community Update
 
Ceph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade in
 
Ceph: A decade in the making and still going strong
Ceph: A decade in the making and still going strongCeph: A decade in the making and still going strong
Ceph: A decade in the making and still going strong
 
Data Policies for the Kafka-API with WebAssembly | Alexander Gallego, Vectorized
Data Policies for the Kafka-API with WebAssembly | Alexander Gallego, VectorizedData Policies for the Kafka-API with WebAssembly | Alexander Gallego, Vectorized
Data Policies for the Kafka-API with WebAssembly | Alexander Gallego, Vectorized
 
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and OutlookLinux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
 
[WSO2Con EU 2018] Architecting for a Container Native Environment
[WSO2Con EU 2018] Architecting for a Container Native Environment[WSO2Con EU 2018] Architecting for a Container Native Environment
[WSO2Con EU 2018] Architecting for a Container Native Environment
 
The road to enterprise ready open stack storage as service
The road to enterprise ready open stack storage as serviceThe road to enterprise ready open stack storage as service
The road to enterprise ready open stack storage as service
 
Discoblocks.pptx.pdf
Discoblocks.pptx.pdfDiscoblocks.pptx.pdf
Discoblocks.pptx.pdf
 
DEVIEW 2013
DEVIEW 2013DEVIEW 2013
DEVIEW 2013
 
Welcome to icehouse
Welcome to icehouseWelcome to icehouse
Welcome to icehouse
 
Netflix Container Scheduling and Execution - QCon New York 2016
Netflix Container Scheduling and Execution - QCon New York 2016Netflix Container Scheduling and Execution - QCon New York 2016
Netflix Container Scheduling and Execution - QCon New York 2016
 
Scheduling a fuller house - Talk at QCon NY 2016
Scheduling a fuller house - Talk at QCon NY 2016Scheduling a fuller house - Talk at QCon NY 2016
Scheduling a fuller house - Talk at QCon NY 2016
 

Recently uploaded

What is Fashion PLM and Why Do You Need It
What is Fashion PLM and Why Do You Need ItWhat is Fashion PLM and Why Do You Need It
What is Fashion PLM and Why Do You Need ItWave PLM
 
Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...
Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...
Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...kellynguyen01
 
Learn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdf
Learn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdfLearn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdf
Learn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdfkalichargn70th171
 
The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...
The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...
The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...ICS
 
Unit 1.1 Excite Part 1, class 9, cbse...
Unit 1.1 Excite Part 1, class 9, cbse...Unit 1.1 Excite Part 1, class 9, cbse...
Unit 1.1 Excite Part 1, class 9, cbse...aditisharan08
 
Advancing Engineering with AI through the Next Generation of Strategic Projec...
Advancing Engineering with AI through the Next Generation of Strategic Projec...Advancing Engineering with AI through the Next Generation of Strategic Projec...
Advancing Engineering with AI through the Next Generation of Strategic Projec...OnePlan Solutions
 
The Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdf
The Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdfThe Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdf
The Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdfkalichargn70th171
 
XpertSolvers: Your Partner in Building Innovative Software Solutions
XpertSolvers: Your Partner in Building Innovative Software SolutionsXpertSolvers: Your Partner in Building Innovative Software Solutions
XpertSolvers: Your Partner in Building Innovative Software SolutionsMehedi Hasan Shohan
 
DNT_Corporate presentation know about us
DNT_Corporate presentation know about usDNT_Corporate presentation know about us
DNT_Corporate presentation know about usDynamic Netsoft
 
5 Signs You Need a Fashion PLM Software.pdf
5 Signs You Need a Fashion PLM Software.pdf5 Signs You Need a Fashion PLM Software.pdf
5 Signs You Need a Fashion PLM Software.pdfWave PLM
 
HR Software Buyers Guide in 2024 - HRSoftware.com
HR Software Buyers Guide in 2024 - HRSoftware.comHR Software Buyers Guide in 2024 - HRSoftware.com
HR Software Buyers Guide in 2024 - HRSoftware.comFatema Valibhai
 
Cloud Management Software Platforms: OpenStack
Cloud Management Software Platforms: OpenStackCloud Management Software Platforms: OpenStack
Cloud Management Software Platforms: OpenStackVICTOR MAESTRE RAMIREZ
 
Unveiling the Tech Salsa of LAMs with Janus in Real-Time Applications
Unveiling the Tech Salsa of LAMs with Janus in Real-Time ApplicationsUnveiling the Tech Salsa of LAMs with Janus in Real-Time Applications
Unveiling the Tech Salsa of LAMs with Janus in Real-Time ApplicationsAlberto González Trastoy
 
ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...
ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...
ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...Christina Lin
 
The Evolution of Karaoke From Analog to App.pdf
The Evolution of Karaoke From Analog to App.pdfThe Evolution of Karaoke From Analog to App.pdf
The Evolution of Karaoke From Analog to App.pdfPower Karaoke
 
KnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptx
KnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptxKnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptx
KnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptxTier1 app
 
Project Based Learning (A.I).pptx detail explanation
Project Based Learning (A.I).pptx detail explanationProject Based Learning (A.I).pptx detail explanation
Project Based Learning (A.I).pptx detail explanationkaushalgiri8080
 
Der Spagat zwischen BIAS und FAIRNESS (2024)
Der Spagat zwischen BIAS und FAIRNESS (2024)Der Spagat zwischen BIAS und FAIRNESS (2024)
Der Spagat zwischen BIAS und FAIRNESS (2024)OPEN KNOWLEDGE GmbH
 
Optimizing AI for immediate response in Smart CCTV
Optimizing AI for immediate response in Smart CCTVOptimizing AI for immediate response in Smart CCTV
Optimizing AI for immediate response in Smart CCTVshikhaohhpro
 
Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝soniya singh
 

Recently uploaded (20)

What is Fashion PLM and Why Do You Need It
What is Fashion PLM and Why Do You Need ItWhat is Fashion PLM and Why Do You Need It
What is Fashion PLM and Why Do You Need It
 
Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...
Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...
Short Story: Unveiling the Reasoning Abilities of Large Language Models by Ke...
 
Learn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdf
Learn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdfLearn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdf
Learn the Fundamentals of XCUITest Framework_ A Beginner's Guide.pdf
 
The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...
The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...
The Real-World Challenges of Medical Device Cybersecurity- Mitigating Vulnera...
 
Unit 1.1 Excite Part 1, class 9, cbse...
Unit 1.1 Excite Part 1, class 9, cbse...Unit 1.1 Excite Part 1, class 9, cbse...
Unit 1.1 Excite Part 1, class 9, cbse...
 
Advancing Engineering with AI through the Next Generation of Strategic Projec...
Advancing Engineering with AI through the Next Generation of Strategic Projec...Advancing Engineering with AI through the Next Generation of Strategic Projec...
Advancing Engineering with AI through the Next Generation of Strategic Projec...
 
The Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdf
The Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdfThe Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdf
The Essentials of Digital Experience Monitoring_ A Comprehensive Guide.pdf
 
XpertSolvers: Your Partner in Building Innovative Software Solutions
XpertSolvers: Your Partner in Building Innovative Software SolutionsXpertSolvers: Your Partner in Building Innovative Software Solutions
XpertSolvers: Your Partner in Building Innovative Software Solutions
 
DNT_Corporate presentation know about us
DNT_Corporate presentation know about usDNT_Corporate presentation know about us
DNT_Corporate presentation know about us
 
5 Signs You Need a Fashion PLM Software.pdf
5 Signs You Need a Fashion PLM Software.pdf5 Signs You Need a Fashion PLM Software.pdf
5 Signs You Need a Fashion PLM Software.pdf
 
HR Software Buyers Guide in 2024 - HRSoftware.com
HR Software Buyers Guide in 2024 - HRSoftware.comHR Software Buyers Guide in 2024 - HRSoftware.com
HR Software Buyers Guide in 2024 - HRSoftware.com
 
Cloud Management Software Platforms: OpenStack
Cloud Management Software Platforms: OpenStackCloud Management Software Platforms: OpenStack
Cloud Management Software Platforms: OpenStack
 
Unveiling the Tech Salsa of LAMs with Janus in Real-Time Applications
Unveiling the Tech Salsa of LAMs with Janus in Real-Time ApplicationsUnveiling the Tech Salsa of LAMs with Janus in Real-Time Applications
Unveiling the Tech Salsa of LAMs with Janus in Real-Time Applications
 
ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...
ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...
ODSC - Batch to Stream workshop - integration of Apache Spark, Cassandra, Pos...
 
The Evolution of Karaoke From Analog to App.pdf
The Evolution of Karaoke From Analog to App.pdfThe Evolution of Karaoke From Analog to App.pdf
The Evolution of Karaoke From Analog to App.pdf
 
KnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptx
KnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptxKnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptx
KnowAPIs-UnknownPerf-jaxMainz-2024 (1).pptx
 
Project Based Learning (A.I).pptx detail explanation
Project Based Learning (A.I).pptx detail explanationProject Based Learning (A.I).pptx detail explanation
Project Based Learning (A.I).pptx detail explanation
 
Der Spagat zwischen BIAS und FAIRNESS (2024)
Der Spagat zwischen BIAS und FAIRNESS (2024)Der Spagat zwischen BIAS und FAIRNESS (2024)
Der Spagat zwischen BIAS und FAIRNESS (2024)
 
Optimizing AI for immediate response in Smart CCTV
Optimizing AI for immediate response in Smart CCTVOptimizing AI for immediate response in Smart CCTV
Optimizing AI for immediate response in Smart CCTV
 
Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Naraina Delhi 💯Call Us 🔝8264348440🔝
 

What's New with Ceph - Ceph Day Silicon Valley

  • 1. WHAT’S NEW IN CEPH Ceph Day Silicon Valley at The University of California Santa Cruz Silicon Valley Campus Neha Ojha, Patrick Donnelly
  • 2. PRIORITIES ● Community ● Management and usability ● Performance ● Core Ceph ○ RADOS ○ RBD ○ RGW ○ CephFS ● Container platforms ● Dashboard
  • 4. CEPHALOCON ● Cephalocon APAC ○ Mar 2018 ○ Community organized (not us!) ○ ~1000 attendees ○ Overwhelming amount of technical content ○ Highlighted huge opportunity to build developer community in APAC ● Next Cephalocon ○ Europe in the spring! ○ 2 days, 3-4 tracks ○ Finalizing plans (venue and timing…)
  • 5. CEPH DAYS ● One day, regional events ○ https://ceph.com/cephday ● Upcoming ○ Ceph Day Berlin - Nov 12 (day before OpenStack Summit)
  • 7. AUTOMATION AND MANAGEMENT ● Focus on “hands off” operation ● Hidden/automated pg_num selection ○ Enable pg_num decreases as well as increases ○ Automated, hands-off management of pool pg_num based on utilization, workload, etc. ● Automated tuning ○ Manage cache sizes, configurables, etc based on user-provided memory envelope ○ Conditional defaults of performance-related functions based on device types ● Additional guard rails ○ ‘ceph osd safe-to-destroy’, ‘ok-to-stop’ checks ○ Include safe-to-destroy check in ‘ceph osd destroy/purge’
  • 8. TELEMETRY AND INSIGHTS ● Phone-home via upstream telemetry or downstream Insights ● Centralized collection of crash reports ○ Alerting for transient failures (daemon crash + restart) ○ Phoned home to track failures and bugs in the wild, prioritize bugs, etc. ● Enablement for proactive/preemptive support ● Disk failure prediction ○ Preemptive evacuation of failing devices ○ Self-contained prediction or higher quality prediction via SaaS GRPC - thanks to Rick Chen @ProphetStor
  • 10. NEW IN MIMIC ● Centralized Config ○ Stored on monitors - one place to update, validate, see history ○ Manageable by the Dashboard in future ● Ceph-volume ○ Replacement for ceph-disk - no udev, predictable, no longer race condition prone ● Asynchronous Recovery ○ No blocking I/O for recovery ○ Better client I/O performance during recovery
  • 12. HARDWARE IS CHANGING ● The world is moving to flash and NVMe ○ Substantially lower latencies and higher throughputs ○ Capacities rivaling HDDs ○ Massively lower $/IOPS ○ $/bit is still a few years away ● HDD-based storage is becoming niche space ○ WD just shut down a factory this week ● Storage software must adapt to survive ○ Not only top-line performance (max IOPS); also IOPS per CPU core
  • 13. PROJECT CRIMSON ● Reimplementing Ceph OSD data path ○ Kefu, Casey (Red Hat) ○ Chunmei, Lisa, Yingxin (Intel) ● Seastar (from Scylla community) ○ run to completion model ○ explicit sharding of data and processing across CPU cores ● DPDK, SPDK bring network and storage drivers into userspace ● Current status ○ Working messenger, various infrastructure pieces (e.g., config mgmt, auth) ○ Shared caches and simplified data path coming next ○ Initial prototypes will be against MemStore (non-blocking) ● Kefu is presenting progress at Scylla Summit 2018 (Nov 6-7 in SF)
  • 14. OTHER PERFORMANCE ● New logging infrastructure (lttng-based) ● Sample-based tracing ○ OpenTracing, Jaeger ● OpTracker improvements ● Mutex -> std::mutex ○ Compile out lockdep for production builds ● Auth signature check optimizations ● BlueStore allocator improvements
  • 15. MSGR2 ● New on-wire protocol ○ Improved protocol feature negotiation ○ Multiplexing (maybe, eventually) ○ IANA-assigned port number for mons (ce4h = 3300) ● Encryption over the wire ● Probably no signature support? ○ fast mode (with no cryptographic integrity checks) or ○ secure mode (full encryption) ● Dual-stack support (IPv4 + IPv6) ● Kerberos authentication ○ Use kerberos credentials to map to Ceph roles, issue Ceph CLI commands
  • 16. MISC ● OSD memory stability ○ PG log length limited ○ OSD internally adjusts caches to stay within bounds ● QoS ○ min, max, and priority-based limits and reservations ● Improved introspection, utilization metrics (Nautilus) ● Future: ‘ceph top’ ○ Sampling-based real-time view of client workload
  • 17. RBD
  • 18. BLOCK ● Improved orchestration/management of async mirroring ○ Point-in-time consistent DR with failover/failback, mirrored snapshots, etc. ● Security namespaces for RBD ○ Simple RBAC-style CephX cap profiles ○ Basic provisioned space quotas ● Client-side caching - PR in progress ● Transparent live image migration ● Simplified pool- and image-level configuration overrides ● Future: ‘rbd top’ for real-time per-image workload
  • 19. RGW
  • 20. RGW / OBJECT ● Ongoing performance and scalability improvements ● Security (Nautilus) ○ Cloud-based (STS, keystone) and enterprise (OPA, kerberos/AD) ● pub/sub API (Nautilus) ○ Recently prototyped for OpenWhisk project; targets Nautilus ● Sync from public cloud (Nautilus) ○ RGW sync to S3 added in Mimic; this adds other direction ● Tiering ○ Push individual objects to external cloud (Nautilus) ○ Push entire buckets to external cloud
  • 21. MULTI- AND HYBRID CLOUD ● Next generation of applications will primarily consume object storage ○ Block is great for backing VMs and containers, ○ and file will serve legacy workloads and even scale to huge data sets, ○ but most cat pictures/videos/etc will land in objects ● Modern IT infrastructure spans multiple data centers, public/private clouds ● In the public cloud, it will be hard to beat native storage pricing (e.g., S3) ● RGW should expand to encompass “data services” ○ Data portability (especially paired with application portability) ○ Data placement (vs capacity, bandwidth, compliance/regulatory regimes) ○ Lifecycle management ○ Introspection (what am I storing and where?) ○ Policy and automation for all of the above
  • 23. MULTI-MDS ROBUSTNESS ● Multi-MDS stable and supported since Luminous ○ Mimic makes adding/removing MDS easier. ● Snapshots stable and supported since Mimic. ○ Requires v4.17+ kernel for kclient. ● Lots of work remaining to make dynamic load balancer more robust ○ Generate realistic workloads at scale (many MDS daemons, lots of RAM for each) ○ Balancer tuning
  • 24. 24 ● Directory (subtree) snapshots anywhere in the file system hierarchy. ● Note: for the kernel client, use the latest kernel. ● Mimic will change the default allow_new_snaps to true. Existing file systems must turn the flag on after upgrade. Credit to Zheng Yan (Red Hat). $ cat foo/file1 Hello $ mkdir /cephfs/foo/.snap/2017-10-06 $ echo “world!” >> foo/file1 $ cat foo/file1 Hello World! $ cat foo/.snap/2017-10-06/file1 Hello bar file2 / foo file1 SNAPSHOTS STABLE
  • 25. 25 ● Byte limit: $ setfattr -n ceph.quota.max_bytes -v 100000000 /some/dir ● File limit: $ setfattr -n ceph.quota.max_files -v 10000 /some/dir ● Caveat: limits are enforced eventually. ● Kernel changes not yet merged upstream Cooperatively fixed by Luis Henriques (SUSE) and Zheng Yan (Red Hat). dir2 file2 quota config: max_files=20,max_bytes=1048576 / dir1 file1 quota config: max_files=200,max_bytes=10485760 (KERNEL) QUOTA SUPPORT
  • 26. ● Bring volume/subvolume management into ceph-mgr ○ Change ceph-volume-client.py to simply wrap new functions ○ “Formalize” volume/subvolume concepts ○ Modify Rook, Kubernetes provisioners, Manila to all consume/share same interface ● Scale-out NFS ○ Cluster-managed ganesha gateways with active/active ○ Robust cluster-coherent NFSv4 recovery ○ Protect Ceph from untrusted clients VOLUME MANAGEMENT & NFS GATEWAYS
  • 27. cephfs-shell ● New alternative client that doesn’t requiring mounting CephFS ● Outreachy project Summer 2018 by Pavani Rajula. $ cephfs-shell CephFS:~/>>> mkdir foo CephFS:~/>>> cd foo CephFS:~/foo>>> put /etc/hosts hosts CephFS:~/foo>>> cat hosts 127.0.0.1 localhost.localdomain localhost ... ...
  • 28. PERFORMANCE ● Buffered create/unlink by clients ○ Use write/buffer capability on directory ○ Pass out allocated inodes to each session for creating files ○ Asynchronously flush create/unlink to MDS ● Sharding metadata within an MDS (“sub-ranks”) to scale across cores ○ Mostly lock-less is a design goal ○ Fast zero-copy metadata/cap export between sub-ranks
  • 30. ROOK ● Native, robust operator for kubernetes and openshift ● Intelligent deployment of Ceph daemons ○ e.g., add/remove/move mon daemons while maintaining quorum ○ e.g., intelligently schedule RGW/iSCSI/NFS gateways across nodes ● Integration with SDN functions ○ e.g., schedule Ganesha NFS gateways and attach them to tenant Neutron networks via Kuryr ● Upgrade orchestration ○ Update Rook operator pod (triggered via CLI or dashboard?) ○ Rook updates Ceph daemons in prescribed order, gated with health, safety, availability checks ○ Rook manages any release-specific steps (like forcing scrubs, declaring upgrade “done”, etc.) ● Existing and happy user community, CNCF member project
  • 31. K8S STRATEGY ● Align with Ceph with kubernetes community interest/adoption (Rook, CNCF) ● Enable early kubernetes/openshift adopters (IT users, not IT) ○ Provision Rook/Ceph clusters layered over existing infrastructure ● Displace of legacy storage on prem for kubernetes ○ Allow multiple kube clusters, Rook instances to share an external Ceph cluster ● Enable kubernetes as a service (e.g., provided by IT) ○ Enable multi-tenant workflows for Rook storage classes (e.g., Pools) ● Maximize consistency of experience on public cloud ● Expose underlying Ceph federation capabilities (especially object) ● Also: use kubernetes “under the hood” for standalone Ceph
  • 33. DASHBOARD ● Converged community investment on built-in web dashboard ○ Hybrid of openATTIC and John’s original dashboard proof of concept ○ Self-hosted by ceph-mgr with easy, tight integration with other cluster management and automation functions ● Currently mostly “logical” cluster functions ○ Management of Ceph services (pools, RBD images, file systems, configuration, etc.) ○ Subsuming ceph-metrics and openATTIC grafana metrics ● Orchestrator abstraction ○ Allow ceph-mgr and dashboard (or CLI, APIs) to drive ansible, Rook, DeepSea, etc ○ Provision or deprovision Ceph daemons, add/remove nodes, replace disks, etc. ○ Abstracts/hides choice of orchestration layer, enabling generalized automation, GUI, docs, UX ● Indirectly laying foundation for stable and versioned management API
  • 34. Questions? Neha Ojha - nojha@redhat.com Patrick Donnelly - pdonnell@redhat.com
  • 36. CEPH-CSI ● Replace upstream Kubernetes and Rook flexvol with ceph-csi ● Development driven by CERN, Cisco, with help from Huamin (Red Hat) ● Stretch goal is to replace Rook’s flexvol with ceph-csi in next 0.9 release
  • 37. MULTI-CLUSTER CEPHFS ● Geo-replication ○ Loosely-consistent and point-in-time consistent DR replication (Nautilus) ○ Active/active async replication, with associated consistency caveats? ● Sync and share (NextCloud) integration ○ Concurrent access via usual POSIX (kcephfs, NFS, etc) mounts and NextCloud to same files, with revisions