The document discusses Ceph storage performance on all-flash storage systems. It describes how SanDisk optimized Ceph for all-flash environments by tuning the OSD to handle the high performance of flash drives. The optimizations allowed over 200,000 IOPS per OSD using 12 CPU cores. Testing on SanDisk's InfiniFlash storage system showed it achieving over 1.5 million random read IOPS and 200,000 random write IOPS at 64KB block size. Latency was also very low, with 99% of operations under 5ms for reads. The document outlines reference configurations for the InfiniFlash system optimized for small, medium and large workloads.
Intel - optimizing ceph performance by leveraging intel® optane™ and 3 d nand...inwin stack
Kenny Chang (張任伯) (Storage Solution Architect, Intel)
With the trend that Solid State Drive (SSD) becomes more affordable, more and more cloud providers are trying to provide high performance, highly reliable storage for their customers with SSDs. Ceph is becoming one of most open source scale-out storage solutions in worldwide market. More and more customers have strong demands that using SSD in Ceph to build high performance storage solutions for their Openstack clouds.
The disrupted Intel® Optane SSDs based on 3D Xpoint technology fills the performance gap between DRAM and NAND based SSD while the Intel® 3D NAND TLC is reducing cost gap between SSD and traditional spindle hard drive and makes it possible for all flash storage. In this session, we will
1) Discuss OpenStack storage Ceph reference design on the first Intel Optane (3D Xpoint) and P4500 TLC NAND based all-flash Ceph cluster, it delivers multi-million IOPS with extremely low latency as well as increase storage density with competitive dollar-per-gigabyte costs
2) Share Ceph bluestore tunings and optimizations, latency analysis, TCO model, IOPS/TB, IOPS/$ based on the reference architecture to demonstrate this high performance, cost effective solution.
Intel - optimizing ceph performance by leveraging intel® optane™ and 3 d nand...inwin stack
Kenny Chang (張任伯) (Storage Solution Architect, Intel)
With the trend that Solid State Drive (SSD) becomes more affordable, more and more cloud providers are trying to provide high performance, highly reliable storage for their customers with SSDs. Ceph is becoming one of most open source scale-out storage solutions in worldwide market. More and more customers have strong demands that using SSD in Ceph to build high performance storage solutions for their Openstack clouds.
The disrupted Intel® Optane SSDs based on 3D Xpoint technology fills the performance gap between DRAM and NAND based SSD while the Intel® 3D NAND TLC is reducing cost gap between SSD and traditional spindle hard drive and makes it possible for all flash storage. In this session, we will
1) Discuss OpenStack storage Ceph reference design on the first Intel Optane (3D Xpoint) and P4500 TLC NAND based all-flash Ceph cluster, it delivers multi-million IOPS with extremely low latency as well as increase storage density with competitive dollar-per-gigabyte costs
2) Share Ceph bluestore tunings and optimizations, latency analysis, TCO model, IOPS/TB, IOPS/$ based on the reference architecture to demonstrate this high performance, cost effective solution.
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...Odinot Stanislas
(FR)
Voici un excellent document qui explique étape après étape comment installer, monitorer et surtout correctement benchmarker ses SSD PCIe/NVMe (pas si simple que ça). Autre élément clé : comment analyser la charge I/O de véritables applications? Combien d'IOPS, en read, en write, quelle bande passante et surtout quel impact sur la durée de vie des SSD? Bref à mettre en toute les mains, et un merci à mon collègue Andrey Kudryavtsev.
(EN)
An excellent content which describe step by step how to install, monitor and benchmark PCIe/NVMe SSD (many trick not so simple). Another key learning: how to measure real I/O activities on a real workload? How many R/W IOPS, block size, throughtput, and finally what's the impact on SSD endurance and (real)life? A must read, and a huge thanks to my colleague Andrey Kudryavtsev.
Auteurs/Authors:
Andrey Kudryavtsev, SSD Solution Architect, Intel Corporation
Zhdan Bybin, Application Engineer, Intel Corporation
The flash market started out monolithically. Flash was a single media type (high performance, high endurance SLC flash). Flash systems also had a single purpose of accelerating the response time of high-end databases. But now there are several flash options. Users can choose between high performance flash or highly dense, medium performance flash systems. At the same time, high capacity hard disk drives are making a case to be the archival storage medium of choice. How does an IT professional choose?
Yesterday's thinking may still believe NVMe (NVM Express) is in transition to a production ready solution. In this session, we will discuss how the evolution of NVMe is ready for production, the history and evolution of NVMe and the Linux stack to address where NVMe has progressed today to become the low latency, highly reliable database key value store mechanism that will drive the future of cloud expansion. Examples of protocol efficiencies and types of storage engines that are optimizing for NVMe will be discussed. Please join us for an exciting session where in-memory computing and persistence have evolved.
Deploying All-Flash Cloud Infrastructure without Breaking the BankWestern Digital
Watch the full webinar here: http://bit.ly/1PsFrUo
As the amount of data companies generate and consume increases year over year, cloud computing is becoming more relied upon to help businesses improve efficiencies and reduce costs. However, the challenge is maintaining a balance between performance, capacity and cost for the varying types of workloads. Join Venkat Kolli as he discusses how to achieve flexibility and performance at economies of scale with the only all-flash storage solution specifically tuned for OpenStack environments.
Recorded Presentation: https://www.brighttalk.com/webcast/12587/175373
IBM POWER8 processor is the fastest available on the market, redefining Open Source performance. With this amazing processor, IBM and members of the OpenPower Foundation design innovative and cost-effective systems, delivering the infrastructure of choice for the most demanding workloads, in terms of throughput, scalability and reliability.
In this talk in english, Thibaud Besson will browse the key characteristics of Power Systems, why they are the most relevant for today's challenges, both from a technical and economical standpoint. Finally, we will review the possibilities you have to get your hands on one of these outstanding plateforms for your Open Source applications.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
2. Forward-Looking Statements
During our meeting today we will make forward-looking statements.
Any statement that refers to expectations, projections or other characterizations of future events or
circumstances is a forward-looking statement, including those relating to market growth, industry
trends, future products, product performance and product capabilities. This presentation also
contains forward-looking statements attributed to third parties, which reflect their projections as of the
date of issuance.
Actual results may differ materially from those expressed in these forward-looking statements due
to a number of risks and uncertainties, including the factors detailed under the caption “Risk Factors”
and elsewhere in the documents we file from time to time with the SEC, including our annual and
quarterly reports.
We undertake no obligation to update these forward-looking statements, which speak only as
of the date hereof or as of the date of issuance by a third party, as the case may be.
5. Old Model
Monolithic, large upfront
investments, and fork-lift
upgrades
Proprietary storage OS
Costly: $$$$$
New SD-AFS Model
Disaggregate storage, compute, and software
for better scaling and costs
Best-in-class solution components
Open source software - no vendor lock-in
Cost-efficient: $
Software-defined All-Flash Storage
7. Storage performance is hugely affected by seemingly small details
All HW is not equal – Switches, NICs, HBAs, SSDs all matter
• Drivers abstraction doesn’t hide dynamic behavior
All SW is not equal – Distro, Patches, Drivers, Configuration matter
Typically large delta between “default” and “tuned” system perf
What’s a user to do?
Software Defined Storage – what’s NOT new
9. The InfiniFlash™ System
9
64-512 TB
JBOD of flash in
3U
Up to 2M IOPS,
<1ms latency,
Up to 15 GB/s
throughput
Energy Efficient
~400W power draw
Connect
up to 8 servers
Simple yet
Scalable
12. InfiniFlash™
8TB Flash-Card Innovation
• Enterprise-grade power-fail safe
• Latching integrated & monitored
• Directly samples air temp
• New flash form factor not SSD-based
Non-disruptive Scale-Up & Scale-
Out
• Capacity on demand
• Serve high growth Big Data
• 3U chassis starting at 64TB up to
512TB
• 8 to 64 8TB flash cards (SAS)
• Compute on demand
• Serve dynamic apps without
IOPS/TB bottlenecks
• Add up to 8 servers
17. InfiniFlash IF550 (HW + SW)
Ultra-dense High Capacity Flash storage
Highly scalable performance
Cinder, Glance and Swift storage
Enterprise-Class storage features
Ceph Optimized for SanDisk flash
18. InfiniFlash SW + HW Advantage
Software Storage System
Software tuned for
Hardware
• Extensive Ceph mods
Hardware Configured
for Software
• Density, power, architecture
Ceph has over 50 tuning parameters
that results in 5x – 6x performance
improvement
19. IF550 - Enhancing Ceph for Enterprise Consumption
• SanDisk Ceph Distro ensures packaging with stable, production-ready code with consistent quality
• All Ceph Performance improvements developed by SanDisk are contributed back to community
19
SanDisk / RedHat
or Community
Distribution
Out-of-the Box
configurations tuned for
performance with Flash
Sizing & planning tool
InfiniFlash drive
management
integrated into Ceph
management
Ceph installer built for InfiniFlash
High performance iSCSI storage
log collection tool
Enterprise hardened SW + HW QA
21. Starting working with Ceph over 2.5 years ago (Dumpling)
Aligned on vision of scale-out Enterprise storage
• Multi-protocol design
• Cluster / Cloud oriented
• Open Source Commitment
SanDisk’s engagement with Ceph
• Flash levels of performance
• Enterprise quality
• Support tools for our product offerings
Ceph and SanDisk
22. Optimising Ceph for the all-flash Future
Ceph optimized for HDD
• Tuning AND algorithm changes needed for Flash optimization
Quickly determined that the OSD was the major bottleneck
• OSD maxed out at about 1000 IOPS on fastest CPUs (using ~4.5
cores)
Examined and rejected multiple OSDs per SSD
• Failure Domain / Crush rules would be a nightmare
23. SanDisk: OSD Read path Optimisation
Context switches matter at flash rates
• Too much “put it in a queue for another thread” and lock contention
Socket handling matters too!
• Too many “get 1 byte” calls to the kernel for sockets
• Disable Nagle’s algorithm to shorten operation latency
Lots of other simple things
• Eliminate repeated look-ups, Redundant string copies, etc
Contributed improvements to Emperor, Firefly and Giant releases
Now obtain about >200K IOPS / OSD using around 12 CPU cores/OSD (Jewel)
24. SanDisk: OSD Write path Optimization
Write path strategy was classic HDD
• Journal writes for minimum foreground latency
• Process journal in batches in the background
Inefficient for flash
Modified buffering/writing strategy for Flash (Jewel!)
• 2.5x write throughput and avg latency is ½ of Hammer
26. Test Configuration
2 x InfiniFlash systems 256TB
each
8 x OSD nodes
• 2x E5-2697, 14C 2.6Hz v3 ,
8x 16GB DDR4 ECC
2133Mhz
1x Mellanox X3 Dual 40GbE
• Ubuntu 14.04.02 LTS 64bit
8 – 10 Client nodes
Ceph Version: sndk-ifos-1.3.0.317,
based on Ceph 10.2.1 (Jewel)
27. IFOS Block IOPS Performance
Highlights
• 4k numbers are cpu bound, increase in server
cpu will improve IOPS by ~11%
• 64k and higher block size bandwidth is close to
raw box bandwidth.
• 256k Random Read numbers can be increased
further based on number of clients, able to
achieve > 90% drive saturation with 14 clients.
1521231
347628
82456
6
22 21
0
5
10
15
20
25
0
200000
400000
600000
800000
1000000
1200000
1400000
1600000
4 64 256
BandwidthinGB/s
IOPs
Block Size
Random Read
Sum of
IOPs
201465
55648
16289
0.8
3.5
4.1
0
1
2
3
4
5
0
50000
100000
150000
200000
250000
4 64 256
BandwidthinGB/s
IOPs
Block Size
Random Write
Sum
of…
Write Performance is on 2 x copy configuration.
28.
29. IFOS Block workload Latency Performance
Environment
• Librbd IO read latency measured on Golden Config with 2 way replication at host level having 8 osd nodes, IO duration 20min.
• fio Read IO profile :64k block with 2 num Jobs & io-depth 16 with 10 clients(each client with one rbd)
• fio Write IO profile :64k block with 2 num Jobs & io-depth 16 with 10 clients(each client with one rbd)
64K Random Read 64K Random Write
Average latency : 6.3ms
Latency Range
(µs) Percentile
500 2.21
750 0.22
1000 7.43
2000 62.72
4000 26.11
10000 1.27
20000 0.03
50000 0.01
Latency Range
(µs) Percentile
1000 0.33
2000 43.16
4000 28.11
10000 21
20000 3.31
50000 2.17
100000 1.35
250000 0.4
500000 0.11
750000 0.04
1000000 0.01
2000000 0.01
Average latency : 1.7ms
2.21 0.22
7.43
62.72
26.12
1.28 0.03 0.01
0
20
40
60
80
Percentile
Latency range in us
Random Read Latency
Histogram
99 percent of
the IOPs is
within 5ms
latency
0.33
43.16
28.11
21
3.312.171.350.40.110.040.010.01
0
10
20
30
40
50
Percentile
Latency Range in us
Random Write Latency
Histogram
99% IOPS : 178367.31 99% IOPS : 227936.61
30.
31. IFOS Object Performance
Erasure Coding provides equivalent of 3x
replica storage with only 1.2x storage
Object performance is on par with Block
performance
Higher node clusters = higher EC ratio = more
storage savings
• Replication Configuration
• OSD level replication with 2 copies
• Erasure Coding Configuration
• Node level Erasure coding with Couchy-Good 6+2
• Couchy Good is better suited to InfiniFlash vs.
Reed Solomon
0
5
10
15
Repl (2x) -
Read
EC (6+2)
Read
Repl (2x) -
Write
EC (6+2)
Write
GBps
Protection Scheme
4M Objects Throughout - Erausre
Coding vs. Replication
32. IF550 Reference Configurations
Workload Small Medium Large
Small Block I/O
2x IF150
o 128TB to 256 TB
Flash per enclosure
using performance
card (4TB)
1 OSD Server per 4-8 cards
o Dual E5-2680
o 64GB RAM
2x IF150
o 128TB to 256 TB
Flash per enclosure
using performance
card (4TB)
1 OSD Server per 4-8 cards
o Dual E5-2687
o 128GB RAM
2+ IF150
o 128TB to 256 TB
Flash per enclosure
using performance
card (4TB)
1 OSD Server per 4-8 cards
o Dual E5-2697+
o 128GB RAM
Throughput
2x IF150
o 128TB to 512 TB
Flash per enclosure
1 OSD Server per 16 cards
o Dual E5-2660
o 64GB RAM
2x IF150
o 128TB to 512 TB
Flash per enclosure
1 OSD Server per 16 cards
o Dual E5-2680
o 128GB RAM
2+ IF150
o 128TB to 512 TB
Flash per enclosure
1 OSD Server per 16 cards
o Dual E5-2680
o 128GB RAM
Mixed
2x IF150
o 128TB to 512 TB
Flash per enclosure
1 OSD Server per 8-16
cards
o Dual E5-2680
o 64GB RAM
2x IF150
o 128TB to 512 TB
Flash per enclosure
1 OSD Server per 8-16
cards
o Dual E5-2690+
o 128GB RAM
2+ IF150
o 128TB to 256 TB
Flash per enclosure
(optional
performance CARD)
1 OSD Server per 8-16
cards
o Dual E5-2695+
o 128GB RAM
33. InfiniFlash TCO Advantage
Reduce the replica count from 3 to 2
Less compute, less HW and SW
• TCO analysis based on a US customer’s OPEX & Cost
data for a 5PB deployment
33
$-
$5
$10
$15
$20
$25
$30
InfiniFlash External AFA DAS SSD node DAS 10k HDD
node
Millions
5 Year TCO
CAPEX OPEX
0 20 40
InfiniFlash
External AFA
DAS SSD node
DAS 10k HDD
node
Data Center
Racks
$- $2,000 $4,000
InfiniFlash
External AFA
DAS SSD node
DAS 10k HDD
node
Thousands
Total Energy Cost
37. SanDisk: Potential Future Improvements
RDMA intra-cluster communication
• Significant reduction in CPU / IOP
BlueStore
• Significant reduction in write amplification -> even higher write
performance
Memory allocation
• tcmalloc/jemalloc/AsyncMessenger tuning shows up to 3x IOPS vs.
default *
Erasure Coding for Blocks (native)
* https://drive.google.com/file/d/0B2gTBZrkrnpZY3U3TUU3RkJVeVk/view
38. Time to fix the write path algorithm
Review of FileStore
• What’s wrong with FileStore
• XFS + levelDB
• Missing transactional semantics for metadata and data
• Missing virtual-copy and merge semantics
• BTRFS implementation of these isn’t general enough
Snapshot/rollback overhead too expensive for frequent use
Transaction semantics aren’t crash proof
Bad Write amp.
Bad jitter due to unpredictable file system
Bad CPU utilization, syncfs is VERY expensive
39. BlueStore
One, Two or Three raw block devices
• Data, Metadata/WAL and KV Journaling
• When combined no fixed partitioning is
needed
Use a single transactional KV store for all
metadata
• Semantics are well matched to
ObjectStore transactions
Use raw block device for data storage
• Support Flash, PMR and SMR HDD
ObjectStore
BlueStore
KeyValueDB
Data
MetaData
BlueFS
Operation Decoder
Journal
Client/Gateway Operations Peer-to-Peer Cluster
Management
Network Interface
40. BlueStore vs FileStore
1 800GB P3700 card (4 OSDs per), 64GB ram, 2 x Intel(R) Xeon(R) CPU E5-2650 v3 @ 2.30GHz, 1 x Intel 40GbE link
client fio processes and mon were on the same nodes as the OSDs.
41. Emerging Storage Solutions (EMS) SanDisk Confidential 41
KV Store Options
RocksDB is a Facebook extension of levelDB
– Log Structured Merge (LSM) based
– Ideal when metadata is on HDD
– Merge is effectively host-based GC when run on flash
ZetaScale™ from SanDisk® now open sourced
– B-tree based
– Ideal when metadata is on Flash
– Uses device-based GC for max performance
42. BlueStore ZetaScale v RocksDB Performance
Test Setup:
1 OSD, 8TB SAS SSD, 10GB ram, Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz , fio, 32 thds, 64 iodepth, 6TB dataset, 30
min
0.436
1.005
3.970
0.95
2.83
9.29
0.000
1.000
2.000
3.000
4.000
5.000
6.000
7.000
8.000
9.000
10.000
0/100 70/30 100/0
IOPsinKs
Read/Write Ratio
Random Read/Write 4K IOPs per OSD
BlueStore(RocksDB)
BlueStore(ZetaScale)
43. The InfiniFlash™ System ...
Power
70% Less
Speed
40x Faster
than SAN
Density
10X Higher
Reliability
20x Better
AFR
Cost
Up to 80%
Lower TCO
Video continues to drive the need for storage, and Point-Of-View cameras like GoPro are producing compelling high resolution videos on our performance cards. People using smartphones to make high resolution videos choose our performance mobile cards also, driving the need for higher capacities.
There is a growing customer base for us around the world, with one billion additional people joining the Global Middle Class between 2013 and 2020. These people will use smart mobile devices as their first choice to spend discretionary income on, and will expand their storage using removable cards and USB drives.
We are not standing still, but creating new product categories to allow people to expand and share their most cherished memories.
___________________________________________________________
Eg. If you’re using it for small blocks, you need more CPUs. However large objects can use less servers. Your choice on how you want to deploy it.
All these are listed as various ineffeciencies. Originally about 10K IOPS before doing all the optimisations
Ran a 1PB test on Hammer with 256TB scaling. Almost linearly scaling.
You can’t get these numbers easily elsewhere.
Typically latency is around 10ms for R and 20-40ms for W even on flash!
EC is a customer configurable option. A lot more writes with 3 copy.
SanDisk is working on block EC….right now just object
The point here is flash is about the same as HDD
FileStore – existing backend storage for CEPH, many deficits. BlueStore is the new architecture. This is a preview… Tech preview for the rest of this year. By the L release it will be production.
We will switch to KV store and get rid of the journal. A journal invokes too many writes. Almost 4.
KV Store will be RockDB but SD will introduce a flash optimised KV Store basedon ZetaScale.
Today's flash solutions and arrays can address most of these problems – they are low power, high performance, somewhat scalable (though not to 10s of Pbs) and highly reliable but there is one thing that is holds it back – something missing – Favorable Economics – it’s simply way too expensive for @ scale workloads making flash out of reach. So we went to work as a team – our first investment was the best investment we ever made – a very clean sheet of paper !!! We knew that today’s HDD based storage solutions and today’s all flash arrays would not do the trick. We had to create something brand new that looks like nothing the world has seen.
Substantiation
Low Power 20 enclosures down to 2 – 100 watts per enclosure 24 Drives per enclosure (9w HDD, 7W SSD) = 93% power reductions or 1/16 the power. From his TCO Calculator.
HDD
480 Drives -9w
20 Enclosures – 100w
4320 + 2000 = 6320w
SSD
46 SSDs – 7w 176TB
2 Enclosures – 100w
536 Watts
Extreme Performance
30x faster NoSQL transactions MongoDB Solution Brief
Scalable
4,500 virtual desktops in one rack Fusion ioVDI and VMware Horizon View: Reference Architecture for VDI
Reliable
Accelerate Oracle Backup Using SanDisk Solid State Drives (SSDs)
Accelerate Oracle Backup Using SanDisk Solid State Drives (SSDs)
Breakthrough Economics
~3x faster Hadoop jobs with half the servers Increasing Hadoop Performance with SanDisk SSDs