SlideShare a Scribd company logo
1
STORAGE VIRTUALIZATION: AN INSIDER’S GUIDE
Jon William Toigo
CEO Toigo Partners International
Chairman Data Management Institute
Copyright © 2013 by the Data Management Institute LLC. All Rights Reserved. Trademarks and tradenames for products
discussed in this document are the property of their respective owners. Opinions expressed here are those of the author.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 2
STORAGE VIRTUALIZATION:
AN INSIDER’S GUIDE
Part 3: Managing Storage Performance
When most business or IT managers are queried for their definition of
storage efficiency, the answer quickly turns to performance. Storage
needs to be nimble, delivering not only capacity sufficient to meet the
growing volume of data but also the speed to write data as rapidly as it is
presented and to retrieve it immediately upon requests from
applications and decision-makers. Hardware vendors have long
marketed speed enhancements as a key differentiator and main selling
point of their latest storage array, switch or interconnect, charging a
premium for the latest, fastest technology.
Storage virtualization has the potential to break speed barriers without
breaking the storage budget.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 3
STORAGE VIRTUALIZATION: AN INSIDER’S GUIDE
Managing Storage Performance
STORAGE PERFORMANCE: THE MEASURE OF IT EFFICIENCY?
Rightly or wrongly, the yardstick most commonly applied to evaluating how well IT departments
are delivering services to their users is speed. The speed at which databases post transactions
(“write operations”) and the speed with which requested information is painted onto the
screens of user client devices (“read operations”) shape the perceptions of users regarding the
efficiency of IT operations overall.
Input/output speed is, of course, a relative thing and hardly a precise or accurate measure of
overall IT performance. I/O performance is influenced by many variables. At root, it reflects
the balance that has been struck between processors, networking and storage infrastructures –
three technologies that are constantly changing in terms of their performance characteristics.
But speed is also influenced by a broad set of factors linked to workload, traffic patterns, and
operating system and application software stack design.
Truth be told, when one considers the successful completion of the trillions of transactions that
traverse IT infrastructure each working day, the volume of digital data that is processed and
handled by information technology, and the myriad transformations to which the data is
subjected as it moves between physical hardware and logical abstractions, one could readily
conclude that IT administrators are the modern equivalent of miracle workers.
Still, I/O speed is a “touch point” between IT and the business it serves. As such, it creates and
reinforces perceptions of overall IT performance with practical ramifications for strategies,
budgets, and careers. Speed is also leveraged by vendors to promote their wares, with boasts
about “the fastest performance of any server/switch/interconnect/array” frequently cited in
marketing literature and sales campaigns to differentiate competing products in the market.
Discussions of speed tend to narrow quickly to a focus on storage infrastructure. These days,
mechanical storage media manifests much slower I/O rate than networks and backplanes or
processors and memories. Yet, to borrow the line of a contemporary storage vendor, storage is
where information lives.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 4
Thus, considerable effort has been made
to optimize storage performance at the
device level and to architect storage
infrastructure to get the most speed for
the money. Managing storage
performance over time is therefore a key
management service expected of IT
administrators.
Storage performance management is a
multifaceted task. It actually begins with
the assessment of application
requirements and the selection of appropriate media to host application data. Key criteria in
media selection can include data output characteristics -- whether the data being generated is
transactional in nature, consisting of short block writes, or if it comprises of large files that may
extend beyond the size of a specific disk drive. Access frequency and numbers of concurrent
accesses expected of data, and the duration of high frequency and high concurrency accesses
(in seconds, minutes, days, or weeks) are also important criteria.
Cost is also important in media selection. Generally speaking, the faster the storage, the more
expensive it is, so striking a balance between speed and expense usually sets pragmatic
parameters on media selection. Differences between media performance and cost are
illustrated below.
For over 30 years, one approach used to accelerate the speed of storage infrastructure has
been to house many disks in an array, then to create logical drive comprising many individual
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 5
physical drives so they are presented as a single unit. In its simplest form, the intention is to
apply the greatest number of read/write heads to the task of capturing data to the disk.
Spreading out the workload in a parallel fashion provides an overall performance capability that
exceeds the write speeds of any one disk drive.
Additionally, data may be written only to a part of each drive – an area requiring the smallest
amount of read/write head movement – so that latency (the amount of time needed to orient
the head over the writing platter surface – can be minimized. This is called “short stroking”
and, while effective at delivering speed improvement, is very wasteful of disk media, expensive
in terms of the numbers of disk that need to be deployed, and heavy on power consumption.
In addition to simple disk arrays and short stroking, caching has also been used to optimize disk
performance. A cache controller, consisting of memory and logic, provides a means to capture
and acknowledge writes received at a disk array and to queue actual data placement operations
so that the performance of the application is not hampered by the slower speed of disk
operations themselves.
Read caching is slightly different. Data that is accessed frequently from disk may be read into
memory on the cache controller and served to applications from that location, where
performance is several orders of magnitude faster than the speed at which data is located and
read from physical disk. Considerable work has been done on read/write caching over the
years to help optimize disk performance, contributing to the costliness of arrays as much as to
their performance.
The most recent innovation in the area of disk performance acceleration is the industry’s
current initiatives around flash random access memory-based solid state disk (FLASH SSD).
Some vendors have chosen to create FLASH SSD in the same form factor as 3.5 or 2.5 inch disk
drives and use these devices as targets for hosting data directly. Writes and reads are
significantly faster using solid state memory than using rotational magnetic media, but there
are caveats. For one, data written to the cells of a FLASH SSD must be erased before the cells
can be written again. (With a hard disk, new data simply overwrites old data.) That additional
erase operation causes FLASH SSD to present non-linear performance.
Also, FLASH SSD currently manifests a memory wear problem: a maximum of 250,000 writes
can be made to a single cell location before it wears out. Vendors typically work around these
and other issues by delivering far more capacity on FLASH SSD than is advertised, then
swapping out written cell banks and worn out cells with capacity hidden on the drive.
Bottom line: while the advertised drive life of a FLASH SSD drive is said to be on par with
magnetic disk (about five years), they may burn out or operate at much reduced performance
much more quickly than a disk drive depending on the workload. In a laptop computer
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 6
operating personal productivity applications, a FLASH SSD might be a comfortable, if pricey,
replacement for a conventional mechanical disk drive. However, as storage for a high
performance transaction processing application generating millions of write operations per
second, FLASH SSD units may need to be replaced weekly.
An alternative use of FLASH SSD to accelerate performance is to use it in a disk optimization
role. Some vendors are using confusing terminology to describe this architecture, calling it
auto-tiering, so a bit of explanation may be required to understand it.
The confusion is around the word tiering. Traditional storage tiering is a protocol, a way of
doing things with data over time to optimize the available capacity of disk resources.
Traditional tiering conceives of different classes of storage ranging from fast (high speed, low
capacity, short stroking) disk arrays intended to capture data from demanding transactional
applications such as databases as quickly as possible to slower and slower arrays comprising
higher and higher capacity media with slower and slower performance characteristics.
The traditional tiering protocol sees data moving between these storage classes over time as
the frequency with which data is being accessed declines. The protocol addresses the problem
of capacity, performance and cost elegantly by placing data on storage that is most appropriate
to its use.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 7
Some vendors of storage equipment have sought to automate this traditional tiering process
and call the resulting capability of their array controller “auto-tiering.” This is a bit of an
oxymoron, since the vendor typically sells an array with both fast and slow media in the same
cabinet, but obviates the cost-savings that accrue by charging extra for the auto-tiering
software on the array controller. In a few cases, vendors have begun adding a stand of FLASH
SSD drives as the fastest tier of storage, “Tier 0,” thereby moving each class of mechanical disk
down a notch in the tiering protocol. In general, the effect is to dramatically increase the cost
of the overall storage platform.
To confuse matters more, some vendors
use the term “auto-tiering” in an
altogether different way. What they are
describing is a method for using a
memory device, typically a FLASH SSD
drive, as a disk optimizer.
In this approach, data is written by
applications to fast disk. However, if the
data is determined to be “hot” – accessed
frequently or by many concurrent read
requests – it is temporarily copied to the
FLASH SSD device so it can be serviced at
much higher speed. Once the data
“cools” and access frequency or
concurrency is reduced, read requests are re-pointed back to the hard disk, where normal I/O
requests are handled.
These disk optimizations, depending on the data being serviced, can be a cost-effective use of
FLASH SSD technology. It remains to be seen whether the technique is best applied at the
physical array level, or better exposed as a sharable service across all arrays in the
infrastructure. More discussion on this point is provided below.
Just as media selection and protocols such as caching and tiering can influence storage
infrastructure performance, so can the choice of the interconnect used to attach storage to
servers. Contemporary storage has no shortage of interconnect standards. Most derive from
the Small Computer Systems Interface (SCSI) standard developed over thirty years ago.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 8
In the late 1990s, a proliferation of interconnect technologies began to occur based on the
serialization of parallel SCSI to enable more devices to be connected via a single interconnecting
bus. The most popular interconnects today – Fibre Channel Protocol, Serial Attached SCSI, and
even iSCSI – are all manifestations of the same thing: SCSI serialized. The key differences have
been the transport mechanism employed for sending SCSI commands and data serially across a
wire. Fibre Channel uses copper or fiber optic cable, specialized switches and host bus adapters
installed on servers to extend the backplane of the server over distance – an architecture
known as a fabric, but commonly misrepresented as a “network” (as in Storage Area Network
or SAN). Serial Attached SCSI is pursuing the same general approach today.
Some vendors preferred to use already ubiquitous Ethernet networks and the Internet
Engineering Task Force protocol suite, TCP/IP, as the transport for serialized SCSI, producing a
standard called iSCSI (SCSI over IP). A network interface card and LAN switching equipment are
used to create this kind of interconnect between storage and servers, with iSCSI operating as an
application across a specialized and usually segregated LAN. Again, this is not a real SAN; it
simply uses a LAN as a transport for extending the SCSI bus backplane to numerous target
devices.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 9
The key takeaway from the above is that all storage is direct-attached and all storage
interconnects are essentially extensions of the bus backplane of a server, rather than
“networked storage.” This is important to understand only because of its impact on issues such
as storage management. A true network protocol has a management layer: a FC SAN does not.
Bottom line: there may be many paths through the storage interconnect fabric but there is
nothing in the underlying interconnecting protocols to help manage traffic across paths or to
balance the load. Evidence of this deficit can be found in just about every SAN deployed today,
where the typical utilization efficiency of a given port on a FC switch hovers at about 15% of
optimal. Even iSCSI Ethernet switch ports are underutilized. Performance management
requires that paths between servers and storage be optimized so that some ports and links do
not become congested, slowing the delivery of commands and data to and from servers and
storage, while other paths go largely unused.
In a nutshell, performance management entails the
management of media, protocols and paths to
achieve the greatest possible speed from the
storage infrastructure in response to data from
applications and end users. Achieving managed
performance in contemporary storage
infrastructure confronts several challenges
including infrastructure complexity, array isolation,
lack of classification of target storage arrays
according to performance characteristics, and the
absence of quality of service (QoS) and load
balancing functionality.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 10
PERFORMANCE MANAGEMENT WITH STORAGE VIRTUALIZATION
The good news is that storage virtualization can help with many of these issues. As noted in
previous papers in this series, storage virtualization is the establishment of a logical or virtual
controller above the physical storage infrastructure layer – including interconnecting switches,
cabling and arrays. Storage virtualization is a technology for simplifying the presentation of
storage resources so that those resources can be managed efficiently in terms of capacity,
performance and availability.
Speed through Caching
The first thing many administrators notice after virtualizing their storage is a significant
performance bump. All I/O is serviced faster than it was when storage was presented directly
from physical layer devices. The speed improvement differs from shop to shop, but a 200%
improvement is pretty common, and some companies using DataCore Software’s
SANsymphony-V have reported increases of up to 300% from the storage infrastructure below
the virtual controller.
The truth is that this
performance bump relates to
caching. In the DataCore
approach to storage
virtualization: the virtualization
engine takes the form of
software loaded on a commodity
server running the Microsoft
Windows™ Server 2008 server
operating system. Essentially,
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 11
SANsymphony-V aggregates all of the storage mounted to the server, enabling the creation of
one or more storage pools, then handles all writes to and reads from these logical storage pools
using server memory as a cache. As with caching in a physical array, reads and writes to/from
the SANsymphony-V memory cache are always faster than reads and writes to/from a physical
disk drive: hence, the performance improvement.
Of course, given the diversity of workload directed to the virtual controller and its increasingly
arbitrary characteristics – especially as companies embrace server and desktop virtualization –
a “static” cache (one that simply queues all write requests) is no guarantor of improved
performance over time. In a virtual server environment, for example, many applications may
share the same I/O paths and create different traffic levels and unpredictable peak loads that
make intelligent provisioning of storage challenging.
DataCore Software’s innovation, called “adaptive caching,” addresses this issue by intelligently
balancing workloads based on performance characteristics and other factors. Metaphorically,
adaptive caching can be viewed as shock absorbers that normalize differences in terrain so that
horsepower and torque are translated effectively into maximum wheel speed.
A potential side benefit of this raw speed improvement is the possibility it enables for
eliminating cost-ineffective short stroking arrays. With a speed improvement of up to 3X,
short-stroke arrays may become a footnote in storage history.
Speed through Tiering
In addition to improving the overall speed of storage infrastructure using adaptive caching,
storage virtualization also enables the tiering of storage without requiring the re-cabling or re-
addressing of arrays in the infrastructure. Virtualization enables the administrator to set up
multiple pools of storage based on criteria such as the performance characteristics of member
arrays. Fast storage may be pooled together to provide capture storage for transactional
workload, while slower storage arrays may be pooled into a resource that is more appropriate
for files. Arrays with high capacity but slow spindles may be pooled together and targeted as
archival repositories.
With pools of storage thus defined, migrating data between them in accordance with some sort
of hierarchical storage management or archive application is greatly simplified in a storage
virtualization setting. No storage is isolated and grouping together arrays with common speeds
and feeds characteristics, or arrays interconnected by different interconnecting technologies
(Fibre Channel, iSCSI, SAS, etc.) is easily accomplished. That’s good news for administrators
who want to deploy traditional storage tiering.
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 12
As mentioned above, current advances in tiering, so-called sub-LUN tiering, are gaining
mindshare as a mechanism for optimizing the performance of disk drives in an array. This is in
addition to caching; it is a purposeful movement of selected data to faster media based on the
amount of access being made to that data. When the data cools off, they are demoted to more
cost effective disks.
With storage virtualization, sub-LUN tiering can be delivered on a much grander scale. For
example, with its burgeoning auto-tiering technology, DataCore Software enables
administrators to leverage FLASH SSD drives in its virtualization engine, on the SANsymphony-V
server itself, to provide the equivalent of sub-LUN tiering as a service for all of the storage
arrays that the administrator has deployed to date. This can save companies the cost of
refreshing existing hardware with new SSD-ready gear, while delivering the performance
optimization promised from sub-LUN tiering techniques.
Speed through Path Management
The third way that storage virtualization can help improve performance is by helping to resolve
the performance of the “plumbing” of the storage infrastructure – the wires, cables, switches
and protocols that are used to connect the storage to the servers. At this infrastructure layer, it
is typical to find terrible inefficiencies, with some data paths and switch ports hardly used at all,
while others are overloaded and congested.
Very few companies have deployed “taps” and meters, or other performance monitoring
technologies to understand the I/O traffic traversing their Ethernet and Fibre Channel
infrastructure. Often, this reflects the fact that storage infrastructure has evolved over time,
using different interconnects and models for connecting equipment together. Today,
especially with the onset of server virtualization, the old plumbing models are creating major
slowdowns in storage I/O.
How can storage virtualization help to right the situation? In the case of DataCore
SANsymphony-V, all connections between storage arrays and the storage virtualization nodes
Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 13
are inventoried when the storage is added to a virtual pool. So the administrator knows how
equipment is connected together – what kinds of interconnects and protocols are being used
and what switch ports are handling traffic.
Next, DataCore uses a multipathing I/O driver in order to leverage all interconnecting paths
between the SANsymphony-V server and physical storage in an intelligent way. SANsymphony-
V also examines queue depths to determine which paths to the physical storage are congested,
and routes I/O in a balanced way across all available paths to the storage in a target pool. This
load balancing is performed behind the scenes, though reporting and monitoring is extensive.
So, administration of traffic across links does not require a deep knowledge of link speeds and
feeds or fabric design.
CONCLUSION OF PART 3
The above is not intended to suggest that performance management is a simple matter, even in
a virtualized storage environment. Many issues that might cause performance problems have
nothing whatsoever to do with storage infrastructure or interconnects. Applications, server
hypervisors and even operating systems can impact I/O performance.
The good news is that with the right
storage virtualization solution,
performance issues that have to do with
storage infrastructure can be resolved
readily. At a minimum, the technology
can eliminate most explanations for I/O
slowdown so that troubleshooting can
move upstream to other potential
explanations. If the impediments to IT
performance are linked to storage,
virtualizing storage with products such
as DataCore Software’s SANsymphony-V
can help administrators deliver off-the-
charts storage performance.

More Related Content

What's hot

Fluid Data Storage:Driving Flexibility in the Data Center
Fluid Data Storage:Driving Flexibility in the Data Center Fluid Data Storage:Driving Flexibility in the Data Center
Fluid Data Storage:Driving Flexibility in the Data Center
Kingfin Enterprises Limited
 
Virtual SAN- Deep Dive Into Converged Storage
Virtual SAN- Deep Dive Into Converged StorageVirtual SAN- Deep Dive Into Converged Storage
Virtual SAN- Deep Dive Into Converged Storage
DataCore Software
 
Building a data warehouse of call data records
Building a data warehouse of call data recordsBuilding a data warehouse of call data records
Building a data warehouse of call data recordsDavid Walker
 
Struggling with data management
Struggling with data managementStruggling with data management
Struggling with data managementDavid Walker
 
Rethinking Storage Infrastructures by Utilizing the Value of Flash
Rethinking Storage Infrastructures by Utilizing the Value of FlashRethinking Storage Infrastructures by Utilizing the Value of Flash
Rethinking Storage Infrastructures by Utilizing the Value of Flash
Jonathan Long
 
1cloudstar cloud store.v1.1
1cloudstar cloud store.v1.11cloudstar cloud store.v1.1
1cloudstar cloud store.v1.1
1CloudStar
 
Net App Syncsort Integrated Backup Solution Sheet
Net App Syncsort Integrated Backup Solution SheetNet App Syncsort Integrated Backup Solution Sheet
Net App Syncsort Integrated Backup Solution Sheet
Michael Hudak
 
Introduction to Hadoop
Introduction to HadoopIntroduction to Hadoop
Introduction to Hadoop
MindsMapped Consulting
 
Storage Characteristics Of Call Data Records In Column Store Databases
Storage Characteristics Of Call Data Records In Column Store DatabasesStorage Characteristics Of Call Data Records In Column Store Databases
Storage Characteristics Of Call Data Records In Column Store DatabasesDavid Walker
 
Accenture hana-in-memory-pov
Accenture hana-in-memory-povAccenture hana-in-memory-pov
Accenture hana-in-memory-povK Thomas
 
Dw hk-white paper
Dw hk-white paperDw hk-white paper
Dw hk-white paperjuly12jana
 
White paper whitewater-datastorageinthecloud
White paper whitewater-datastorageinthecloudWhite paper whitewater-datastorageinthecloud
White paper whitewater-datastorageinthecloudAccenture
 
Backup to Disk Infographic
Backup to Disk InfographicBackup to Disk Infographic
Backup to Disk Infographic
IT Brand Pulse
 
HP Autonomy - Three Ways to Preserve and Protect your Virtual Infrastructure
HP Autonomy - Three Ways to Preserve and Protect your Virtual InfrastructureHP Autonomy - Three Ways to Preserve and Protect your Virtual Infrastructure
HP Autonomy - Three Ways to Preserve and Protect your Virtual Infrastructure
VMUG IT
 
Presentation
PresentationPresentation
Presentation
Anoush Ghamsari
 
Terracotta Hadoop & In-Memory Webcast
Terracotta Hadoop & In-Memory WebcastTerracotta Hadoop & In-Memory Webcast
Terracotta Hadoop & In-Memory Webcast
Terracotta, a product line at Software AG
 
A brief introduction to data storage
A brief introduction to data storageA brief introduction to data storage
A brief introduction to data storage
Preludedx
 
Hitachi Cloud Solutions Profile
Hitachi Cloud Solutions Profile Hitachi Cloud Solutions Profile
Hitachi Cloud Solutions Profile Hitachi Vantara
 
In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...
In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...
In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...
Hitachi Vantara
 

What's hot (20)

Fluid Data Storage:Driving Flexibility in the Data Center
Fluid Data Storage:Driving Flexibility in the Data Center Fluid Data Storage:Driving Flexibility in the Data Center
Fluid Data Storage:Driving Flexibility in the Data Center
 
Virtual SAN- Deep Dive Into Converged Storage
Virtual SAN- Deep Dive Into Converged StorageVirtual SAN- Deep Dive Into Converged Storage
Virtual SAN- Deep Dive Into Converged Storage
 
Building a data warehouse of call data records
Building a data warehouse of call data recordsBuilding a data warehouse of call data records
Building a data warehouse of call data records
 
Struggling with data management
Struggling with data managementStruggling with data management
Struggling with data management
 
Rethinking Storage Infrastructures by Utilizing the Value of Flash
Rethinking Storage Infrastructures by Utilizing the Value of FlashRethinking Storage Infrastructures by Utilizing the Value of Flash
Rethinking Storage Infrastructures by Utilizing the Value of Flash
 
1cloudstar cloud store.v1.1
1cloudstar cloud store.v1.11cloudstar cloud store.v1.1
1cloudstar cloud store.v1.1
 
Net App Syncsort Integrated Backup Solution Sheet
Net App Syncsort Integrated Backup Solution SheetNet App Syncsort Integrated Backup Solution Sheet
Net App Syncsort Integrated Backup Solution Sheet
 
Introduction to Hadoop
Introduction to HadoopIntroduction to Hadoop
Introduction to Hadoop
 
Storage Characteristics Of Call Data Records In Column Store Databases
Storage Characteristics Of Call Data Records In Column Store DatabasesStorage Characteristics Of Call Data Records In Column Store Databases
Storage Characteristics Of Call Data Records In Column Store Databases
 
Accenture hana-in-memory-pov
Accenture hana-in-memory-povAccenture hana-in-memory-pov
Accenture hana-in-memory-pov
 
Dw hk-white paper
Dw hk-white paperDw hk-white paper
Dw hk-white paper
 
White paper whitewater-datastorageinthecloud
White paper whitewater-datastorageinthecloudWhite paper whitewater-datastorageinthecloud
White paper whitewater-datastorageinthecloud
 
Backup to Disk Infographic
Backup to Disk InfographicBackup to Disk Infographic
Backup to Disk Infographic
 
HP Autonomy - Three Ways to Preserve and Protect your Virtual Infrastructure
HP Autonomy - Three Ways to Preserve and Protect your Virtual InfrastructureHP Autonomy - Three Ways to Preserve and Protect your Virtual Infrastructure
HP Autonomy - Three Ways to Preserve and Protect your Virtual Infrastructure
 
Presentation
PresentationPresentation
Presentation
 
Terracotta Hadoop & In-Memory Webcast
Terracotta Hadoop & In-Memory WebcastTerracotta Hadoop & In-Memory Webcast
Terracotta Hadoop & In-Memory Webcast
 
A brief introduction to data storage
A brief introduction to data storageA brief introduction to data storage
A brief introduction to data storage
 
Connect July-Aug 2014
Connect July-Aug 2014Connect July-Aug 2014
Connect July-Aug 2014
 
Hitachi Cloud Solutions Profile
Hitachi Cloud Solutions Profile Hitachi Cloud Solutions Profile
Hitachi Cloud Solutions Profile
 
In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...
In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...
In the Age of Unstructured Data, Enterprise-Class Unified Storage Gives IT a ...
 

Similar to Insiders Guide- Managing Storage Performance

SSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-ProcessingSSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-Processing
Samsung Business USA
 
Demartek lenovo s3200_sql_server_evaluation_2016-01
Demartek lenovo s3200_sql_server_evaluation_2016-01Demartek lenovo s3200_sql_server_evaluation_2016-01
Demartek lenovo s3200_sql_server_evaluation_2016-01
Lenovo Data Center
 
Virtualizing Business Critical Applications
Virtualizing Business Critical ApplicationsVirtualizing Business Critical Applications
Virtualizing Business Critical Applications
DataCore Software
 
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage SwitzerlandWhy is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
INFINIDAT
 
Performance Tuning
Performance TuningPerformance Tuning
Performance Tuning
Jannet Peetz
 
Insiders Guide- Full Business Value of Storage Assets
Insiders Guide- Full Business Value of Storage AssetsInsiders Guide- Full Business Value of Storage Assets
Insiders Guide- Full Business Value of Storage AssetsDataCore Software
 
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
IBM India Smarter Computing
 
Efficient and scalable multitenant placement approach for in memory database ...
Efficient and scalable multitenant placement approach for in memory database ...Efficient and scalable multitenant placement approach for in memory database ...
Efficient and scalable multitenant placement approach for in memory database ...
CSITiaesprime
 
How do you know you really need ssd
How do you know you really need ssdHow do you know you really need ssd
How do you know you really need ssd
John McDonald
 
Our Hero Flash eBook
Our Hero Flash eBookOur Hero Flash eBook
Our Hero Flash eBook
thinkASG
 
Xd planning guide - storage best practices
Xd   planning guide - storage best practicesXd   planning guide - storage best practices
Xd planning guide - storage best practices
Nuno Alves
 
Workload Centric Scale-Out Storage for Next Generation Datacenter
Workload Centric Scale-Out Storage for Next Generation DatacenterWorkload Centric Scale-Out Storage for Next Generation Datacenter
Workload Centric Scale-Out Storage for Next Generation Datacenter
Cloudian
 
Streamlining Backup: Enhancing Data Protection with Backup Appliances
Streamlining Backup: Enhancing Data Protection with Backup AppliancesStreamlining Backup: Enhancing Data Protection with Backup Appliances
Streamlining Backup: Enhancing Data Protection with Backup Appliances
MaryJWilliams2
 
Demartek Lenovo Storage S3200 i a mixed workload environment_2016-01
Demartek Lenovo Storage S3200  i a mixed workload environment_2016-01Demartek Lenovo Storage S3200  i a mixed workload environment_2016-01
Demartek Lenovo Storage S3200 i a mixed workload environment_2016-01
Lenovo Data Center
 
Demartek lenovo s3200_mixed_workload_environment_2016-01
Demartek lenovo s3200_mixed_workload_environment_2016-01Demartek lenovo s3200_mixed_workload_environment_2016-01
Demartek lenovo s3200_mixed_workload_environment_2016-01
Lenovo Data Center
 
[IC Manage] Workspace Acceleration & Network Storage Reduction
[IC Manage] Workspace Acceleration & Network Storage Reduction[IC Manage] Workspace Acceleration & Network Storage Reduction
[IC Manage] Workspace Acceleration & Network Storage ReductionPerforce
 
Configuration and Deployment Guide For Memcached on Intel® Architecture
Configuration and Deployment Guide For Memcached on Intel® ArchitectureConfiguration and Deployment Guide For Memcached on Intel® Architecture
Configuration and Deployment Guide For Memcached on Intel® Architecture
Odinot Stanislas
 
Comparison of In-memory Data Platforms
Comparison of In-memory Data PlatformsComparison of In-memory Data Platforms
Comparison of In-memory Data Platforms
Amir Mahdi Akbari
 
The economics of backup 5 ways disk backup can help your business
The economics of backup 5 ways disk backup can help your businessThe economics of backup 5 ways disk backup can help your business
The economics of backup 5 ways disk backup can help your businessServium
 
The Fundamental Characteristics of Storage concepts for DBAs
The Fundamental Characteristics of Storage concepts for DBAsThe Fundamental Characteristics of Storage concepts for DBAs
The Fundamental Characteristics of Storage concepts for DBAs
Alireza Kamrani
 

Similar to Insiders Guide- Managing Storage Performance (20)

SSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-ProcessingSSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-Processing
 
Demartek lenovo s3200_sql_server_evaluation_2016-01
Demartek lenovo s3200_sql_server_evaluation_2016-01Demartek lenovo s3200_sql_server_evaluation_2016-01
Demartek lenovo s3200_sql_server_evaluation_2016-01
 
Virtualizing Business Critical Applications
Virtualizing Business Critical ApplicationsVirtualizing Business Critical Applications
Virtualizing Business Critical Applications
 
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage SwitzerlandWhy is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
 
Performance Tuning
Performance TuningPerformance Tuning
Performance Tuning
 
Insiders Guide- Full Business Value of Storage Assets
Insiders Guide- Full Business Value of Storage AssetsInsiders Guide- Full Business Value of Storage Assets
Insiders Guide- Full Business Value of Storage Assets
 
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
 
Efficient and scalable multitenant placement approach for in memory database ...
Efficient and scalable multitenant placement approach for in memory database ...Efficient and scalable multitenant placement approach for in memory database ...
Efficient and scalable multitenant placement approach for in memory database ...
 
How do you know you really need ssd
How do you know you really need ssdHow do you know you really need ssd
How do you know you really need ssd
 
Our Hero Flash eBook
Our Hero Flash eBookOur Hero Flash eBook
Our Hero Flash eBook
 
Xd planning guide - storage best practices
Xd   planning guide - storage best practicesXd   planning guide - storage best practices
Xd planning guide - storage best practices
 
Workload Centric Scale-Out Storage for Next Generation Datacenter
Workload Centric Scale-Out Storage for Next Generation DatacenterWorkload Centric Scale-Out Storage for Next Generation Datacenter
Workload Centric Scale-Out Storage for Next Generation Datacenter
 
Streamlining Backup: Enhancing Data Protection with Backup Appliances
Streamlining Backup: Enhancing Data Protection with Backup AppliancesStreamlining Backup: Enhancing Data Protection with Backup Appliances
Streamlining Backup: Enhancing Data Protection with Backup Appliances
 
Demartek Lenovo Storage S3200 i a mixed workload environment_2016-01
Demartek Lenovo Storage S3200  i a mixed workload environment_2016-01Demartek Lenovo Storage S3200  i a mixed workload environment_2016-01
Demartek Lenovo Storage S3200 i a mixed workload environment_2016-01
 
Demartek lenovo s3200_mixed_workload_environment_2016-01
Demartek lenovo s3200_mixed_workload_environment_2016-01Demartek lenovo s3200_mixed_workload_environment_2016-01
Demartek lenovo s3200_mixed_workload_environment_2016-01
 
[IC Manage] Workspace Acceleration & Network Storage Reduction
[IC Manage] Workspace Acceleration & Network Storage Reduction[IC Manage] Workspace Acceleration & Network Storage Reduction
[IC Manage] Workspace Acceleration & Network Storage Reduction
 
Configuration and Deployment Guide For Memcached on Intel® Architecture
Configuration and Deployment Guide For Memcached on Intel® ArchitectureConfiguration and Deployment Guide For Memcached on Intel® Architecture
Configuration and Deployment Guide For Memcached on Intel® Architecture
 
Comparison of In-memory Data Platforms
Comparison of In-memory Data PlatformsComparison of In-memory Data Platforms
Comparison of In-memory Data Platforms
 
The economics of backup 5 ways disk backup can help your business
The economics of backup 5 ways disk backup can help your businessThe economics of backup 5 ways disk backup can help your business
The economics of backup 5 ways disk backup can help your business
 
The Fundamental Characteristics of Storage concepts for DBAs
The Fundamental Characteristics of Storage concepts for DBAsThe Fundamental Characteristics of Storage concepts for DBAs
The Fundamental Characteristics of Storage concepts for DBAs
 

More from DataCore Software

Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...
Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...
Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...
DataCore Software
 
NVMe and Flash – Make Your Storage Great Again!
NVMe and Flash – Make Your Storage Great Again!NVMe and Flash – Make Your Storage Great Again!
NVMe and Flash – Make Your Storage Great Again!
DataCore Software
 
Zero Downtime, Zero Touch Stretch Clusters from Software-Defined Storage
Zero Downtime, Zero Touch Stretch Clusters from Software-Defined StorageZero Downtime, Zero Touch Stretch Clusters from Software-Defined Storage
Zero Downtime, Zero Touch Stretch Clusters from Software-Defined Storage
DataCore Software
 
From Disaster to Recovery: Preparing Your IT for the Unexpected
From Disaster to Recovery: Preparing Your IT for the UnexpectedFrom Disaster to Recovery: Preparing Your IT for the Unexpected
From Disaster to Recovery: Preparing Your IT for the Unexpected
DataCore Software
 
How to Integrate Hyperconverged Systems with Existing SANs
How to Integrate Hyperconverged Systems with Existing SANsHow to Integrate Hyperconverged Systems with Existing SANs
How to Integrate Hyperconverged Systems with Existing SANs
DataCore Software
 
How to Avoid Disasters via Software-Defined Storage Replication & Site Recovery
How to Avoid Disasters via Software-Defined Storage Replication & Site RecoveryHow to Avoid Disasters via Software-Defined Storage Replication & Site Recovery
How to Avoid Disasters via Software-Defined Storage Replication & Site Recovery
DataCore Software
 
Cloud Infrastructure for Your Data Center
Cloud Infrastructure for Your Data CenterCloud Infrastructure for Your Data Center
Cloud Infrastructure for Your Data Center
DataCore Software
 
Building a Highly Available Data Infrastructure
Building a Highly Available Data InfrastructureBuilding a Highly Available Data Infrastructure
Building a Highly Available Data Infrastructure
DataCore Software
 
TUI Case Study
TUI Case StudyTUI Case Study
TUI Case Study
DataCore Software
 
Thorntons Case Study
Thorntons Case StudyThorntons Case Study
Thorntons Case Study
DataCore Software
 
Top 3 Challenges Impacting Your Data and How to Solve Them
Top 3 Challenges Impacting Your Data and How to Solve ThemTop 3 Challenges Impacting Your Data and How to Solve Them
Top 3 Challenges Impacting Your Data and How to Solve Them
DataCore Software
 
Business Continuity for Mission Critical Applications
Business Continuity for Mission Critical ApplicationsBusiness Continuity for Mission Critical Applications
Business Continuity for Mission Critical Applications
DataCore Software
 
Dynamic Hyper-Converged Future Proof Your Data Center
Dynamic Hyper-Converged Future Proof Your Data CenterDynamic Hyper-Converged Future Proof Your Data Center
Dynamic Hyper-Converged Future Proof Your Data Center
DataCore Software
 
Community Health Network Delivers Unprecedented Availability for Critical Hea...
Community Health Network Delivers Unprecedented Availability for Critical Hea...Community Health Network Delivers Unprecedented Availability for Critical Hea...
Community Health Network Delivers Unprecedented Availability for Critical Hea...
DataCore Software
 
Case Study: Mission Community Hospital
Case Study: Mission Community HospitalCase Study: Mission Community Hospital
Case Study: Mission Community Hospital
DataCore Software
 
Emergency Communication of Southern Oregon
Emergency Communication of Southern OregonEmergency Communication of Southern Oregon
Emergency Communication of Southern Oregon
DataCore Software
 
DataCore At VMworld 2016
DataCore At VMworld 2016DataCore At VMworld 2016
DataCore At VMworld 2016
DataCore Software
 
Integrating Hyper-converged Systems with Existing SANs
Integrating Hyper-converged Systems with Existing SANs Integrating Hyper-converged Systems with Existing SANs
Integrating Hyper-converged Systems with Existing SANs
DataCore Software
 
Fighting the Hidden Costs of Data Storage
Fighting the Hidden Costs of Data StorageFighting the Hidden Costs of Data Storage
Fighting the Hidden Costs of Data Storage
DataCore Software
 
Can $0.08 Change your View of Storage?
Can $0.08 Change your View of Storage?Can $0.08 Change your View of Storage?
Can $0.08 Change your View of Storage?
DataCore Software
 

More from DataCore Software (20)

Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...
Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...
Software-Defined Storage Accelerates Storage Cost Reduction and Service-Level...
 
NVMe and Flash – Make Your Storage Great Again!
NVMe and Flash – Make Your Storage Great Again!NVMe and Flash – Make Your Storage Great Again!
NVMe and Flash – Make Your Storage Great Again!
 
Zero Downtime, Zero Touch Stretch Clusters from Software-Defined Storage
Zero Downtime, Zero Touch Stretch Clusters from Software-Defined StorageZero Downtime, Zero Touch Stretch Clusters from Software-Defined Storage
Zero Downtime, Zero Touch Stretch Clusters from Software-Defined Storage
 
From Disaster to Recovery: Preparing Your IT for the Unexpected
From Disaster to Recovery: Preparing Your IT for the UnexpectedFrom Disaster to Recovery: Preparing Your IT for the Unexpected
From Disaster to Recovery: Preparing Your IT for the Unexpected
 
How to Integrate Hyperconverged Systems with Existing SANs
How to Integrate Hyperconverged Systems with Existing SANsHow to Integrate Hyperconverged Systems with Existing SANs
How to Integrate Hyperconverged Systems with Existing SANs
 
How to Avoid Disasters via Software-Defined Storage Replication & Site Recovery
How to Avoid Disasters via Software-Defined Storage Replication & Site RecoveryHow to Avoid Disasters via Software-Defined Storage Replication & Site Recovery
How to Avoid Disasters via Software-Defined Storage Replication & Site Recovery
 
Cloud Infrastructure for Your Data Center
Cloud Infrastructure for Your Data CenterCloud Infrastructure for Your Data Center
Cloud Infrastructure for Your Data Center
 
Building a Highly Available Data Infrastructure
Building a Highly Available Data InfrastructureBuilding a Highly Available Data Infrastructure
Building a Highly Available Data Infrastructure
 
TUI Case Study
TUI Case StudyTUI Case Study
TUI Case Study
 
Thorntons Case Study
Thorntons Case StudyThorntons Case Study
Thorntons Case Study
 
Top 3 Challenges Impacting Your Data and How to Solve Them
Top 3 Challenges Impacting Your Data and How to Solve ThemTop 3 Challenges Impacting Your Data and How to Solve Them
Top 3 Challenges Impacting Your Data and How to Solve Them
 
Business Continuity for Mission Critical Applications
Business Continuity for Mission Critical ApplicationsBusiness Continuity for Mission Critical Applications
Business Continuity for Mission Critical Applications
 
Dynamic Hyper-Converged Future Proof Your Data Center
Dynamic Hyper-Converged Future Proof Your Data CenterDynamic Hyper-Converged Future Proof Your Data Center
Dynamic Hyper-Converged Future Proof Your Data Center
 
Community Health Network Delivers Unprecedented Availability for Critical Hea...
Community Health Network Delivers Unprecedented Availability for Critical Hea...Community Health Network Delivers Unprecedented Availability for Critical Hea...
Community Health Network Delivers Unprecedented Availability for Critical Hea...
 
Case Study: Mission Community Hospital
Case Study: Mission Community HospitalCase Study: Mission Community Hospital
Case Study: Mission Community Hospital
 
Emergency Communication of Southern Oregon
Emergency Communication of Southern OregonEmergency Communication of Southern Oregon
Emergency Communication of Southern Oregon
 
DataCore At VMworld 2016
DataCore At VMworld 2016DataCore At VMworld 2016
DataCore At VMworld 2016
 
Integrating Hyper-converged Systems with Existing SANs
Integrating Hyper-converged Systems with Existing SANs Integrating Hyper-converged Systems with Existing SANs
Integrating Hyper-converged Systems with Existing SANs
 
Fighting the Hidden Costs of Data Storage
Fighting the Hidden Costs of Data StorageFighting the Hidden Costs of Data Storage
Fighting the Hidden Costs of Data Storage
 
Can $0.08 Change your View of Storage?
Can $0.08 Change your View of Storage?Can $0.08 Change your View of Storage?
Can $0.08 Change your View of Storage?
 

Insiders Guide- Managing Storage Performance

  • 1. 1 STORAGE VIRTUALIZATION: AN INSIDER’S GUIDE Jon William Toigo CEO Toigo Partners International Chairman Data Management Institute Copyright © 2013 by the Data Management Institute LLC. All Rights Reserved. Trademarks and tradenames for products discussed in this document are the property of their respective owners. Opinions expressed here are those of the author.
  • 2. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 2 STORAGE VIRTUALIZATION: AN INSIDER’S GUIDE Part 3: Managing Storage Performance When most business or IT managers are queried for their definition of storage efficiency, the answer quickly turns to performance. Storage needs to be nimble, delivering not only capacity sufficient to meet the growing volume of data but also the speed to write data as rapidly as it is presented and to retrieve it immediately upon requests from applications and decision-makers. Hardware vendors have long marketed speed enhancements as a key differentiator and main selling point of their latest storage array, switch or interconnect, charging a premium for the latest, fastest technology. Storage virtualization has the potential to break speed barriers without breaking the storage budget.
  • 3. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 3 STORAGE VIRTUALIZATION: AN INSIDER’S GUIDE Managing Storage Performance STORAGE PERFORMANCE: THE MEASURE OF IT EFFICIENCY? Rightly or wrongly, the yardstick most commonly applied to evaluating how well IT departments are delivering services to their users is speed. The speed at which databases post transactions (“write operations”) and the speed with which requested information is painted onto the screens of user client devices (“read operations”) shape the perceptions of users regarding the efficiency of IT operations overall. Input/output speed is, of course, a relative thing and hardly a precise or accurate measure of overall IT performance. I/O performance is influenced by many variables. At root, it reflects the balance that has been struck between processors, networking and storage infrastructures – three technologies that are constantly changing in terms of their performance characteristics. But speed is also influenced by a broad set of factors linked to workload, traffic patterns, and operating system and application software stack design. Truth be told, when one considers the successful completion of the trillions of transactions that traverse IT infrastructure each working day, the volume of digital data that is processed and handled by information technology, and the myriad transformations to which the data is subjected as it moves between physical hardware and logical abstractions, one could readily conclude that IT administrators are the modern equivalent of miracle workers. Still, I/O speed is a “touch point” between IT and the business it serves. As such, it creates and reinforces perceptions of overall IT performance with practical ramifications for strategies, budgets, and careers. Speed is also leveraged by vendors to promote their wares, with boasts about “the fastest performance of any server/switch/interconnect/array” frequently cited in marketing literature and sales campaigns to differentiate competing products in the market. Discussions of speed tend to narrow quickly to a focus on storage infrastructure. These days, mechanical storage media manifests much slower I/O rate than networks and backplanes or processors and memories. Yet, to borrow the line of a contemporary storage vendor, storage is where information lives.
  • 4. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 4 Thus, considerable effort has been made to optimize storage performance at the device level and to architect storage infrastructure to get the most speed for the money. Managing storage performance over time is therefore a key management service expected of IT administrators. Storage performance management is a multifaceted task. It actually begins with the assessment of application requirements and the selection of appropriate media to host application data. Key criteria in media selection can include data output characteristics -- whether the data being generated is transactional in nature, consisting of short block writes, or if it comprises of large files that may extend beyond the size of a specific disk drive. Access frequency and numbers of concurrent accesses expected of data, and the duration of high frequency and high concurrency accesses (in seconds, minutes, days, or weeks) are also important criteria. Cost is also important in media selection. Generally speaking, the faster the storage, the more expensive it is, so striking a balance between speed and expense usually sets pragmatic parameters on media selection. Differences between media performance and cost are illustrated below. For over 30 years, one approach used to accelerate the speed of storage infrastructure has been to house many disks in an array, then to create logical drive comprising many individual
  • 5. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 5 physical drives so they are presented as a single unit. In its simplest form, the intention is to apply the greatest number of read/write heads to the task of capturing data to the disk. Spreading out the workload in a parallel fashion provides an overall performance capability that exceeds the write speeds of any one disk drive. Additionally, data may be written only to a part of each drive – an area requiring the smallest amount of read/write head movement – so that latency (the amount of time needed to orient the head over the writing platter surface – can be minimized. This is called “short stroking” and, while effective at delivering speed improvement, is very wasteful of disk media, expensive in terms of the numbers of disk that need to be deployed, and heavy on power consumption. In addition to simple disk arrays and short stroking, caching has also been used to optimize disk performance. A cache controller, consisting of memory and logic, provides a means to capture and acknowledge writes received at a disk array and to queue actual data placement operations so that the performance of the application is not hampered by the slower speed of disk operations themselves. Read caching is slightly different. Data that is accessed frequently from disk may be read into memory on the cache controller and served to applications from that location, where performance is several orders of magnitude faster than the speed at which data is located and read from physical disk. Considerable work has been done on read/write caching over the years to help optimize disk performance, contributing to the costliness of arrays as much as to their performance. The most recent innovation in the area of disk performance acceleration is the industry’s current initiatives around flash random access memory-based solid state disk (FLASH SSD). Some vendors have chosen to create FLASH SSD in the same form factor as 3.5 or 2.5 inch disk drives and use these devices as targets for hosting data directly. Writes and reads are significantly faster using solid state memory than using rotational magnetic media, but there are caveats. For one, data written to the cells of a FLASH SSD must be erased before the cells can be written again. (With a hard disk, new data simply overwrites old data.) That additional erase operation causes FLASH SSD to present non-linear performance. Also, FLASH SSD currently manifests a memory wear problem: a maximum of 250,000 writes can be made to a single cell location before it wears out. Vendors typically work around these and other issues by delivering far more capacity on FLASH SSD than is advertised, then swapping out written cell banks and worn out cells with capacity hidden on the drive. Bottom line: while the advertised drive life of a FLASH SSD drive is said to be on par with magnetic disk (about five years), they may burn out or operate at much reduced performance much more quickly than a disk drive depending on the workload. In a laptop computer
  • 6. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 6 operating personal productivity applications, a FLASH SSD might be a comfortable, if pricey, replacement for a conventional mechanical disk drive. However, as storage for a high performance transaction processing application generating millions of write operations per second, FLASH SSD units may need to be replaced weekly. An alternative use of FLASH SSD to accelerate performance is to use it in a disk optimization role. Some vendors are using confusing terminology to describe this architecture, calling it auto-tiering, so a bit of explanation may be required to understand it. The confusion is around the word tiering. Traditional storage tiering is a protocol, a way of doing things with data over time to optimize the available capacity of disk resources. Traditional tiering conceives of different classes of storage ranging from fast (high speed, low capacity, short stroking) disk arrays intended to capture data from demanding transactional applications such as databases as quickly as possible to slower and slower arrays comprising higher and higher capacity media with slower and slower performance characteristics. The traditional tiering protocol sees data moving between these storage classes over time as the frequency with which data is being accessed declines. The protocol addresses the problem of capacity, performance and cost elegantly by placing data on storage that is most appropriate to its use.
  • 7. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 7 Some vendors of storage equipment have sought to automate this traditional tiering process and call the resulting capability of their array controller “auto-tiering.” This is a bit of an oxymoron, since the vendor typically sells an array with both fast and slow media in the same cabinet, but obviates the cost-savings that accrue by charging extra for the auto-tiering software on the array controller. In a few cases, vendors have begun adding a stand of FLASH SSD drives as the fastest tier of storage, “Tier 0,” thereby moving each class of mechanical disk down a notch in the tiering protocol. In general, the effect is to dramatically increase the cost of the overall storage platform. To confuse matters more, some vendors use the term “auto-tiering” in an altogether different way. What they are describing is a method for using a memory device, typically a FLASH SSD drive, as a disk optimizer. In this approach, data is written by applications to fast disk. However, if the data is determined to be “hot” – accessed frequently or by many concurrent read requests – it is temporarily copied to the FLASH SSD device so it can be serviced at much higher speed. Once the data “cools” and access frequency or concurrency is reduced, read requests are re-pointed back to the hard disk, where normal I/O requests are handled. These disk optimizations, depending on the data being serviced, can be a cost-effective use of FLASH SSD technology. It remains to be seen whether the technique is best applied at the physical array level, or better exposed as a sharable service across all arrays in the infrastructure. More discussion on this point is provided below. Just as media selection and protocols such as caching and tiering can influence storage infrastructure performance, so can the choice of the interconnect used to attach storage to servers. Contemporary storage has no shortage of interconnect standards. Most derive from the Small Computer Systems Interface (SCSI) standard developed over thirty years ago.
  • 8. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 8 In the late 1990s, a proliferation of interconnect technologies began to occur based on the serialization of parallel SCSI to enable more devices to be connected via a single interconnecting bus. The most popular interconnects today – Fibre Channel Protocol, Serial Attached SCSI, and even iSCSI – are all manifestations of the same thing: SCSI serialized. The key differences have been the transport mechanism employed for sending SCSI commands and data serially across a wire. Fibre Channel uses copper or fiber optic cable, specialized switches and host bus adapters installed on servers to extend the backplane of the server over distance – an architecture known as a fabric, but commonly misrepresented as a “network” (as in Storage Area Network or SAN). Serial Attached SCSI is pursuing the same general approach today. Some vendors preferred to use already ubiquitous Ethernet networks and the Internet Engineering Task Force protocol suite, TCP/IP, as the transport for serialized SCSI, producing a standard called iSCSI (SCSI over IP). A network interface card and LAN switching equipment are used to create this kind of interconnect between storage and servers, with iSCSI operating as an application across a specialized and usually segregated LAN. Again, this is not a real SAN; it simply uses a LAN as a transport for extending the SCSI bus backplane to numerous target devices.
  • 9. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 9 The key takeaway from the above is that all storage is direct-attached and all storage interconnects are essentially extensions of the bus backplane of a server, rather than “networked storage.” This is important to understand only because of its impact on issues such as storage management. A true network protocol has a management layer: a FC SAN does not. Bottom line: there may be many paths through the storage interconnect fabric but there is nothing in the underlying interconnecting protocols to help manage traffic across paths or to balance the load. Evidence of this deficit can be found in just about every SAN deployed today, where the typical utilization efficiency of a given port on a FC switch hovers at about 15% of optimal. Even iSCSI Ethernet switch ports are underutilized. Performance management requires that paths between servers and storage be optimized so that some ports and links do not become congested, slowing the delivery of commands and data to and from servers and storage, while other paths go largely unused. In a nutshell, performance management entails the management of media, protocols and paths to achieve the greatest possible speed from the storage infrastructure in response to data from applications and end users. Achieving managed performance in contemporary storage infrastructure confronts several challenges including infrastructure complexity, array isolation, lack of classification of target storage arrays according to performance characteristics, and the absence of quality of service (QoS) and load balancing functionality.
  • 10. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 10 PERFORMANCE MANAGEMENT WITH STORAGE VIRTUALIZATION The good news is that storage virtualization can help with many of these issues. As noted in previous papers in this series, storage virtualization is the establishment of a logical or virtual controller above the physical storage infrastructure layer – including interconnecting switches, cabling and arrays. Storage virtualization is a technology for simplifying the presentation of storage resources so that those resources can be managed efficiently in terms of capacity, performance and availability. Speed through Caching The first thing many administrators notice after virtualizing their storage is a significant performance bump. All I/O is serviced faster than it was when storage was presented directly from physical layer devices. The speed improvement differs from shop to shop, but a 200% improvement is pretty common, and some companies using DataCore Software’s SANsymphony-V have reported increases of up to 300% from the storage infrastructure below the virtual controller. The truth is that this performance bump relates to caching. In the DataCore approach to storage virtualization: the virtualization engine takes the form of software loaded on a commodity server running the Microsoft Windows™ Server 2008 server operating system. Essentially,
  • 11. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 11 SANsymphony-V aggregates all of the storage mounted to the server, enabling the creation of one or more storage pools, then handles all writes to and reads from these logical storage pools using server memory as a cache. As with caching in a physical array, reads and writes to/from the SANsymphony-V memory cache are always faster than reads and writes to/from a physical disk drive: hence, the performance improvement. Of course, given the diversity of workload directed to the virtual controller and its increasingly arbitrary characteristics – especially as companies embrace server and desktop virtualization – a “static” cache (one that simply queues all write requests) is no guarantor of improved performance over time. In a virtual server environment, for example, many applications may share the same I/O paths and create different traffic levels and unpredictable peak loads that make intelligent provisioning of storage challenging. DataCore Software’s innovation, called “adaptive caching,” addresses this issue by intelligently balancing workloads based on performance characteristics and other factors. Metaphorically, adaptive caching can be viewed as shock absorbers that normalize differences in terrain so that horsepower and torque are translated effectively into maximum wheel speed. A potential side benefit of this raw speed improvement is the possibility it enables for eliminating cost-ineffective short stroking arrays. With a speed improvement of up to 3X, short-stroke arrays may become a footnote in storage history. Speed through Tiering In addition to improving the overall speed of storage infrastructure using adaptive caching, storage virtualization also enables the tiering of storage without requiring the re-cabling or re- addressing of arrays in the infrastructure. Virtualization enables the administrator to set up multiple pools of storage based on criteria such as the performance characteristics of member arrays. Fast storage may be pooled together to provide capture storage for transactional workload, while slower storage arrays may be pooled into a resource that is more appropriate for files. Arrays with high capacity but slow spindles may be pooled together and targeted as archival repositories. With pools of storage thus defined, migrating data between them in accordance with some sort of hierarchical storage management or archive application is greatly simplified in a storage virtualization setting. No storage is isolated and grouping together arrays with common speeds and feeds characteristics, or arrays interconnected by different interconnecting technologies (Fibre Channel, iSCSI, SAS, etc.) is easily accomplished. That’s good news for administrators who want to deploy traditional storage tiering.
  • 12. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 12 As mentioned above, current advances in tiering, so-called sub-LUN tiering, are gaining mindshare as a mechanism for optimizing the performance of disk drives in an array. This is in addition to caching; it is a purposeful movement of selected data to faster media based on the amount of access being made to that data. When the data cools off, they are demoted to more cost effective disks. With storage virtualization, sub-LUN tiering can be delivered on a much grander scale. For example, with its burgeoning auto-tiering technology, DataCore Software enables administrators to leverage FLASH SSD drives in its virtualization engine, on the SANsymphony-V server itself, to provide the equivalent of sub-LUN tiering as a service for all of the storage arrays that the administrator has deployed to date. This can save companies the cost of refreshing existing hardware with new SSD-ready gear, while delivering the performance optimization promised from sub-LUN tiering techniques. Speed through Path Management The third way that storage virtualization can help improve performance is by helping to resolve the performance of the “plumbing” of the storage infrastructure – the wires, cables, switches and protocols that are used to connect the storage to the servers. At this infrastructure layer, it is typical to find terrible inefficiencies, with some data paths and switch ports hardly used at all, while others are overloaded and congested. Very few companies have deployed “taps” and meters, or other performance monitoring technologies to understand the I/O traffic traversing their Ethernet and Fibre Channel infrastructure. Often, this reflects the fact that storage infrastructure has evolved over time, using different interconnects and models for connecting equipment together. Today, especially with the onset of server virtualization, the old plumbing models are creating major slowdowns in storage I/O. How can storage virtualization help to right the situation? In the case of DataCore SANsymphony-V, all connections between storage arrays and the storage virtualization nodes
  • 13. Copyright © 2013 by The Data Management Institute LLC. All Rights Reserved. 13 are inventoried when the storage is added to a virtual pool. So the administrator knows how equipment is connected together – what kinds of interconnects and protocols are being used and what switch ports are handling traffic. Next, DataCore uses a multipathing I/O driver in order to leverage all interconnecting paths between the SANsymphony-V server and physical storage in an intelligent way. SANsymphony- V also examines queue depths to determine which paths to the physical storage are congested, and routes I/O in a balanced way across all available paths to the storage in a target pool. This load balancing is performed behind the scenes, though reporting and monitoring is extensive. So, administration of traffic across links does not require a deep knowledge of link speeds and feeds or fabric design. CONCLUSION OF PART 3 The above is not intended to suggest that performance management is a simple matter, even in a virtualized storage environment. Many issues that might cause performance problems have nothing whatsoever to do with storage infrastructure or interconnects. Applications, server hypervisors and even operating systems can impact I/O performance. The good news is that with the right storage virtualization solution, performance issues that have to do with storage infrastructure can be resolved readily. At a minimum, the technology can eliminate most explanations for I/O slowdown so that troubleshooting can move upstream to other potential explanations. If the impediments to IT performance are linked to storage, virtualizing storage with products such as DataCore Software’s SANsymphony-V can help administrators deliver off-the- charts storage performance.