Tape storage remains a highly cost effective solution for backup and archival storage despite perceptions that it is outdated. Recent developments show continued improvements in tape technology with areal density demonstrations of 29.5 Gbit/in2 and LTO roadmaps extending to 12.8 TB capacities. The Long Term File System also allows tapes to be accessed like removable drives, opening new use cases. Together these advances ensure tape storage will continue to be relevant for the next decade.
Z4R: Intro to Storage and DFSMS for z/OSTony Pearson
This session covers basic storage concepts for z/OS operating system with examples for Flash, Disk and Tape devices and how to use DFSMS policy-based management. Presented at IBM TechU in Johannesburg, South Africa September 2019
Cloud Infrastructure and Services (CIS) - WebinarEMC
Between 2012 and 2020, the patch of the digital universe that CIOs and their IT staffs need to manage will become not just bigger but also more complex. The skills, experience, and resources to manage all these bits of data will become scarcer and more specialized, requiring a new, flexible, and scalable IT infrastructure that extends beyond the enterprise: cloud computing. By 2020, nearly 40% of the information in the digital universe will be "touched" by cloud computing providers - meaning that a byte. The Cloud Infrastructure and Services (CIS) session educates participants about cloud deployment and service models, cloud infrastructure, and the key considerations in migrating to cloud computing.
EMC IT's Journey to Cloud : BUSINESS PRODUCTION BACKUP & RECOVERY SYSTEMSEMC
EMC IT's Journey to Cloud
PHASE 2: BUSINESS PRODUCTION
BACKUP & RECOVERY SYSTEMS
Discover how EMC uses its next generation deduplication, backup, and archiving.
Z4R: Intro to Storage and DFSMS for z/OSTony Pearson
This session covers basic storage concepts for z/OS operating system with examples for Flash, Disk and Tape devices and how to use DFSMS policy-based management. Presented at IBM TechU in Johannesburg, South Africa September 2019
Cloud Infrastructure and Services (CIS) - WebinarEMC
Between 2012 and 2020, the patch of the digital universe that CIOs and their IT staffs need to manage will become not just bigger but also more complex. The skills, experience, and resources to manage all these bits of data will become scarcer and more specialized, requiring a new, flexible, and scalable IT infrastructure that extends beyond the enterprise: cloud computing. By 2020, nearly 40% of the information in the digital universe will be "touched" by cloud computing providers - meaning that a byte. The Cloud Infrastructure and Services (CIS) session educates participants about cloud deployment and service models, cloud infrastructure, and the key considerations in migrating to cloud computing.
EMC IT's Journey to Cloud : BUSINESS PRODUCTION BACKUP & RECOVERY SYSTEMSEMC
EMC IT's Journey to Cloud
PHASE 2: BUSINESS PRODUCTION
BACKUP & RECOVERY SYSTEMS
Discover how EMC uses its next generation deduplication, backup, and archiving.
OSBConf 2015 | Contemporary and cost efficient backups to to tape by josef we...NETWAYS
Recently IBM demonstrated a 220 TB Tape Cartridge. I will show the future of Tape Technology and the enhancement made in Tape Storage. Also I give an outlook in Hard-Disk and in Flash Technology. The roadmap in areal density and capacity growth in those different technology will force us to rethink our backup storage architecture in the future. I will discuss and compare those different storage technologies areal density, roadmap, bit error rate, cost and power consumption. I will calculate some example related to backup environment where not only huge data are stored but also many data processed daily.
S de2784 footprint-reduction-edge2015-v2Tony Pearson
Data footprint reduction is the umbrella term for technologies like Thin Provisioning, Space-efficient snapshots, Data deduplication, and Real-time Compression.
Google File System was innovatively created by Google engineers and it is ready for production in record time. The success of Google is to attributed the efficient search algorithm, and also to the underlying commodity hardware. As Google run number of application then Google’s goal became to build a vast storage network out of inexpensive commodity hardware. So Google create its own file system, named as Google File System that is GFS. Google File system is one of the largest file system in operation. Generally Google File System is a scalable distributed file system of large distributed data intensive apps. In the design phase of Google file system, in which the given stress includes component failures , files are huge and files are mutated by appending data. The entire file system is organized hierarchically in directories and identified by pathnames. The architecture comprises of multiple chunk servers, multiple clients and a single master. Files are divided into chunks, and that is the key design parameter. Google File System also uses leases and mutation order in their design to achieve atomicity and consistency. As of there fault tolerance, Google file system is highly available, replicas of chunk servers and master exists.
Cost analysis for acquisition of 250 terabytes of storage growing at 25% per year for five years. Products from EMC, NetAPP, NEC, Dot Hill were compared to a software defined storage solution based on SUSE Enterprise Storage software.
Google File System was innovatively created by Google engineers and it is ready for production in record time. The success of Google is to attributed the efficient search algorithm, and also to the underlying commodity hardware. As Google run number of application then Google’s goal became to build a vast storage network out of inexpensive commodity hardware. So Google create its own file system, named as Google File System that is GFS. Google File system is one of the largest file system in operation. Generally Google File System is a scalable distributed file system of large distributed data intensive apps. In the design phase of Google file system, in which the given stress includes component failures , files are huge and files are mutated by appending data. The entire file system is organized hierarchically in directories and identified by pathnames. The architecture comprises of multiple chunk servers, multiple clients and a single master. Files are divided into chunks, and that is the key design parameter. Google File System also uses leases and mutation order in their design to achieve atomicity and consistency. As of there fault tolerance, Google file system is highly available, replicas of chunk servers and master exists.
The tape Industry began in 1952 and the disk Industry in 1956. In 1952, the world’s first
successful commercial tape drive was delivered, the IBM 726 with 12,500 bytes of capacity per
reel. In 1956 the world’s first disk drive was delivered by IBM, the Ramac 350 with 5 megabytes
of capacity. Though no one knew it at the time, two key and lasting events linking disk and tape
for the foreseeable future had just occurred
Snapshots have been a key feature of primary storage infrastructures that IT professionals have relied on for years. But storage systems have traditionally been able to support only a limited number of active snapshots. And snapshots, being pointers and not actual data, are also susceptible to a primary storage system failure. As a result, most IT professionals use snapshots sparingly for protecting data. In this webinar Storage Switzerland and Nexenta show you how primary storage can be architected so that snapshots are able to meet almost all of the data protection requirements an organization has.
OSBConf 2015 | Contemporary and cost efficient backups to to tape by josef we...NETWAYS
Recently IBM demonstrated a 220 TB Tape Cartridge. I will show the future of Tape Technology and the enhancement made in Tape Storage. Also I give an outlook in Hard-Disk and in Flash Technology. The roadmap in areal density and capacity growth in those different technology will force us to rethink our backup storage architecture in the future. I will discuss and compare those different storage technologies areal density, roadmap, bit error rate, cost and power consumption. I will calculate some example related to backup environment where not only huge data are stored but also many data processed daily.
S de2784 footprint-reduction-edge2015-v2Tony Pearson
Data footprint reduction is the umbrella term for technologies like Thin Provisioning, Space-efficient snapshots, Data deduplication, and Real-time Compression.
Google File System was innovatively created by Google engineers and it is ready for production in record time. The success of Google is to attributed the efficient search algorithm, and also to the underlying commodity hardware. As Google run number of application then Google’s goal became to build a vast storage network out of inexpensive commodity hardware. So Google create its own file system, named as Google File System that is GFS. Google File system is one of the largest file system in operation. Generally Google File System is a scalable distributed file system of large distributed data intensive apps. In the design phase of Google file system, in which the given stress includes component failures , files are huge and files are mutated by appending data. The entire file system is organized hierarchically in directories and identified by pathnames. The architecture comprises of multiple chunk servers, multiple clients and a single master. Files are divided into chunks, and that is the key design parameter. Google File System also uses leases and mutation order in their design to achieve atomicity and consistency. As of there fault tolerance, Google file system is highly available, replicas of chunk servers and master exists.
Cost analysis for acquisition of 250 terabytes of storage growing at 25% per year for five years. Products from EMC, NetAPP, NEC, Dot Hill were compared to a software defined storage solution based on SUSE Enterprise Storage software.
Google File System was innovatively created by Google engineers and it is ready for production in record time. The success of Google is to attributed the efficient search algorithm, and also to the underlying commodity hardware. As Google run number of application then Google’s goal became to build a vast storage network out of inexpensive commodity hardware. So Google create its own file system, named as Google File System that is GFS. Google File system is one of the largest file system in operation. Generally Google File System is a scalable distributed file system of large distributed data intensive apps. In the design phase of Google file system, in which the given stress includes component failures , files are huge and files are mutated by appending data. The entire file system is organized hierarchically in directories and identified by pathnames. The architecture comprises of multiple chunk servers, multiple clients and a single master. Files are divided into chunks, and that is the key design parameter. Google File System also uses leases and mutation order in their design to achieve atomicity and consistency. As of there fault tolerance, Google file system is highly available, replicas of chunk servers and master exists.
The tape Industry began in 1952 and the disk Industry in 1956. In 1952, the world’s first
successful commercial tape drive was delivered, the IBM 726 with 12,500 bytes of capacity per
reel. In 1956 the world’s first disk drive was delivered by IBM, the Ramac 350 with 5 megabytes
of capacity. Though no one knew it at the time, two key and lasting events linking disk and tape
for the foreseeable future had just occurred
Snapshots have been a key feature of primary storage infrastructures that IT professionals have relied on for years. But storage systems have traditionally been able to support only a limited number of active snapshots. And snapshots, being pointers and not actual data, are also susceptible to a primary storage system failure. As a result, most IT professionals use snapshots sparingly for protecting data. In this webinar Storage Switzerland and Nexenta show you how primary storage can be architected so that snapshots are able to meet almost all of the data protection requirements an organization has.
3. Backup and Recovery are top concerns for IT Managers Reasons for Data Losses? Source: “For Small Businesses, Bad Backup Can Lead to Data Loss”, http://www.businessnewsdaily.com/data-loss-backup-physical-online-1077/ 42% of midmarket companies have experienced data loss. 32% lost files forever .
4. Hard Lessons Learned August 2011 April 2011 April 2011 : Amazon's huge EC2 cloud services crash permanently destroyed some data…. Amazon didn't say explicitly whether it was human error that touched off the event ….. August 2011: Lightning Strikes Out Amazon’s Cloud „ … . Separately, and independent from the power issue in the affected availability zone, we’ve discovered an error in the EBS software that cleans up unused snapshots. During a recent run of this EBS software in the EU-West Region, one or more blocks in a number of EBS snapshots were incorrectly deleted. The root cause was a software error that caused the snapshot references to a subset of blocks to be missed during the reference counting process ….. „ „… .. Not only are 18 days of backups for some volumes gone, the majority of the remaining snapshots have been corrupted by Amazon’s backup system . ….“
5. Hard Lessons Learned Feb 2011: Google restore gmail from tape … . I know what some of you are thinking: how could this happen if we have multiple copies of your data, in multiple data centers? Well, in some rare instances software bugs can affect several copies of the data. That’s what happened here. Some copies of mail were deleted, and we’ve been hard at work over the last 30 hours getting it back for the people affected by this issue. To protect your information from these unusual bugs, we also back it up to tape . Since the tapes are offline, they’re protected from such software bugs . So what caused this problem? We released a storage software update that introduced the unexpected bug, which caused 0.02% of Gmail users to temporarily lose access to their email. When we discovered the problem, we immediately stopped the deployment of the new software and reverted to the old version. Tape is, last line of defense !
16. Disclaimers The performance data contained herein was obtained in a controlled environment based on the use of specific data. Actual results that may be obtained in other operating environments may vary significantly. These values do not constitute a guarantee of performance. Product data is accurate as of initial publication and is subject to change without notice. No part of this presentation may be reproduced or transmitted in any form without written permission from IBM Corporation. References in this document to IBM products, programs, or services do not imply that IBM intends to make these available in all countries in which IBM operates. Any reference to an IBM program product in this document is not intended to state or imply that only IBM's program product may be used. Any functionally equivalent program may be used instead. The information provided in this document has not been submitted to any formal IBM test and is distributed "As Is" basis without any warranty either express or implied. The use of this information or the implementation of any of these techniques is a customer responsibility and depends on the customer's ability to evaluate and integrate them into their operating environment. While each item may have been reviewed by IBM for accuracy in a specific situation, there is no guarantee that the same or similar results will be obtained elsewhere. Customers attempting to adapt these techniques to their own environments do so at their own risk.
17.
Editor's Notes
This presentation is designed to help IBM Sales to discuss Backup/Restore topics and explore business. The will be able to easily design a Backup/Restore Solution in customer environement and align storage hardware. Based on the IBM Framework, part I released in 2009. New: Blueprints
3/12 Of course, we protect data with backup systems because we might need to recover is some day. While disasters or major business disruptions are rare, compared to ordinary file restore requests; they ’re frequent enough to warrant a significant investment. If a disaster is declared, what to we want to recover? We want to recover all the information needed to run the business – and we want to do it as quickly as possible. So I ask you, how do we improve the efficiency of recovery systems, if up to 70% of the data is old or duplicate? When we recognize how these challenges are related, and understand the current inefficiencies; the case for improved Data Protection and Retention is compelling. Inefficient Data Protection and Retention processes can be a Gold Mine of efficiency opportunities. For example, if a disaster is declared, can you restore the 30% or so of your data that is active, or do you have to restore everything? How much does your organization spend protecting data that is duplicate, expired, or infrequently accessed? Data Protection and Retention improvements can improve service levels and control costs by helping clients eliminate unnecessary process steps and workload. In many cases, efficiency improvements free more than enough captive resources to enable improvements in recovery time objectives and recovery point objectives.
2000 60GB 62 sec 2011 4000GB 40 sec Published Values for Bit Error Rate (BER) Tape drives (Midrange and Enterprise) Midrange (typical LTO Ultrium drive) 1 x 10E17 bits Enterprise (typical mainframe drive) 1 x 10E19 bits Disk (FC, SAS, SATA) Enterprise FC/SAS 1 x 10E16 bits Enterprise SATA 1 x 10E15 bits Desktop SATA 1 x 10E14 bits In addition, today’s tape technologies like LTO Ultrium drives, perform a read after write verification process. That is, a set of read elements read the data immediately after it is written to tape to verify that the data was written accurately. LTO Ultrium tape also includes advanced servo tracking mechanisms to help provide precision tracking of the read/write heads with the tape.