SAN ARCHITECTURE AND HARDWARE DEVICES : Overview, creating a Network for storage; SAN Hardware devices, The fibre channel switch, Host Bus adaptors; Putting the storage in SAN; Fabric operation from a Hardware perspective
Important Terminology for the Users of Web-based ServicesHTS Hosting
The rapid growth of the World Wide Web and the increased use of web-based services make it essential for the users of such services to be aware of the most important and frequently used terms with regard to web-based services.
Learn about Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive Enterprise Workloads. This IBM Redpaper discusses server performance imbalance that can be found in typical application environments and how to address this issue with the 16 Gb Fibre Channel technology to provide required levels of performance and availability for the storage-intensive applications. For more information on Pure Systems, visit http://ibm.co/18vDnp6.
Visit the official Scribd Channel of IBM India Smarter Computing at http://bit.ly/VwO86R to get access to more documents.
A brief study on Storage Area Network (SAN), SAN architecture & its importance. It focuses on the techniques and the technologies that have evolved around SAN & its Security.
Network Attached Storage (NAS) Initiative Gary Wilhelm
Consolidation of departmental file services is one of the initiatives defined by Carolina Counts. The project has been researched and funded and is now in the implementation phase. We will cover the specifics of the solution and the details of the business and support models.
IBM Spectrum Scale Best Practices for Genomics Medicine WorkloadsUlf Troppens
Genomics medicine requires physicians, data scientists and researchers to analyze huge amounts of genomics data quickly. The IBM Spectrum Scale Best Practices for Genomics Medicine Workload provides composable infrastructure that enables IT architects to customize deployments for varying functional and performance needs. The described scale-out architecture is capable to store, access and manage genomics data from a few 100 TB to tens of PB. The solution integrates compute resources and an easy-to-use Web User Interface to submit high-throughput batch jobs to analyze genomics data sets. While the best practices are optimized for genomics medicine workloads, most of the settings are generic and applicable to other workloads and industries.
Class lecture by Prof. Raj Jain on Storage Virtualization. The talk covers Disk Arrays, Data Access Methods, SCSI (Small Computer System Interface), Advanced Technology Attachment (ATA), ESCON and FICON, Fibre Chanel, Fibre Channel Devices, Fibre Channel Protocol Layers, Fibre Channel Flow Control, Fibre Channel Classes of Service, What is Storage Virtualization?, Benefits of Storage Virtualization, Virtualizing Storage, RAID Levels, Nested RAIDs, Synchronous vs. Asynchronous Replication, Virtual Storage Area Network (VSAN), Physical Storage Network, Virtual Storage Network, SAN vs. NAS, iSCSI (Internet Small Computer System Interface), iFCP (Internet Fiber Channel Protocol), FCIP (Fibre Channel over IP), FCoE (Fibre Channel over Ethernet), Virtual File Systems. Video recording available in YouTube.
SoftLayer Storage Services Overview (for Interop Las Vegas 2015)Michael Fork
Introduction to SoftLayer's Storage Services. Topics covered include Block and File offerings Endurance, Performance, Mass Storage Servers (QuantaStor), and Backup (EVault, R1Soft), Object Storage (OpenStack Swift), CDN, Data Transfer Service, and Aspera.
SAN ARCHITECTURE AND HARDWARE DEVICES : Overview, creating a Network for storage; SAN Hardware devices, The fibre channel switch, Host Bus adaptors; Putting the storage in SAN; Fabric operation from a Hardware perspective
Important Terminology for the Users of Web-based ServicesHTS Hosting
The rapid growth of the World Wide Web and the increased use of web-based services make it essential for the users of such services to be aware of the most important and frequently used terms with regard to web-based services.
Learn about Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive Enterprise Workloads. This IBM Redpaper discusses server performance imbalance that can be found in typical application environments and how to address this issue with the 16 Gb Fibre Channel technology to provide required levels of performance and availability for the storage-intensive applications. For more information on Pure Systems, visit http://ibm.co/18vDnp6.
Visit the official Scribd Channel of IBM India Smarter Computing at http://bit.ly/VwO86R to get access to more documents.
A brief study on Storage Area Network (SAN), SAN architecture & its importance. It focuses on the techniques and the technologies that have evolved around SAN & its Security.
Network Attached Storage (NAS) Initiative Gary Wilhelm
Consolidation of departmental file services is one of the initiatives defined by Carolina Counts. The project has been researched and funded and is now in the implementation phase. We will cover the specifics of the solution and the details of the business and support models.
IBM Spectrum Scale Best Practices for Genomics Medicine WorkloadsUlf Troppens
Genomics medicine requires physicians, data scientists and researchers to analyze huge amounts of genomics data quickly. The IBM Spectrum Scale Best Practices for Genomics Medicine Workload provides composable infrastructure that enables IT architects to customize deployments for varying functional and performance needs. The described scale-out architecture is capable to store, access and manage genomics data from a few 100 TB to tens of PB. The solution integrates compute resources and an easy-to-use Web User Interface to submit high-throughput batch jobs to analyze genomics data sets. While the best practices are optimized for genomics medicine workloads, most of the settings are generic and applicable to other workloads and industries.
Class lecture by Prof. Raj Jain on Storage Virtualization. The talk covers Disk Arrays, Data Access Methods, SCSI (Small Computer System Interface), Advanced Technology Attachment (ATA), ESCON and FICON, Fibre Chanel, Fibre Channel Devices, Fibre Channel Protocol Layers, Fibre Channel Flow Control, Fibre Channel Classes of Service, What is Storage Virtualization?, Benefits of Storage Virtualization, Virtualizing Storage, RAID Levels, Nested RAIDs, Synchronous vs. Asynchronous Replication, Virtual Storage Area Network (VSAN), Physical Storage Network, Virtual Storage Network, SAN vs. NAS, iSCSI (Internet Small Computer System Interface), iFCP (Internet Fiber Channel Protocol), FCIP (Fibre Channel over IP), FCoE (Fibre Channel over Ethernet), Virtual File Systems. Video recording available in YouTube.
SoftLayer Storage Services Overview (for Interop Las Vegas 2015)Michael Fork
Introduction to SoftLayer's Storage Services. Topics covered include Block and File offerings Endurance, Performance, Mass Storage Servers (QuantaStor), and Backup (EVault, R1Soft), Object Storage (OpenStack Swift), CDN, Data Transfer Service, and Aspera.
In this session, we’ll focus exclusively on OpenStack Swift, OpenStack’s object store capability. We’ll review the architecture, use cases, deployment strategies and common obstacles as we “open up the covers” on this exciting element of the OpenStack architecture.
Introduction to types of cloud storage and overview and comparison of the SoftLayer Storage Services. Topics covered include Block and File offerings"Codename: Prime", Consistent Performance, Mass Storage Servers (QuantaStor), and Backup (EVault, R1Soft), Object Storage (OpenStack Swift), CDN, Data Transfer Service, and Aspera.
With AWS, you can choose the right storage service for the right use case. Given the myriad of choices, from object storage to block storage, this session will profile details and examples of some of the choices available to you, with details on real world deployments from customers who are using Amazon Simple Storage Service (Amazon S3), Amazon Elastic Block Store (Amazon EBS), Amazon Glacier, and AWS Storage Gateway.
Learn how AWS customers save money, time and effort by using AWS's backup and archive services. Organizations of all sizes rely on AWS services to durably safeguard their data off-premises at a surprisingly low cost. This session will illustrate backup and archive architectures that AWS customers are benefitting from today.
Integrating On-premises Enterprise Storage Workloads with AWS (ENT301) | AWS ...Amazon Web Services
AWS gives designers of enterprise storage systems a completely new set of options. Aimed at enterprise storage specialists and managers of cloud-integration teams, this session gives you the tools and perspective to confidently integrate your storage workloads with AWS. We show working use cases, a thorough TCO model, and detailed customer blueprints. Throughout we analyze how data-tiering options measure up to the design criteria that matter most: performance, efficiency, cost, security, and integration.
6 Storage Workloads Ideal for Microsoft AzureBuurst
Is your organization looking to move on-premises storage workloads to Microsoft Azure?
We’ve helped hundreds of our customers move their storage workloads to Azure--without re-architecting their applications. We’ll review the 6 on-premises storage workloads ideal to move to Azure today.
In this webinar, we covered:
-6 Ideal Workloads to migrate including disaster recovery, cloud backup and more!
-Lessons Learned from helping customers do cloud workload migrations
-How to migrate on-premises file storage to Azure
-How to extend native Azure storage capabilities
-What cloud storage offers that on-premises storage options can’t
-Demo: deploy a virtual NAS on Azure in minutes
Learn more at https://www.softnas.com/azure.
This session is for IT pros working with compliance managers to deliver solutions that lower costs and still meet compliance demands. You will learn how to move large scale data stores to the cloud, while remaining compliant with existing regulations. Services mentioned: S3, Glacier and the Vault Lock feature, Snowball, ingestion services.
Compared to storing long-term datasets on-premise, archiving in the cloud is a smart alternative whether you’re looking for an active archive solution, tape replacement, or to fulfill a compliance requirement. Learn how AWS customers are simplifying their archiving strategies and meeting compliance needs using Amazon Glacier.
AWS re:Invent 2016: Three Customer Viewpoints: Private Equity, Managed Servic...Amazon Web Services
Customers seeking to make transformations to their environment typically gain efficiencies in both operational and capital costs. This session presents three different case studies, each from a different category of operations – Private Equity, Managed Services, and Government. In Government, General Dynamics will discuss how implementing AWS storage as part of their DevOps environment, saving $700k per year. Carlyle Group will show how replacing their on-premises storage with cloud based solutions brought new capabilities to their organization. Finally, Cobalt Iron will discuss how moving their customers to cloud backup targets allowed them to deliver 2M cloud backups per month.
Image steganography techniques can be classified into two major categories such as spatial domain techniques and frequency domain techniques.
In spatial domain techniques the secret message is hidden inside the image by applying some manipulation over the different pixels of the image.
In frequency domain techniques the image is transformed to another form by applying a transformation like discrete wavelet transform and then the message is hidden by applying any of the usual embedding techniques.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
2. Section II
Storage Networking
Technologies and Virtualization
In This Section:
• Chapter 5: Direct-Attached Storage and Introduction to SCSI
• Chapter 6: Storage Area Network
• Chapter 7: Network-Attached Storage
• Chapter 8: IP SAN
• Chapter 9: Content-Addressed Storage
• Chapter 10: Storage Virtualization
Ali Broumandnia, Broumandnia@gmail.com 2
3. Chapter 9
Content-Addressed Storage
• 9.1 Fixed Content and Archives
• 9.2 Types of Archives
• 9.3 Features and Benefits of CAS
• 9.4 CAS Architecture
• 9.5 Object Storage and Retrieval in CAS
• 9.6 CAS Examples
• 9.7 Concepts in Practice: EMC Centera
• Summary
Ali Broumandnia, Broumandnia@gmail.com 3
4. Key Concepts:
• Fixed Content and Archives
• Single-Instance Storage
• Object Storage and Retrieval
• Content Authenticity
Ali Broumandnia, Broumandnia@gmail.com 4
5. • In the life cycle of information, data is actively created, accessed, edited, and
changed. As data ages, it becomes less likely to change and eventually
becomes “fixed” but continues to be accessed by multiple applications and
users. This data is called fixed content.
• Accumulations of fixed content such as documents, e-mail messages, web
pages, and digital media throughout an organization have resulted in an
unprecedented growth in the amount of data. It has also introduced the
challenge of managing fixed content.
• Furthermore, users demand assurance that stored content has not changed and
require an immediate online access to fixed content. These requirements
resulted in the development of Content-Addressed Storage (CAS).
Ali Broumandnia, Broumandnia@gmail.com 5
6. • CAS is an object-based system that has been purposely built for
storing fixed content data. It is designed for secure online storage
and retrieval of fixed content. Unlike file-level and block-level
access that use file names and the physical location of data for
storage and retrieval, CAS stores user data and its attributes as
separate objects.
• The stored object is assigned a globally unique address known as a
content address (CA). This address is derived from the object’s
binary representation. CAS provides an optimized and centrally
managed storage solution that can support single-instance
storage (SiS) to eliminate multiple copies of the same data.
Ali Broumandnia, Broumandnia@gmail.com 6
7. What is in this chapter:
• Describes fixed content and archives,
• Traditional solutions deployed for archives and their limitations,
• Features and benefits of CAS,
• CAS architecture,
• Storage and retrieval in a CAS environment,
• Examples of CAS solutions
Ali Broumandnia, Broumandnia@gmail.com 7
8. 9.1 Fixed Content and Archives
• Data is accessed and modified at varying frequencies between the time it is
created and discarded.
• Some data frequently changes.
• Some data that does not typically change, but is allowed to change if required is.
• Another category of data is fixed content, which defines data that cannot be
changed.
• It is mandatary for all organizations to retain some data for an extended period of
time due to government regulations and legal/contractual obligations. Fixed data,
which is retained for future reference or business value, is referred to as fixed
content asset. (see Figure 9-1).
Ali Broumandnia, Broumandnia@gmail.com 8
10. • Organizations make use of these digital assets to generate new
revenues, improve service levels, and leverage historical value. This
demands frequent and quick retrieval of the fixed contents at
multiple locations.
• An archive is a repository where fixed content is placed. Online
data availability in the archive can further increase the business
value of the referenced information.
Ali Broumandnia, Broumandnia@gmail.com 10
11. 9.2 Types of Archives
• An electronic data archive is a repository for data that has fewer access
requirements. It can be implemented as online, nearline, or offline based on the
means of access:
• Online archive: The storage device is directly connected to the host to make
the data immediately available. This is best suited for active archives.
• Nearline archive: The storage device is connected to the host and
information is local, but the device must be mounted or loaded to access the
information.
• Offline archive: The storage device is not directly connected, mounted, or
loaded. Manual intervention is required to provide this service before
information can be accessed.
Ali Broumandnia, Broumandnia@gmail.com 11
12. • An archive is often stored on a write once read many (WORM)
device, such as a CD-ROM.
• Some tape devices also provide this functionality by implementing
filelocking capabilities in the hardware or software.
• Requirements to retain archives have caused corporate archives to
grow at a rate of 50 percent or more per year. At the same time,
organizations must reduce costs while maintaining required service-
level agreements (SLAs). Therefore, it is essential to find a solution
that minimizes the fixed costs of the archive’s operations and
management.
Ali Broumandnia, Broumandnia@gmail.com 12
13. 9.3 Features and Benefits of CAS
• CAS has emerged as an alternative to tape and optical
solutions because it overcomes many of their obvious
deficiencies.
• CAS also meets the demand to improve data
accessibility and to properly protect, dispose of, and
ensure service-level agreements for archived data.
Ali Broumandnia, Broumandnia@gmail.com 13
14. 9.3 Features and Benefits of CAS
• Content authenticity
• Content integrity
• Location independence
• Single-instance storage (SiS)
• Retention enforcement
• Record-level protection and disposition
• Technology independence
Ali Broumandnia, Broumandnia@gmail.com 14
15. Content authenticity:
It assures the genuineness of stored content. This is achieved by
generating a unique content address and automating the process
of continuously checking and recalculating the content address for
stored objects. Content authenticity is assured because the
address assigned to each piece of fixed content is as unique as a
fingerprint. Every time an object is read, CAS uses a hashing
algorithm to recalculate the object’s content address as a
validation step and compares the result to its original content
address. If the object fails validation, it is rebuilt from its mirrored
copy.
Ali Broumandnia, Broumandnia@gmail.com 15
16. Content integrity:
Refers to the assurance that the stored content has not been altered.
Use of hashing algorithm for content authenticity also ensures content
integrity in CAS. If the fixed content is altered, CAS assigns a new
address to the altered content, rather than overwrite the original fixed
content, providing an audit trail and maintaining the fixed content in
its original state. As an integral part of maintaining data integrity and
audit trail capabilities, CAS supports parity RAID protection in addition
to mirroring. Every object in a CAS system is systematically checked in
the background. Over time, every object is tested, guaranteeing
content integrity even in the case of hardware failure, random error, or
attempts to alter the content with malicious intent.
Ali Broumandnia, Broumandnia@gmail.com 16
17. Location independence:
• CAS uses a unique identifier that applications can leverage to
retrieve data rather than a centralized directory, path names, or
URLs. Using a content address to access fixed content makes the
physical location of the data irrelevant to the application requesting
the data. Therefore the location from which the data is accessed is
transparent to the application. This yields complete content mobility
to applications across locations.
Ali Broumandnia, Broumandnia@gmail.com 17
18. Single-instance storage (SiS):
• The unique signature is used to guarantee the storage of only a
single instance of an object. This signature is derived from the
binary representation of the object. At write time, the CAS
system is polled to see if it already has an object with the same
signature. If the object is already on the system, it is not stored,
rather only a pointer to that object is created. SiS simplifies
storage resource management tasks, especially when handling
hundreds of terabytes of fixed content.
Ali Broumandnia, Broumandnia@gmail.com 18
19. Retention enforcement:
• Protecting and retaining data objects is a core requirement of an
archive system. CAS creates two immutable components: a data
object and a meta-object for every object stored. The meta
object stores object’s attributes and data handling policies. For
systems that support object-retention capabilities, the retention
policies are enforced until the policies expire.
Ali Broumandnia, Broumandnia@gmail.com 19
20. Record-level protection and disposition:
• All fixed content is stored in CAS once and is backed up with a
protection scheme. The array is composed of one or more
storage clusters. Some CAS architectures provide an extra level
of protection by replicating the content onto arrays located at a
different location. The disposition of records also follows the
stringent guidelines established by regulators for shredding
and disposing of data in electronic formats.
Ali Broumandnia, Broumandnia@gmail.com 20
21. Technology independence:
The CAS system interface is impervious to technology changes.
As long as the application server is able to map the original
content address the data remains accessible. Although
hardware changes are inevitable, the goal of CAS hardware
vendors is to ensure compatibility across platforms.
Technology independence:
CAS maintains all content on disks that provide sub second “time
to first byte” (200 ms–400 ms) in a single cluster. Random disk
access in CAS enables fast record retrieval.
Ali Broumandnia, Broumandnia@gmail.com 21
22. 9.4 CAS Architecture
• The CAS architecture is shown in Figure 9-2. A client accesses the CAS-Based
storage over a LAN through the server that runs the CAS API (application
programming interface). The CAS API is responsible for performing functions
that enable an application to store and retrieve the data. CAS architecture is a
Redundant Array of Independent Nodes (RAIN). It contains storage nodes and
access nodes networked as a cluster by using a private LAN that is internal to
it. The internal LAN can be reconfigured automatically to detect the
configuration changes such as the addition of storage or access nodes. Clients
access the CAS on a separate LAN, which is used for interconnecting clients
and servers to the CAS. The nodes are configured with low-cost, high-capacity
ATA HDDs. These nodes run an operating system with special software that
implements the features and functionality required in a CAS system.
Ali Broumandnia, Broumandnia@gmail.com 22
24. • When the cluster is installed, the nodes are configured with a “role” defining the
functionality they provide to the cluster. A node can be configured as a storage
node, an access node, or a dual-role node. Storage nodes store and protect
data objects. They are sometimes referred to as back-end nodes. Access nodes
provide connectivity to application servers through the customer’s LAN. They
establish connectivity through a private LAN to the storage nodes in the cluster.
The number of access nodes is determined by the amount of user required
throughput from the cluster. If a node is configured solely as an “access node,”
its disk space cannot be used to store data objects. This configuration is
generally found in older installations of CAS. Storage and retrieval requests are
sent to the access node via the customer’s LAN. Dual-role nodes provide both
storage and access node capabilities. This node configuration is more typical
than a pure access node configuration.
Ali Broumandnia, Broumandnia@gmail.com 24
25. • Almost all CAS products have the same features and options.
Some may be implemented differently, but the following features
are an essential part of any CAS solution:
Ali Broumandnia, Broumandnia@gmail.com 25
26. • Integrity checking: It ensures that the content of the file matches
the digital signature (hashed output or CA).
• Data protection and node resilience: This ensures that the
content stored on the CAS system is available in the event of disk
or node failure.
• Load balancing: Distributes data objects on multiple nodes to
provide maximum throughput, availability, and capacity
Ali Broumandnia, Broumandnia@gmail.com 26
27. • Scalability: Adding more nodes to the cluster without any
interruption to data access and with minimum administrative
overhead.
• Self diagnosis and repair: Automatically detects and repairs
corrupted objects and alert the administrator of any potential
problem.
• Report generation and event notification: Provides on-demand
reporting and event notification. A command-line interface (CLI)
or a graphical user interface (GUI) can produce various types of
reports.
Ali Broumandnia, Broumandnia@gmail.com 27
28. • Fault tolerance: Ensures data availability if a component of the
CAS system fails, through the use of redundant components
data protection schemes.
• Audit trails: Enable documentation of management activity and
any access and disposition of data.
Ali Broumandnia, Broumandnia@gmail.com 28
29. 9.5 Object Storage and Retrieval in CAS
• The process of storing and retrieving objects in CAS is explained
in Figures 9-3 and 9-4, respectively. This process requires an
understanding of the following CAS terminologies:
Ali Broumandnia, Broumandnia@gmail.com 29
30. • Application programming interface (API): A high-level implementation of an
interface that specifies the details of how clients can make service requests.
The CAS API resides on the application server and is responsible for storing
and retrieving the objects in a CAS system.
• Access profile: Used by access applications to authenticate to a CAS cluster
and by CAS clusters to authenticate themselves to each other for replication.
• Virtual pools: Enable a single logical cluster to be broken up into multiple
logical groupings of data.
• Binary large object (BLOB): The actual data without the descriptive
information (metadata). The distinct bit sequence of user data represents the
actual content of a file and is independent of the name and physical
Ali Broumandnia, Broumandnia@gmail.com 30
31. • Content address (CA): An object’s address, which is created by a hash
algorithm run across the binary representation of the object. While
a CA, the hash algorithm considers all aspects of the content, returning a
unique content address to the user’s application.
• C-Clip: A virtual package that contains data (BLOB) and its associated CDF.
The C-Clip ID is the CA that the system returns to the client application. It is
also referred as a C-Clip handle or C-Clip reference.
• C-Clip Descriptor File (CDF): An XML file that the system creates while
making a C-Clip. This file includes CAs for all referenced BLOBs and
metadata. Metadata includes characteristics of CAS objects such as size,
format, and expiration date.
Ali Broumandnia, Broumandnia@gmail.com 31
32. Referring to Figure 9-3, the data object storage process in a
CAS system is as follows:
• 1. End users present the data to be archived to the CAS API via an
application. The application server may also interact directly with
the source (e.g., an X-ray machine) that generated this fixed
content.
• 2. The API separates the actual data (BLOB) from the metadata and
the CA is calculated from the object’s binary representation.
• 3. The content address and metadata of the object are then
inserted into the C-Clip Descriptor File (CDF). The C-clip is then
transferred to and stored on the CAS system.
Ali Broumandnia, Broumandnia@gmail.com 32
33. • 4. The CAS system recalculates the object’s CA as a validation step
and stores the object. This is to ensure that the content of the object
has not changed.
• 5. An acknowledgment is sent to the API after a mirrored copy of the
CDF and a protected copy of the BLOB have been safely stored in the
CAS system. After a data object is stored in the CAS system, the API is
given a C-Clip ID and C-Clip ID is stored local to the application
server.
• 6. Using the C-Clip ID, the application can read the data back from
the CAS system.
Ali Broumandnia, Broumandnia@gmail.com 33
35. The process of data retrieval from CAS follows these
steps:
• 1. The end user or an application requests an object.
• 2. The application queries the local table of C-Clip IDs stored in
the local storage and locates the C-Clip ID for the requested
object.
• 3. Using the API, a retrieval request is sent along with the C-Clip
ID to the CAS system.
• 4. The CAS system delivers the requested information to the
application, which in turn delivers it to the end user.
Ali Broumandnia, Broumandnia@gmail.com 35
37. Summary:
• Understanding the information lifecycle enables storage designers
to adopt different storage technology to meet data archival
requirements. CAS offers a centrally managed networked storage
solution for fixed content. CAS has enabled IT organizations to
realize significant cost savings and improve operational efficiency
and data protection. CAS meets stringent regulatory requirements
that have helped organizations avoid penalties and issues
associated with regulatory noncompliance. CAS eliminates data
duplication with SiS, which reduces IT maintenance costs and
increases long-term return on investment for storage systems.
Ali Broumandnia, Broumandnia@gmail.com 37
38. • Storage networking technologies have provided the primary
architecture that has evolved and matured to meet the business’s
information storage demands. As this evolution continues,
organizations have started using multi-vendor and multi-
generation technologies that must work together without
compromising business requirements. The virtualization
technologies described in the next chapter provides more
flexibility and scalability to meet next-generation storage needs.
Ali Broumandnia, Broumandnia@gmail.com 38