This document discusses how choosing the right NAS platform can help organizations address challenges related to rapidly growing data and flat budgets. It recommends looking for a solution that can scale to meet future capacity and performance needs efficiently over 3-4 years, drive capacity efficiencies through deduplication, integrate well with VMware, simplify storage administration, and streamline upgrades. The document then introduces the Hitachi NAS Platform 4000 series as a solution that can provide these benefits, helping organizations consolidate storage, improve productivity, and reduce costs and complexity.
Storage Analytics: Transform Storage Infrastructure Into a Business EnablerHitachi Vantara
View this webinar session to learn how you can transform your storage infrastructure into a business enabler. You will learn: Tips and tricks to streamline storage performance monitoring across your Hitachi environment. How to define and enforce performance and capacity objectives for key business applications by establishing storage service level management. How to create storage service level management reports that satisfy the needs of multiple IT stakeholders (that is, CIO, architect, administrator). For more information on controlling costs of sprawling storage with storage analytics white paper: http://www.hds.com/assets/pdf/hitachi-white-paper-control-costs-and-sprawling-storage-with-storage-analytics.pdf
A-B-C Strategies for File and Content BrochureHitachi Vantara
Explains each strategy, including archive 1st, back up less, consolidate more, distributed IT efficiency, enable e-discovery and compliance, and facilitate cloud. For more information on Unstructured Data Management Solutions by HDS please visit: http://www.hds.com/solutions/it-strategies/unstructured-data-management.html?WT.ac=us_mg_sol_udm
Reduce Costs and Complexity with Backup-Free StorageHitachi Vantara
The growth in unstructured data stresses traditional backup and restore operations. Numerous, disparate systems with large numbers of files and duplicate copies of data increase backup and restore times and hurt the performance and availability of production systems. Cost and complexity rise, with more backup instances to buy and manage, more care and handling of an increasing numbers of tapes, and more management of offsite storage. In addition, you may need to support analytics, a compliance audit, or legal action that needs information that is stored offsite. By tiering data to an archive, you can reduce total backup volume by at least 30%. By extending that core archive to the edges of your business, your potential gains are worth investigating. View this webcast to learn how to: Lower capital expenses (hardware, software, licensing, and so on). Control maintenance costs. Simplify management complexity. Reduce backup volume, time cost, and time and administrative effort. For more information on Hitachi Data Systems File and Content Solutions please visit: http://www.hds.com/products/file-and-content/?WT.ac=us_mg_pro_filecont
Storage Analytics: Transform Storage Infrastructure Into a Business EnablerHitachi Vantara
View this webinar session to learn how you can transform your storage infrastructure into a business enabler. You will learn: Tips and tricks to streamline storage performance monitoring across your Hitachi environment. How to define and enforce performance and capacity objectives for key business applications by establishing storage service level management. How to create storage service level management reports that satisfy the needs of multiple IT stakeholders (that is, CIO, architect, administrator). For more information on controlling costs of sprawling storage with storage analytics white paper: http://www.hds.com/assets/pdf/hitachi-white-paper-control-costs-and-sprawling-storage-with-storage-analytics.pdf
A-B-C Strategies for File and Content BrochureHitachi Vantara
Explains each strategy, including archive 1st, back up less, consolidate more, distributed IT efficiency, enable e-discovery and compliance, and facilitate cloud. For more information on Unstructured Data Management Solutions by HDS please visit: http://www.hds.com/solutions/it-strategies/unstructured-data-management.html?WT.ac=us_mg_sol_udm
Reduce Costs and Complexity with Backup-Free StorageHitachi Vantara
The growth in unstructured data stresses traditional backup and restore operations. Numerous, disparate systems with large numbers of files and duplicate copies of data increase backup and restore times and hurt the performance and availability of production systems. Cost and complexity rise, with more backup instances to buy and manage, more care and handling of an increasing numbers of tapes, and more management of offsite storage. In addition, you may need to support analytics, a compliance audit, or legal action that needs information that is stored offsite. By tiering data to an archive, you can reduce total backup volume by at least 30%. By extending that core archive to the edges of your business, your potential gains are worth investigating. View this webcast to learn how to: Lower capital expenses (hardware, software, licensing, and so on). Control maintenance costs. Simplify management complexity. Reduce backup volume, time cost, and time and administrative effort. For more information on Hitachi Data Systems File and Content Solutions please visit: http://www.hds.com/products/file-and-content/?WT.ac=us_mg_pro_filecont
Microsoft SQL Server 2012 Data Warehouse on Hitachi Converged PlatformHitachi Vantara
Accelerate breakthrough insights across your organization with Microsoft SQL Server 2012 Data Warehouse running on the mission-critical and ready-to-deploy Hitachi server-storage-networking platform, Hitachi Unified Compute Platform. Amplify infrastructure performance with Hitachi and Microsoft SQL Server 2012 Fast Track Data Warehouse xVelocity in-memory technologies. Learn how your organization can extract 100 million+ records in 2 or 3 seconds versus the 30 minutes required previously. With SQL Server 2012 Fast Track Data Warehouse and Hitachi software, your organization will be able to leverage a data platform that processes any data anywhere. View this webcast and learn:How to reduce deployment time with ready-to-deploy solutions that have been engineered and pre-configured by Hitachi and validated by the Microsoft Fast Track Data Warehouse program. How Hitachi and Microsoft have optimized performance for your data warehouse requirements. How your organization can realize immediate ROI from your data warehouse investment. For more information on Hitachi Unified Compute Platform please visit: http://www.hds.com/products/hitachi-unified-compute-platform/?WT.ac=us_mg_pro_ucp
Power the Creation of Great Work Solution ProfileHitachi Vantara
This solution discusses how quality and speed are critical in solving storage and data management bottlenecks, delivering cost-effective solutions that are highly scalable for post-production tasks. Whether CGI animation, rendering, or transcoding, Hitachi Data Systems powers digital workflows, enabling extraordinary creative and business achievements with HUS and HNAS infrastructure offerings. For more information on Hitachi Unified Storage and Hitachi NAS Platform 4000 Series please visit: http://www.hds.com/products/file-and-content/network-attached-storage/?WT.ac=us_mg_pro_hnasp
Learn more about Hitachi Content Platform Anywhere by visiting http://www.hds.com/products/file-and-content/hitachi-content-platform-anywhere.html
and more information on the Hitachi Content Platform is at http://www.hds.com/products/file-and-content/content-platform
Simplify Data Center Monitoring With a Single-Pane ViewHitachi Vantara
Keeping IT systems up and well tuned requires constant attention, but the task is too often complicated by separate monitoring tools required to watch applications, servers, networks and storage. This white paper discusses how system administrators can consolidate oversight of these components, particularly where DataCore SANsymphony V storage hypervisor virtualizes the storage resources. Such visibility is made possible through the integration of SANsymphony-V with Hitachi IT Operations Analyzer.
Fully leveraging your data, infrastructure, and IT staff has never been more important than it is now, during these times of fiscal responsibility and evolving business demands. In response, businesses need to maximize their IT by getting increased performance, efficiency, and economics out of their infrastructure and resources.
This presentation focuses on three key technologies that provide particularly compelling opportunities to maximize IT:
-All-flash systems that accelerate access to information for faster decision-making, analysis and productivity.
-Unified storage solutions that enable you to process more, and diverse, workloads in less time while driving capacity efficiencies.
-Unified compute solutions that deliver improved orchestration and automation and enhance the productivity of your IT staff, while avoiding costly over- or under-provisioning.
Do more in your data center with the Hitachi Compute Blade 500 blade server. This highly reliable enterprise platform is designed for virtualization and is the ideal platform for cloud computing applications.
The economics of storage virtualization webinarHitachi Vantara
Virtualization in the data center is a stable and proven approach to mke IT more efficient, from desktops to servers and from networks to storage. Whether storage virtualization is host-based, controller-based or through an appliance, it is a core ingredient in economical IT architectures. As with most new technology investments, you need a clear understanding of benefits versus costs.
Big Data – Shining the Light on Enterprise Dark DataHitachi Vantara
Content stored for a business purpose is often without structure or metadata required to determine its original purpose. With Hitachi Data Discovery Suite and Hitachi Content Platform, businesses can uncover dark data that could be leveraged for better business insight and uncover compliance issues that could prevent business risks. View this session and learn: What is enterprise dark data? How can enterprise dark data impact business decisions? How can you augment your underutilized data and deliver more value? How can you decrease the headache and challenges created by dark data? For more information please visit: http://www.hds.com/products/file-and-content/
Hitachi Unified Storage 100 family systems consolidate and manage block, file and object data on a central platform. For more information on our unified storage please visit: http://www.hds.com/products/storage-systems/hitachi-unified-storage-100-family.html?WT.ac=us_mg_pro_hus100
Achieve Higher Quality Decisions Faster for a Competitive Edge in the Oil and...Hitachi Vantara
Hitachi next-generation unified storage solutions meet the challenges of today’s data-intensive oil and gas exploration and production activities. For more information on Hitachi Unified Storage and Hitachi NAS Platform 4000 series please visit: http://www.hds.com/products/file-and-content/network-attached-storage/?WT.ac=us_mg_pro_hnasp
Build the Optimal Mainframe Storage ArchitectureHitachi Vantara
White Paper which discusses the advantages of FICON networked storage. The paper focuses specifically on Hitachi VSP and Brocade 8510. It discusses why networked FICON and describes the Hitachi VSP enterprise storage and the Brocade 8510 director.
For more information on HDS and Brocade Solutions please visit: http://www.hds.com/products/networking/fibrechannel/brocade.html
Explains how backup-free storage reduces cost and complexity; provides benefits of Hitachi Content Platform; includes brief HDS backup use cases.
For more information on our Unstructured Data Management Solutions please check: http://www.hds.com/go/hitachi-abc-ebook-managing-data/
Accelerate the Business Value of Enterprise StorageHitachi Vantara
When it comes to enterprise storage, IT has always had to choose between features and cost. Ongoing tradeoffs between the best technologies to support business operations and an adequate budget to pay for those technologies generally impede an organization’s ability to be competitive, innovative and cost efficient. The entry-enterprise storage market has opened up new opportunities for storage customers – and eliminated the need for tradeoffs. Join this webinar to understand how to accelerate business value with entry-enterprise storage systems and learn about the new Hitachi Data System offering, Hitachi Unified Storage VM. View this WebTech to: Understand the common tradeoffs and challenges within the entry-enterprise storage market. Understand the business value of new entry-enterprise offerings. Learn how Hitachi Unified Storage VM is bringing enterprise-level features to the midrange. For more information on Hitachi Unified Storage VM please visit: http://www.hds.com/products/storage-systems/hitachi-unified-storage-vm.html?WT.ac=us_mg_pro_husvm
Face Data Challenges of Life Science Organizations With Next-Generation Hitac...Hitachi Vantara
Hitachi Unified Storage 100 family drives efficiency at reduced costs and improves the discovery-to-market cycle for life sciences organizations. For more information on Hitachi Unified Storage and Hitachi NAS Platform 4000 series please visit: http://www.hds.com/products/file-and-content/network-attached-storage/?WT.ac=us_mg_pro_hnasp
Microsoft SQL Server 2012 Data Warehouse on Hitachi Converged PlatformHitachi Vantara
Accelerate breakthrough insights across your organization with Microsoft SQL Server 2012 Data Warehouse running on the mission-critical and ready-to-deploy Hitachi server-storage-networking platform, Hitachi Unified Compute Platform. Amplify infrastructure performance with Hitachi and Microsoft SQL Server 2012 Fast Track Data Warehouse xVelocity in-memory technologies. Learn how your organization can extract 100 million+ records in 2 or 3 seconds versus the 30 minutes required previously. With SQL Server 2012 Fast Track Data Warehouse and Hitachi software, your organization will be able to leverage a data platform that processes any data anywhere. View this webcast and learn:How to reduce deployment time with ready-to-deploy solutions that have been engineered and pre-configured by Hitachi and validated by the Microsoft Fast Track Data Warehouse program. How Hitachi and Microsoft have optimized performance for your data warehouse requirements. How your organization can realize immediate ROI from your data warehouse investment. For more information on Hitachi Unified Compute Platform please visit: http://www.hds.com/products/hitachi-unified-compute-platform/?WT.ac=us_mg_pro_ucp
Power the Creation of Great Work Solution ProfileHitachi Vantara
This solution discusses how quality and speed are critical in solving storage and data management bottlenecks, delivering cost-effective solutions that are highly scalable for post-production tasks. Whether CGI animation, rendering, or transcoding, Hitachi Data Systems powers digital workflows, enabling extraordinary creative and business achievements with HUS and HNAS infrastructure offerings. For more information on Hitachi Unified Storage and Hitachi NAS Platform 4000 Series please visit: http://www.hds.com/products/file-and-content/network-attached-storage/?WT.ac=us_mg_pro_hnasp
Learn more about Hitachi Content Platform Anywhere by visiting http://www.hds.com/products/file-and-content/hitachi-content-platform-anywhere.html
and more information on the Hitachi Content Platform is at http://www.hds.com/products/file-and-content/content-platform
Simplify Data Center Monitoring With a Single-Pane ViewHitachi Vantara
Keeping IT systems up and well tuned requires constant attention, but the task is too often complicated by separate monitoring tools required to watch applications, servers, networks and storage. This white paper discusses how system administrators can consolidate oversight of these components, particularly where DataCore SANsymphony V storage hypervisor virtualizes the storage resources. Such visibility is made possible through the integration of SANsymphony-V with Hitachi IT Operations Analyzer.
Fully leveraging your data, infrastructure, and IT staff has never been more important than it is now, during these times of fiscal responsibility and evolving business demands. In response, businesses need to maximize their IT by getting increased performance, efficiency, and economics out of their infrastructure and resources.
This presentation focuses on three key technologies that provide particularly compelling opportunities to maximize IT:
-All-flash systems that accelerate access to information for faster decision-making, analysis and productivity.
-Unified storage solutions that enable you to process more, and diverse, workloads in less time while driving capacity efficiencies.
-Unified compute solutions that deliver improved orchestration and automation and enhance the productivity of your IT staff, while avoiding costly over- or under-provisioning.
Do more in your data center with the Hitachi Compute Blade 500 blade server. This highly reliable enterprise platform is designed for virtualization and is the ideal platform for cloud computing applications.
The economics of storage virtualization webinarHitachi Vantara
Virtualization in the data center is a stable and proven approach to mke IT more efficient, from desktops to servers and from networks to storage. Whether storage virtualization is host-based, controller-based or through an appliance, it is a core ingredient in economical IT architectures. As with most new technology investments, you need a clear understanding of benefits versus costs.
Big Data – Shining the Light on Enterprise Dark DataHitachi Vantara
Content stored for a business purpose is often without structure or metadata required to determine its original purpose. With Hitachi Data Discovery Suite and Hitachi Content Platform, businesses can uncover dark data that could be leveraged for better business insight and uncover compliance issues that could prevent business risks. View this session and learn: What is enterprise dark data? How can enterprise dark data impact business decisions? How can you augment your underutilized data and deliver more value? How can you decrease the headache and challenges created by dark data? For more information please visit: http://www.hds.com/products/file-and-content/
Hitachi Unified Storage 100 family systems consolidate and manage block, file and object data on a central platform. For more information on our unified storage please visit: http://www.hds.com/products/storage-systems/hitachi-unified-storage-100-family.html?WT.ac=us_mg_pro_hus100
Achieve Higher Quality Decisions Faster for a Competitive Edge in the Oil and...Hitachi Vantara
Hitachi next-generation unified storage solutions meet the challenges of today’s data-intensive oil and gas exploration and production activities. For more information on Hitachi Unified Storage and Hitachi NAS Platform 4000 series please visit: http://www.hds.com/products/file-and-content/network-attached-storage/?WT.ac=us_mg_pro_hnasp
Build the Optimal Mainframe Storage ArchitectureHitachi Vantara
White Paper which discusses the advantages of FICON networked storage. The paper focuses specifically on Hitachi VSP and Brocade 8510. It discusses why networked FICON and describes the Hitachi VSP enterprise storage and the Brocade 8510 director.
For more information on HDS and Brocade Solutions please visit: http://www.hds.com/products/networking/fibrechannel/brocade.html
Explains how backup-free storage reduces cost and complexity; provides benefits of Hitachi Content Platform; includes brief HDS backup use cases.
For more information on our Unstructured Data Management Solutions please check: http://www.hds.com/go/hitachi-abc-ebook-managing-data/
Accelerate the Business Value of Enterprise StorageHitachi Vantara
When it comes to enterprise storage, IT has always had to choose between features and cost. Ongoing tradeoffs between the best technologies to support business operations and an adequate budget to pay for those technologies generally impede an organization’s ability to be competitive, innovative and cost efficient. The entry-enterprise storage market has opened up new opportunities for storage customers – and eliminated the need for tradeoffs. Join this webinar to understand how to accelerate business value with entry-enterprise storage systems and learn about the new Hitachi Data System offering, Hitachi Unified Storage VM. View this WebTech to: Understand the common tradeoffs and challenges within the entry-enterprise storage market. Understand the business value of new entry-enterprise offerings. Learn how Hitachi Unified Storage VM is bringing enterprise-level features to the midrange. For more information on Hitachi Unified Storage VM please visit: http://www.hds.com/products/storage-systems/hitachi-unified-storage-vm.html?WT.ac=us_mg_pro_husvm
Face Data Challenges of Life Science Organizations With Next-Generation Hitac...Hitachi Vantara
Hitachi Unified Storage 100 family drives efficiency at reduced costs and improves the discovery-to-market cycle for life sciences organizations. For more information on Hitachi Unified Storage and Hitachi NAS Platform 4000 series please visit: http://www.hds.com/products/file-and-content/network-attached-storage/?WT.ac=us_mg_pro_hnasp
Why hitachi virtual storage platform does so well in a mainframe environment ...Hitachi Vantara
Hitachi VSP is a new paradigm in enterprise array performance. In this session we will discuss how the architecture of VSP enhances its box-wide performance. The results of performance testing with synthetic host I/O generators and the PAI/O driver will also be presented.
How and why to upgrade to hitachi device manager v7 webinarHitachi Vantara
Hitachi Device Manager v7 lets you simplify and control all your storage assets from a centralized console with improved usability, workflow, speed, scalability and task management. Whether you have already upgraded or are considering an upgrade to v7, please join us for this informative webtech session to learn the best practices for upgrading.
Consolidate More: High Performance Primary Deduplication in the Age of Abunda...Hitachi Vantara
Increase productivity, efficiency and environmental savings by eliminating silos, preventing sprawl and reducing complexity by 50%. Using powerful consolidation systems, Hitachi Unified Storage or Hitachi NAS Platform, lets you consolidate existing file servers and NAS devices on to fewer nodes. You can perform the same or even more work with fewer devices and lower overhead, while reducing floor space and associated power and cooling costs. View this webcast to learn how to: Shrink your primary file data without disrupting performance. Increase productivity and utilization of available capacity. Defer additional storage purchases. Save on power, cooling and space costs. For more information please visit: http://www.hds.com/products/file-and-content/network-attached-storage/?WT.ac=us_inside_rm_htchunfds
1- Lower total cost of ownership
2- A platform for rapid reporting and analytics
3- Increased scalability and availability
4- Support for new and emerging applications
5- Flexibility for hybrid environment
6- Greater simplicity
(Original share from Francisco González Jiménez)
1- Lower total cost of ownership
2- A platform for rapid reporting and analytics
3- Increased scalability and availability
4- Support for new and emerging applications
5- Flexibility for hybrid environment
6- Greater simplicity
How companies are managing growth, gaining insights and cutting costs in the ...Virginia Fernandez
6 reasons to upgrade your database:
Reason 1: Lower total cost of ownership
Reason 2: A platform for rapid reporting and analytics
Reason 3: Increased scalability and availability
Reason 4: Support for new and emerging applications
Reason 5: Flexibility for hybrid environments
Reason 6: Greater simplicity
How companies are managing growth, gaining insights
and cutting costs in the era of big data.
Top reasons to change your database:
1. Lower total cost of ownership
2. A platform for rapid reporting
and analytics
3. Increased scalability and
availability
4. Support for new and emerging
applications
5. Flexibility for hybrid environments
6. Greater simplicity
This article takes a look at some of the reasons behind this data explosion, and some of the possible effects if the growth is not managed. We’ll also examine some of the ways in which these problems can be avoided.
IDC Whitepaper: Achieving the full Business Value of VirtualizationDataCore Software
Are you struggling with how to choose the right storage virtualization solution, or just looking to achieve a scalable software-based storage virtualization solution that fits your budget? Consolidate storage and server assets
Increase the number of virtualized servers running on individual physical servers while doubling storage utilization rates for installed storage
Leverage lower-cost/higher-capacity storage tiers that can significantly cut the cost of acquiring new storage assets
Improve application and information availability while shrinking backup times
Significantly reduce the cost to meet the performance and business continuity objectives of virtualized IT organizations
CIO Guide to Using SAP HANA Platform For Big DataSnehanshu Shah
This guide supports CIOs in setting up a system infrastructure for their business that can get the best out of Big Data. We describe what the SAP HANA platform can do and how it integrates with Hadoop and related technologies interplay, looking at data life cycle management and data streaming. Concrete use cases point out the requirements associated with Big Data as well as the opportunities it offers, and how companies are already taking advantage of them.
TierPoint White Paper_When_Virtualization_Meets_Infrastructure_2015sllongo3
In a rapidly changing technology landscape, are your virtualized systems taking full advantage of the cloud? Read about how these solutions can be leveraged for production, development and disaster recovery.
Similar to Hitachi white-paper-future-proof-your-datacenter-with-the-right-nas-platform (20)
Hitachi Vantara and our special guest, Dr. Alison Brooks, Research Director at IDC, discuss:
• How video and other IoT data can help your business become smarter, safer and more efficient.
• How to harness IoT data to gain operational intelligence and achieve better business outcomes.
• How Hitachi’s customers are innovating with IoT to excel.
• Which practical applications and best practices will get you started on your own IoT journey to reach your goals and tackle your challenges.
Virtualizing SAP HANA with Hitachi Unified Compute Platform Solutions: Bring...Hitachi Vantara
Virtualizing SAP HANA with Hitachi Unified Compute Platform Solutions: Bringing Flexibility, Agility and Readiness to the Real-Time Enterprise. VMworld 2015
Hitachi Virtual Infrastructure Integrator (Virtual V2I) is a VMware vCenter plugin plus associated software. It provides data management efficiency for large VM environments. Specifically, the latest release addresses virtual machine backup and recovery and cloning services. Customer want to leverage storage based snapshots as it is scalable, more granular backup from hours between backups to minutes resulting in improved RPO. VMworld 2015.
Economist Intelligence Unit: Preparing for Next-Generation CloudHitachi Vantara
Preparing for next-generation cloud: Lessons learned and insights shared is an Economist Intelligence Unit (EIU) research programme, sponsored by Hitachi Data Systems. In this report, the EIU looks at companies’ experiences with cloud adoption and assesses whether the technology has lived up to expectations. Where the cloud has fallen short of expectations, we set out to understand why. In cases of seamless implementation, we gather best practices from firms using the cloud successfully.
HDS Influencer Summit 2014: Innovating with Information to Address Business N...Hitachi Vantara
Top Executives at HDS share how the company is Innovating with Information to address business needs. Learn how the company is transforming now and into the future. #HDSday.”
Information Innovation Index 2014 UK Research ResultsHitachi Vantara
Hitachi Data Systems releases insights from its inaugural ‘Information Innovation Index’, a UK research report, conducted by independent UK technology market research agency, Vanson Bourne, in which 200 IT decision-makers were surveyed during April 2014 to provide insights into how current approaches to IT are thwarting companies’ ambitions to leverage data to drive innovation and business growth.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
1. Consolidate and Future-Proof Your Data Center With
the Right NAS Platform
Achieve the Correct Balance Among Performance,
Capacity, Environmental Impact and TCO While
Reducing Your NAS Footprint
DATA DRIVEN GLOBAL VISION CLOUD PLATFORM STRATEG
ON POWERFUL RELEVANT PERFORMANCE SOLUTION CLO
VIRTUAL BIG DATA SOLUTION ROI FLEXIBLE DATA DRIVEN V
WHITEPAPER
By Hitachi Data Systems
August 2013
2. WHITE PAPER 2
Contents
Executive Summary 3
Introduction 3
The Reality of Flat Budgets and the Role of TCO 4
Not All NAS Solutions Are Created Equal 5
Scale With Confidence (Capacity and Performance) 5
Drive Capacity Efficiencies 5
Provide Tight VMware Integration 5
Simplify Storage Administration 6
Streamline Hardware and Software Upgrades and Updates 6
Leverage Industry-Leading Storage 6
Reduce Total Cost of Ownership 6
NAS for Today and Tomorrow: Hitachi NAS Platform 4000 Series 7
Win From Both Technical and Economic Perspectives 7
Improved Performance 7
Greater Scalability 8
High-Capacity Efficiency 8
Efficient VM Backup and Recovery 9
Seamless VMware Integration 9
TCO Savings, Simplified Management and Sustainability 9
HNAS in Action: 3 Real-World Examples 9
Software Firm Attains Twice the Performance, 66% Lower Capex 9
Game Maker Consolidates Servers and Storage, Reduces Build
Delivery Time by 67% 9
Global Semiconductor Firm Dedupes 1.2 million Files in 16 Minutes 9
Summary 10
3. WHITE PAPER 3
Consolidate and Future-Proof Your Data Center With the Right
NAS Platform
Executive Summary
In today’s economic climate, it’s a given that companies are under more pressure than ever to reduce costs, do more
with less, and stretch IT resources to the limit. However, when it comes to data growth, this formula is no longer
enough. In fact, for many organizations, data growth is in a near-crisis situation.
Unfortunately, most IT budgets simply cannot keep up with increasing data growth, forcing organizations to make
tradeoffs between what’s best for the business and what’s affordable. This can leave business opportunities to more
nimble competitors as companies look for ways to reduce the total cost of ownership (TCO) of their IT infrastructure.
One solution that’s helping data center administrators meet the data growth challenge is network attached storage
(NAS). The NAS growth story is an impressive one, growing at 60% year over year in many environments. However,
until recently, NAS implementations were made up of corporate “home” drives and small file sharing environments
such as audio, video, presentations and documents, and other rich media. When it came to growth, the answer was
to simply add more disks. While this low-cost strategy has worked in the past, these environments have grown to
include enterprise applications, resulting in inefficiencies and high cost. This area is ripe for improvements, consolida-
tion and cost reduction.
Today, NAS environments have been developed and configured to support:
■■ Fast-growing virtualization environments (virtual machine and virtual desktop environments).
■■ High-speed video streaming and transcoding.
■■ Large-capacity stores and file sharing environments for medium, large and enterprise organizations.
With these improvements to NAS technology, the NAS environment is now a critical piece of the data center for per-
formance and mission-critical applications. This paper examines the challenges facing most organizations within the
NAS environment, and how choosing the right NAS platform can address performance, capacity, efficiency and scal-
ability demands, while reducing overall TCO. In addition, this paper explains how storage economics plays a key role
in helping organizations obtain the maximum value out of their current storage investments.
Introduction
The story of explosive data and capacity growth is nothing new. Organizations have been forced to deal with 25% to
30% growth in data storage and performance requirements for years now. However, the growth is neither going away
nor getting any easier to deal with. In fact, it often feels like a ticking time bomb. Unstructured data such as social
media, digital images, contracts, video, presentations, Web pages and now more mission-critical workloads like
VMware are growing at unprecedented rates. As it relates to the NAS environment, there is seemingly no end to this
growth in sight.
At the same time, many organizations struggle to gain control over the spiraling costs, complexities and risks asso-
ciated with this data growth and data center sprawl. Consequently, the need to better utilize IT resources while
increasing performance and reducing costs has never been more critical. This is where choosing the right platform
can mean the difference between upgrading equipment after just 1 year or building a future-proof system that scales
for 3 or 4 years.
4. WHITE PAPER 4
The Reality of Flat Budgets and the Role of TCO
The story with IT budgets hasn’t changed either. Purse strings in the data center are as tight as ever and with valid
reasons. At the economic level, austerity programs, currency devaluation and government sequestration are all
impacting a company’s ability to support new technology and capacity in the data center.
At the infrastructure level, server sprawl and storage growth, as well as rising power, cooling and floor space costs,
and storage management costs all vie for dwindling budget dollars. This reality can force IT decision-makers to make
some tough choices. One such choice could be canceling vital IT projects that might effectively address data growth
or find other ways to reduce costs further.
With growing demand for capacity and performance in an environment of flat or declining budgets, IT needs a better
solution to meet these demands. It must focus on reducing the TCO of the solution over its useful life, instead of
focusing on acquisition price or cost per GB. Indeed, these upfront capital costs (hardware, software, and so forth)
are becoming less of a focus in purchasing decisions relative to the greater, more comprehensive metric of TCO. TCO
includes acquisition costs and operating costs like management and environmental expenses. In fact, a recent survey
by ESG reported that 65% of respondents felt TCO was the most important criteria for selecting a storage solution.1
A FORMULA FOR ACHIEVING TCO THROUGH THE RIGHT NAS PLATFORM
To understand the full benefits of the right NAS system, it may help to look at a hypothetical example
and ponder some essential questions. Let’s assume that a typical company requires 50TB of capac-
ity. Ideally, the company would like a new NAS solution to last 3 to 4 years without a technical refresh
or hardware upgrade to save on migration and business disruption costs. If this company is growing at
a steady 30% year over year pace, the following questions should be top of mind when considering a
NAS solution:
1. Does the NAS solution scale to meet capacity needs for the next 3 to 4 years? This is important to
note as current growth rates will require year 4 to be well over 100TB.
2. Does the NAS solution scale to meet growing performance demands? Many environments’ perfor-
mance requirements are growing at a rate equal to or higher than the capacity growth rate.
3. If the NAS solution can scale to meet both capacity and performance demands, is this accomplished
efficiently?
For example:
■■ What is the scaling impact? How many more NAS nodes will be needed?
■■ Does the NAS have capacity efficiency technology like deduplication to keep capacity
requirements down?
■■ How much data center space will be necessary?
■■ How much more storage management time and effort will be necessary?
■■ What is the impact on power, cooling and so forth?
4. Does the NAS solution have all the application integration needed to grow the environment over this
period of time? Will VMware integration be needed in year 2 or 3?
5. Can existing storage be leveraged? This is essential to know because many NAS systems require a
separate architecture at an additional cost.
1
ESG Research Report, 2013 IT Spending Intentions Survey, January 2012.
5. WHITE PAPER 5
Not All NAS Solutions Are Created Equal
There is a wide range of NAS solutions on the market, but not all are built to last more than 1 year without significant
upgrades, especially in growing environments. The right NAS solution can address current challenges of “server
sprawl” by consolidating existing systems from many to one. It can address rising capacity and data center costs by
providing superior capacity, management and environmental efficiency, as well as application integration. It can also
provide a “future-proof” investment by scaling up for service level agreement (SLA) and performance requirements
and scaling out for capacity demand (“leg room” for growing resource demands). Let’s examine some of the key attri-
butes an IT decision-maker should look for in a NAS solution.
Scale With Confidence (Capacity and Performance)
A NAS solution should provide a path for adding storage to meet new application or business needs, and to con-
solidate disparate storage into a single point of management. For best scalability, the solution should support shared
storage technologies, including network file system (NFS), server message block (SMB), iSCSI and Fibre Channel
SANs. A NAS solution that delivers linear scalability with each cluster node delivers higher input/output operations per
second (IOPS) and random workload optimization, accelerating overall throughput and productivity.
With the right NAS solution, businesses can scale to multiple nodes in a single storage cluster without incurring
additional overhead or downtime, and meet performance objectives with fewer nodes. A cluster meeting the same
performance with 8 nodes as a 16-node cluster results in lower power, cooling, floor space, and IT management
costs. Having a more powerful, consolidated system up front can lead to overall TCO savings.
Drive Capacity Efficiencies
Unstructured information dominates the data growth curve, especially in environments containing high numbers of
duplicated files and mail servers that store large numbers of file copies. A smart way to improve capacity efficiency
is through deduplication. Doing so reduces the file system size, does not conflict with normal I/O workloads, and
executes separately so it will not impact regular operations.
An optimal NAS solution drives efficiencies to enable significant economic benefits. For example, if a NAS system
could reclaim 40% of a terabyte of duplicated file data, the result is 400GB of space reclaimed for other uses. This
allows organizations to stretch existing capacity to store more data than normal and defer purchases of additional
storage. That means less storage space for the same files, more value and capacity per spindle, and lower overall
operating costs.
Provide Tight VMware Integration
Random workloads and data services, both virtualized and nonvirtualized, can severely tax the storage ecosystem
because I/O varies greatly between peak workload (boot, login and virus scan storms) and steady-state workload.
This also forces the need for almost instantaneous, granular recovery of virtual machines, fast snapshots, deduplica-
tion and improved recovery point and recovery time objectives (RPO/RTO).
As virtual machine (VM) environments grow, storage administrators can be consumed with processes that can
become inefficient and costly. For example, they must address the challenges of cloning VMs, creating and restor-
ing snapshots more frequently, and having fast replication in place to move secondary copies to off-site recovery
locations. To ensure there is no improperly sized storage that results in greater hardware requirements and resulting
costs, administrators need to address storage capacity and density. They can use nondisruptive tiering to free up
inactive or stale user data, while properly sizing storage resources.
The optimal NAS solution efficiently addresses the unique storage requirements of VMware environments. It stream-
lines the entire VM lifecycle by providing fast snapshot, replication and cloning utilities. These utilities save time for IT
administrators, allowing them to grow their VMware deployment and extend data protection best practices while free-
ing up storage capacity and reducing costs.
6. WHITE PAPER 6
Simplify Storage Administration
Streamlining and centralizing storage management tasks reduces complexity while lowering the need for more IT
staff. A NAS platform should provide a unified dashboard view of block, file and object consumers, solve issues
quickly, and simplify operations with fewer management steps. With more comprehensive, easy-to-understand stor-
age analytics, organizations can have the business-oriented view necessary to regain control of fast-growing storage
environments.
Another necessity is seamless integration with VMware vCenter for quick and efficient data management of vir-
tual server and virtual desktop infrastructure (VDI) environments. This integration allows administrators to manage
storage-based snapshots for flexible scheduling and automated backups. By working in unison, these tools provide
end-to-end storage performance monitoring and consolidate complex storage-management statistics to enable
better-informed decisions. IT can then optimize capacity and performance utilization, and improve the use of existing
storage resources, while at the same time planning for future growth.
Streamline Hardware and Software Upgrades and Updates
Disruptive hardware and software upgrades that involve data migration can challenge even the most efficient data
center. Moving data from one system to another can result in application downtime, business disruption, and even
revenue loss. The good news for many administrators is that the right technology is available to alleviate or eliminate
all disruptions associated with these updates.
For the simplest of updates such as software upgrades (including operating system), a critical consideration is
whether the vendor has a history of providing forward- and backward-compatible hardware and software. This will
ensure that data migration is not necessary. When a data migration is necessary, such as when upgrading a NAS
system’s hardware, options like host-based replication can provide a seamless transition. However, this often comes
at the expense of server performance.
Array-based data migration using storage virtualization provides the lowest risk migration by reducing or eliminat-
ing application downtime while preserving host performance. In addition, storage virtualization not only reduces the
outage window to applications to speed up technology refresh, but it can also extend the longevity of the system.
Leverage Industry-Leading Storage
Organizations should not have to buy completely new infrastructure and be limited to having block storage isolated
as one silo and NAS storage as another. True consolidation breaks down silos by leveraging existing storage frame-
works and tightly integrating disparate data types through the storage gateway, which also improves utilization of
3rd-party storage.
Organizations are looking for enterprise-level storage features that are cost-efficient and optimized. The right NAS
solution utilizes a powerful storage foundation that manages and controls other storage systems in a virtualized
environment. At the enterprise level, the storage system should provide a wide range of high-end features, includ-
ing dynamic provisioning and dynamic tiering to improve system performance and storage efficiency at a lower cost.
Dynamic tiering also gives IT the option of mixing high-performance disks such as flash drives with slower, lower-cost
disks, while still achieving the same overall high performance.
Reduce Total Cost of Ownership
Clearly, procuring a system with this combination of superior scale, integration and capacity efficiency against a back-
drop of flat to declining costs may seem like a daunting task. However, looking at TCO demonstrates that including
these elements in the purchasing decision will ultimately reduce costs in growing environments.
7. WHITE PAPER 7
Storage Economics Delivers a Cost Reduction Framework
While choosing the right NAS solution is essential to managing data growth and keeping costs in line with budgets,
storage economics can help organizations get the maximum value from their current storage investments. Storage
Economics from Hitachi Data Systems is a unique methodology that calculates the 34 costs contributing to TCO of
an organization’s storage. Furthermore, Hitachi can assist companies in mapping technologies that are proven to
reduce the cost areas most important to their business.
Storage economics is about understanding the costs and identifying the technology framework to deliver the TCO
and the performance benefits that meet each organization’s unique needs. According to IDC, Storage Economics
from HDS takes an approach that helps to demonstrate the true value of relatively complex storage architectures,
and allows return on investment (ROI) analysis to become a routine part of the storage investment decision-making
process.
The Right NAS Will Reduce TCO and Future-Proof Your Data Center
When companies are able to scale a storage environment more efficiently, the net results are more productivity
and less administrative overhead. With fewer hardware and software components, storage management costs are
reduced. And lower power, cooling and floor space costs can yield savings in a smaller, more cost-effective data
center. When you look at the bottom line after 3 years of deploying the right NAS solution, the overall TCO savings
are significant.
NAS for Today and Tomorrow: Hitachi NAS Platform 4000 Series
Hitachi NAS Platform (HNAS) 4000 series provides the highest performance and usable capacity in the industry to
help you consolidate your storage and save time and money. HNAS is an advanced, high-performance and highly
scalable network attached storage system.
HNAS offers nondisruptive deduplication for maximum capacity efficiency with no degradation of performance. The
platform supports your largest big data workloads and consolidates multiple NAS filers at a lower TCO. Plus, HNAS
provides superior VMware integration, simplifies data management, and increases capacity utilization. At the same
time it offers the scalability and flexibility necessary to manage workloads of enterprise data centers, medium organi-
zations and remote data centers.
Moving to HNAS simplifies your overall IT infrastructure and reduces the complexity of storage management.
Advanced virtualization, tiering and deduplication capabilities give you the efficiency and agility you need, and the
automated operation results in lower operational costs. The bottom line: HNAS offers the exceptional performance
that businesses with performance-driven applications need to improve productivity and competitiveness.
Win From Both Technical and Economic Perspectives
Hitachi NAS Platform 4000 series enables you to gain technical advantages in performance, scale, efficiency, data
protection and VMware integration that today’s organizations need to stay ahead of competitors. HNAS also offers
economic benefits, including reduced TCO, fewer storage management requirements, and an environmentally friendly
footprint.
Improved Performance
From a technical perspective, HNAS helps your organization win in several ways. For example, many organizations
are dealing with thousands of random I/Os coming from email, online transaction processing (OLTP), and so forth.
And organizations deploying VMware over network file systems and scaling these environments are dealing with an
8. WHITE PAPER 8
exponential increase in IOPS requirements year over year. Some solutions address the resulting IOPS bottleneck
by adding more filers; however, this approach means more hardware, software and labor costs, as well as
increased TCO.
HNAS remedies this IOPS bottleneck problem by providing more consistent and linear performance at scale. It
accomplishes this improved performance by separating processing activities into core operations executed within
the FPGA system. Such activities may include snapshots and file system operations, and the data management func-
tions, including file locking within a CPU. In addition, HNAS also provides an “intelligent superflush,” which converts
random workloads to sequential block operations to deliver higher IOPS. The only hybrid-core VMware NFS platform
available, HNAS offers high efficiency and unique performance capabilities.
Because HNAS scales linearly as IOPS requirements increase, there’s no fall off due to CPU utilization. This scalability
provides a much higher level of maximum performance that translates into more users per system, more functions,
and fewer overall file servers and licenses. Reducing the number of file servers simplifies management so IT can focus
on more important issues.
Greater Scalability
It’s no secret that NAS environments are growing fast: in many cases at much higher rates than structured environ-
ments. That’s why organizations are looking to grow their systems out with increased metadata, snapshots, clones
and other requirements. However, some solutions have poor metadata management, issues with scale and LUN size,
as well as limitations on the number of VMs per LUN and snapshots.
Unique HNAS architecture allows you to efficiently traverse the file system, even at tremendous scale. It provides the
highest maximum number of clones and snapshots available: 100 million clones or snapshots per file server. And, it
enables cloning a 1TB VM in 4 seconds.
Scalability is critical to high-growth organizations, and HNAS supports 16TB capacity. To enable the same capac-
ity, competitors would have to deploy at least 3 times the number of systems to meet the same capacity as HNAS
4100. And HNAS also offers a file system that is twice as large. This capacity means there are fewer file systems to
manage, and strong Microsoft®
Windows®
storage consolidation allows more virtual servers and greater capacity
scaling.
High-Capacity Efficiency
Achieving high-capacity efficiency requires storing the maximum amount of data in the minimum space and at the
lowest cost. Capacity efficiency is also one way to determine how hard your system is working and measure a return
on assets. Hitachi NAS Platform capacity efficiency provides another technical win with exceptional deduplication, file
system overhead, and data tiering.
A strong deduplication solution can significantly reduce capacity costs. Yet many organizations struggle with dedu-
plication software that offers limited support and size, and adversely impacts system performance. In addition, some
NAS solutions require as much as 48% overhead for deduplication. Not HNAS.
HNAS solves these issues with a new breed of deduplication that uses auto throttling to eliminate performance
impacts, offers minimal overhead, and provides significantly higher scalability with fewer limitations. Hitachi also offers
industry-leading file tiering with a data migrator solution that automatically moves data between tiers or to Amazon
E3 Clouds, based on characteristics such as file age, type, size and so forth.
9. WHITE PAPER 9
Efficient VM Backup and Recovery
Today’s organizations need to back up and recover everything, and they need to it efficiently. However, some NAS VM
data protection solutions are not only constrained at the volume level, but they also require up to 20% of overhead.
HNAS provides highly efficient VM backup and recovery with up to 100 million snapshots and VM-level snapshots as
well as the capability to leverage a stable NDMP for integration with a wide variety of backup products. VM adminis-
trators can also employ Hitachi NAS Virtual Infrastructure Integrator software, which allows them to apply scheduling
and retention policies directly to HNAS, making VM RTO and RPO faster. For example, HNAS can restore a 1TB VM
in seconds.
Seamless VMware Integration
Integration with VMware gives an organization a technical edge, especially in achieving lower TCO in virtual environ-
ments. Most analysts agree that VM environments over NFS are growing at about 70% a year, making NFS the future
of virtual environments.
With an efficient, seamlessly integrated solution like HNAS, businesses can provision, back up and protect a NFS file
system and data store. By employing Hitachi NAS Virtual Infrastructure Integrator and a site recovery manager (SRM)
application they can complete these tasks in just 4 minutes.
TCO Savings, Simplified Management and Sustainability
With a host of technical advantages over competing solutions, added HNAS efficiencies offer more productivity, less
administration and clear economic benefits. For example, HNAS can quickly lower TCO by reducing NAS server
requirements.
If an organization has a growing NAS environment, HNAS can reduce TCO by up to 30% in 1 year. A key to
these lower costs is the reduction in storage management when scaling out NAS. A company simply needs fewer
resources to manage the same environment with HNAS. And fewer nodes and hardware makes HNAS a truly green,
environmentally friendly solution with reduced power, cooling and floor space.
HNAS in Action: 3 Real-World Examples
To illustrate how Hitachi NAS Platform helps organizations gain technical and economic advantages over competi-
tors, here are 3 examples from actual deployments.
Software Firm Attains Twice the Performance, 66% Lower Capex
When a leading design automation software firm needed to update its NAS solution, HNAS was one of the finalists
selected. After a head-to-head proof of concept that included a truly random mixed read and write test, HNAS was
2 times as fast and 1/3 the price of the competitor’s solution.
Game Maker Consolidates Servers and Storage, Reduces Build Delivery Time by 67%
A large video game maker challenged with data growth, I/O bottlenecks, system sprawl and backup issues needed
a cost-effective solution. With HNAS and VMware, they consolidated 32 servers and 3 storage systems and cut their
build delivery time by up to 67%.
Global Semiconductor Firm Dedupes 1.2 million Files in 16 Minutes
A global semiconductor company was looking to boost its primary storage efficiency. HNAS deduped 1.2 million files
in 16 minutes, with minimal impact on file server activity.
10. WHITE PAPER 10
Summary
Data growth is here to stay, and so is the demand for faster access to that data. However, IT budgets are not rising
fast enough to keep pace with the current solutions to the data growth challenges. The right NAS solution can
process more and increasingly diverse workloads faster, consolidate and scale with greater efficiently through dedu-
plication, and integrate tightly through intelligent tiering and automated data protection. Just as important, the right
NAS solution can fully align with an organization’s storage economics by delivering the highest value and lowest stor-
age TCO.
Hitachi NAS Platform provides all the storage features required to support a wide range of workloads, applications
and deployments, with a significantly smaller hardware and environmental footprint. Simplifying data protection, stor-
age efficiencies, high performance and massive scalability are all reasons why HNAS is an optimal solution for storage
economics.
HNAS also provides peace of mind that organizations can address data growth from all sources, including structured,
unstructured and virtualized, from one centrally managed storage system. At the same time, they realize the most
value from infrastructure investments.
For additional information about HNAS, please visit http://www.hds.com/products/file-and-content/
network-attached-storage/.