Panzura & Scality - Cloud Storage made seamless - Cloud Expo New York City 2012Marc Villemade
This is the full presentation I did with Ranajit Nevatia from Panzura (@ranajitN) at Cloud Computing Expo NY in June 2012.
It introduces and explains the concepts of Structured and Unstructured data and why Object Storage will prevail when it comes to the latter.
AWS Partner Presentation – Panzura – AWS Cloud Storage for the Enterprise 2012Amazon Web Services
Panzura provides a global cloud storage system that makes cloud storage seamless. It allows storage of both structured and unstructured data in the cloud with fast local read/write performance. This provides enterprises with unlimited, low-cost storage while maintaining high performance access to data anywhere in the world.
The document discusses several use cases for Panzura storage:
1) Archiving large files like medical images, market data, and seismic data in the cloud for lower costs, global access, and centralized management.
2) Backing up structured data to the cloud for infinite scalability, no tapes, and fast restores.
3) Sharing files globally between offices for collaboration without performance issues over the WAN.
Leveraging Swift Storage Policies using Scality RINGNicolas Trangez
We present several use-cases for Swift Storage Policies, and how they are implemented in the Scality RING Swift back-end.
#vBrownBag session at the Vancouver OpenStack Summit, May 2015
This document summarizes Dell's strategy to deliver a unified "Fluid Data" storage architecture through strategic acquisitions and integration of software and hardware. Key points include:
1) Dell has acquired several companies to expand its capabilities in file systems, data protection, deduplication, and memory virtualization.
2) Dell's converged infrastructure solutions integrate computing, networking, and storage blades in a single chassis for high density and performance.
3) Dell's AppAssure software provides data protection, replication, disaster recovery and universal recovery capabilities across virtual, physical and cloud environments for business continuity.
4) Dell aims to help customers optimize data, automate tiering, protect data, replicate and recover
The document describes the Quantum NDX-8 and NDX-8d, which provide NAS-based data protection and primary storage. The NDX-8d appliance offers backup and disaster recovery capabilities using deduplication to reduce data by 90% and retain 2+ years of data online. It provides fast restores, controls data growth, and less media management. The NDX-8 functions as primary storage with twice the processor/RAM of competitors and seamless integration with the NDX-8d for backup. Both the NDX-8 and NDX-8d are affordable starting around $3,700 and $4,700 respectively.
The Isilon X200 scale-out storage node uses a modular architecture and OneFS operating system to provide flexible scalability from a few terabytes to multiple petabytes within a single global namespace. It delivers high concurrent access and scalable capacity. SSD technology accelerates namespace intensive operations. The X200 is simple to deploy, manage, and scales performance and capacity on-the-fly in 60 seconds.
Currie Munce, VP of HGST, gave the talk at CAISS Annual Conference 2012, as part of the panel discussion: Storage Component Technologies - Enable Big Data and Make Better Cloud Computing.
Panzura & Scality - Cloud Storage made seamless - Cloud Expo New York City 2012Marc Villemade
This is the full presentation I did with Ranajit Nevatia from Panzura (@ranajitN) at Cloud Computing Expo NY in June 2012.
It introduces and explains the concepts of Structured and Unstructured data and why Object Storage will prevail when it comes to the latter.
AWS Partner Presentation – Panzura – AWS Cloud Storage for the Enterprise 2012Amazon Web Services
Panzura provides a global cloud storage system that makes cloud storage seamless. It allows storage of both structured and unstructured data in the cloud with fast local read/write performance. This provides enterprises with unlimited, low-cost storage while maintaining high performance access to data anywhere in the world.
The document discusses several use cases for Panzura storage:
1) Archiving large files like medical images, market data, and seismic data in the cloud for lower costs, global access, and centralized management.
2) Backing up structured data to the cloud for infinite scalability, no tapes, and fast restores.
3) Sharing files globally between offices for collaboration without performance issues over the WAN.
Leveraging Swift Storage Policies using Scality RINGNicolas Trangez
We present several use-cases for Swift Storage Policies, and how they are implemented in the Scality RING Swift back-end.
#vBrownBag session at the Vancouver OpenStack Summit, May 2015
This document summarizes Dell's strategy to deliver a unified "Fluid Data" storage architecture through strategic acquisitions and integration of software and hardware. Key points include:
1) Dell has acquired several companies to expand its capabilities in file systems, data protection, deduplication, and memory virtualization.
2) Dell's converged infrastructure solutions integrate computing, networking, and storage blades in a single chassis for high density and performance.
3) Dell's AppAssure software provides data protection, replication, disaster recovery and universal recovery capabilities across virtual, physical and cloud environments for business continuity.
4) Dell aims to help customers optimize data, automate tiering, protect data, replicate and recover
The document describes the Quantum NDX-8 and NDX-8d, which provide NAS-based data protection and primary storage. The NDX-8d appliance offers backup and disaster recovery capabilities using deduplication to reduce data by 90% and retain 2+ years of data online. It provides fast restores, controls data growth, and less media management. The NDX-8 functions as primary storage with twice the processor/RAM of competitors and seamless integration with the NDX-8d for backup. Both the NDX-8 and NDX-8d are affordable starting around $3,700 and $4,700 respectively.
The Isilon X200 scale-out storage node uses a modular architecture and OneFS operating system to provide flexible scalability from a few terabytes to multiple petabytes within a single global namespace. It delivers high concurrent access and scalable capacity. SSD technology accelerates namespace intensive operations. The X200 is simple to deploy, manage, and scales performance and capacity on-the-fly in 60 seconds.
Currie Munce, VP of HGST, gave the talk at CAISS Annual Conference 2012, as part of the panel discussion: Storage Component Technologies - Enable Big Data and Make Better Cloud Computing.
Tandberg Data's AccuVault is an all-in-one data protection appliance available in desktop and 1U configurations. It uses the company's AccuGuard Enterprise software to provide centralized, automated backup and disaster recovery for small to medium-sized networks. AccuVault's data deduplication capabilities reduce bandwidth usage and storage needs. It is well-suited to protect Windows servers, virtual servers, workstations and popular applications like Exchange and SQL.
Novell File Management Suite for Microsoft Active Directory EnvironmentsNovell
The new Novell File Management Suite offers tremendous benefits to organizations using Novell Open Enterprise Server or NetWare, but it doesn't stop there. The product is fully cross-platform, allowing pure Windows environments to tap into its capabilities too. Attend this session to see how Novell File Management Suite can be used to introduce advanced data management capabilities in almost any file system environment. You'll see live product demos and learn specific ways it can bring efficiencies to your Windows environment.
The document summarizes Quantum's new DXi8500 and Scalar i6000 offerings for enterprise data storage and management. The DXi8500 provides the fastest deduplication performance up to 8.8 TB/hr and largest single appliance capacity of 320TB. The Scalar i6000 features dual robotics and an Active Vault for large-scale data access. Quantum Vision 4.1 software provides enhanced management across these enterprise resources.
removing nodes. NuoDB automatically redistributes data and transactions
across available resources to maintain high performance and availability.
The document introduces NuoDB, a new cloud data management system that takes a completely new
approach to databases by rewriting the rules for relational databases to work in the cloud. NuoDB uses a
distributed, shared-nothing architecture that elastically scales to handle large transaction volumes and
global users. It provides the reliability of ACID transactions while scaling out simply on commodity cloud
resources to meet the demands of modern web-scale applications.
If you want to start managing your storage infrastructure intelligently, understanding your data is the place to start. Join this session to learn how you can do just that with the new Novell File Reporter, available through Novell File Management Suite. You'll learn how it can help you get a handle on file storage chaos by conducting file system inventories and generating usage reports. These reports can then be used to create and fine tune policies for the other suite components to implement.
Quantum RDX 8000 - The Speed Of Disk. The Protection of Tape. Quantum
The Quantum RDX 8000 provides the speed of disk storage with the protection and removability of tape. It is an 8 slot removable disk library that connects over iSCSI network and works with RDX cartridges from 160GB to 1TB. The RDX 8000 offers scalability, flexibility, and can be used for backup with software like Symantec Backup Exec or with deduplication via DATASTOR Shield to reduce storage needs. It provides an affordable solution starting at $3,500.
The document introduces Quantum Lattus, a wide area storage solution for managing large volumes of data across multiple sites. It uses fountain coding algorithms to distribute encoded data across object storage nodes, providing self-healing capabilities and 15 nines of durability. Lattus can scale from 500TB to hundreds of petabytes, offers both file system and REST access, and provides lower costs than other solutions through reduced maintenance needs, power/cooling, and high redundancy.
The document provides an overview of the Isilon scale-out storage platform. It summarizes that Isilon provides simple yet powerful storage solutions that allow businesses to manage their data instead of their storage. It can scale virtually unlimited while staying simple. Key features highlighted include accelerating processes, unlimited growth capabilities, high data protection, and low management needs.
The Scality RING is a software-defined storage platform that runs on standard x86 servers and is designed to scale linearly across multiple sites and thousands of servers to store petabytes of data. It provides a single, distributed storage system with unlimited storage capacity and high durability through data replication and erasure coding. The hardware-agnostic RING architecture ensures continuous availability during hardware upgrades and failures.
In this presentation from the DDN User Meeting at SC13, Jeff Denworth provides a product update,
Watch the video presentation: http://insidehpc.com/2013/11/13/ddn-user-meeting-coming-sc13-nov-18/
David Pechon is an IT professional with experience in virtualization, storage, and identity management. He has worked for the US Army, banks, and consulting companies. While RAID provides data redundancy, it does not replace the need for backups, as RAID cannot protect against accidental or intentional deletion of data. Storage technologies have evolved from drum memory and floppy disks to today's hybrid arrays, all-flash arrays, and converged infrastructure solutions. The differences between SANs and NAS depend on factors like latency requirements, data access patterns, distance between servers and storage, and administrative needs. Newer solutions aim to provide the benefits of both SAN and NAS through unified storage and software-defined storage.
The document summarizes a presentation on evolving a new analytical platform. It discusses defining the platform to include tools for the whole research cycle beyond just business intelligence (BI), with SQL Server 2008 R2 as an example of defining the platform. It also discusses what is working with existing platforms and what is still missing, including the need for more scalable data storage and processing.
The HiTiME project aims to develop a system that can recognize entities like people, organizations, locations, dates and professions in historical text documents. The system splits documents into words, recognizes entities using named entity recognition and stores the output in a database. It also aims to integrate with other systems at the International Institute of Social History to improve search, metadata and visualization of historical data. Some planned improvements include using additional natural language processing tools, disambiguating entities, recognizing composite entities, and integrating with applications like the Basic Word Sequence Analysis tool.
The Riverbed Whitewater appliance accelerates access to cloud storage by leveraging deduplication technology. It provides high-performance LAN access to cloud storage, enabling organizations to extend their storage to the cloud without compromising on performance or availability. Whitewater dramatically optimizes and accelerates backup and archiving to cloud storage, allowing unlimited scalability in storage infrastructures. It integrates seamlessly with existing storage management, data protection and archiving tools.
The document discusses a document management solution called Thinking Team that is affordable, easy to use, and addresses common information management problems. It offers flexible and pragmatic document storage, retrieval, and collaboration capabilities. Thinking Team has three pricing tiers and various optional extensions, and claims to put users back in control of information flow with convenience and low effort.
In this deck from the DDN User Group Meeting at SC14, Alex Bouzari presents: DDN Vision.
"Bouzari will discuss DDN’s vision of enabling organizations to maximize the value of their most valuable asset: Information, and how DDN has applied that vision, and a relentless focus on customer satisfaction, to quietly grow into the world’s largest privately held data storage company, named by Inc. magazine in 2010 and 2011 as one of the fastest-growing companies in America. By developing technology which solves the 21st century challenges of massive data creation and complex information analytics — what many are calling the “Big Data Era,” and leveraging an extensive network of go-to-market partners which includes IBM, HP, Dell, and Sony, DDN has successfully deployed thousands of systems in enterprises, universities and government agencies worldwide."
Watch the video presentation:
This document provides an overview of a presentation by Advanced Systems Group on top technology trends for virtualization. It discusses flash storage technologies, the importance of disaster recovery (DR), and architecting for the cloud. The presentation covers various flash storage options and their performance characteristics. It emphasizes the need for DR to address hardware failures, data corruption, and natural disasters. It also discusses best practices for virtualization including cluster sizing, resource allocation, and security considerations for virtual machines.
DDN GS7K - Easy-to-deploy, High Performance Scale-Out Parallel File System Ap...inside-BigData.com
In this deckt, Uday Mohan from DataDirect Networks presents: DDN GS7K - Easy-to-deploy, High Performance Scale-Out Parallel File System Appliance.
High performance computing is critical in commercial markets, spanning a wide range of applications across multiple industries, and this trend is only growing. The GS7K from DDN will help bring the latest high-performance storage technologies to more of these markets, connecting companies to their next innovations faster while satisfying their enterprise standards.”
Watch the video presentation: http://wp.me/p3RLHQ-d99
Drobo storage devices provide affordable storage solutions for small and medium-sized businesses, with features like thin provisioning, automated storage management, and scalability. Models include 8-bay and 12-bay devices for file sharing or iSCSI storage, with capacities up to 24TB. Drobo uses its innovative BeyondRAID technology to protect data like RAID without limitations, allowing mixed drive sizes, instant expansion, and dual disk redundancy.
Webinar: NAS vs. Object Storage: 10 Reasons Why Object Storage Will WinStorage Switzerland
This webinar discusses why object storage is better suited than network attached storage (NAS) for storing unstructured data. The presenters are from Storage Switzerland and Caringo. They will explain the challenges of rapidly growing unstructured data and how object storage addresses them through scalability, cost effectiveness, and continuous data integrity. Specifically, they cite 10 reasons why object storage will replace NAS in enterprises: 1) it ends refresh cycles by using a continuous refreshing architecture, 2) integrates cloud storage, 3) easily verifies data integrity, 4) replacement parts are redundant and easy to maintain, 5) has power/cooling advantages to operate more cost effectively, 6) is more cost effective to manage with one system instead of dozens
Tandberg Data's AccuVault is an all-in-one data protection appliance available in desktop and 1U configurations. It uses the company's AccuGuard Enterprise software to provide centralized, automated backup and disaster recovery for small to medium-sized networks. AccuVault's data deduplication capabilities reduce bandwidth usage and storage needs. It is well-suited to protect Windows servers, virtual servers, workstations and popular applications like Exchange and SQL.
Novell File Management Suite for Microsoft Active Directory EnvironmentsNovell
The new Novell File Management Suite offers tremendous benefits to organizations using Novell Open Enterprise Server or NetWare, but it doesn't stop there. The product is fully cross-platform, allowing pure Windows environments to tap into its capabilities too. Attend this session to see how Novell File Management Suite can be used to introduce advanced data management capabilities in almost any file system environment. You'll see live product demos and learn specific ways it can bring efficiencies to your Windows environment.
The document summarizes Quantum's new DXi8500 and Scalar i6000 offerings for enterprise data storage and management. The DXi8500 provides the fastest deduplication performance up to 8.8 TB/hr and largest single appliance capacity of 320TB. The Scalar i6000 features dual robotics and an Active Vault for large-scale data access. Quantum Vision 4.1 software provides enhanced management across these enterprise resources.
removing nodes. NuoDB automatically redistributes data and transactions
across available resources to maintain high performance and availability.
The document introduces NuoDB, a new cloud data management system that takes a completely new
approach to databases by rewriting the rules for relational databases to work in the cloud. NuoDB uses a
distributed, shared-nothing architecture that elastically scales to handle large transaction volumes and
global users. It provides the reliability of ACID transactions while scaling out simply on commodity cloud
resources to meet the demands of modern web-scale applications.
If you want to start managing your storage infrastructure intelligently, understanding your data is the place to start. Join this session to learn how you can do just that with the new Novell File Reporter, available through Novell File Management Suite. You'll learn how it can help you get a handle on file storage chaos by conducting file system inventories and generating usage reports. These reports can then be used to create and fine tune policies for the other suite components to implement.
Quantum RDX 8000 - The Speed Of Disk. The Protection of Tape. Quantum
The Quantum RDX 8000 provides the speed of disk storage with the protection and removability of tape. It is an 8 slot removable disk library that connects over iSCSI network and works with RDX cartridges from 160GB to 1TB. The RDX 8000 offers scalability, flexibility, and can be used for backup with software like Symantec Backup Exec or with deduplication via DATASTOR Shield to reduce storage needs. It provides an affordable solution starting at $3,500.
The document introduces Quantum Lattus, a wide area storage solution for managing large volumes of data across multiple sites. It uses fountain coding algorithms to distribute encoded data across object storage nodes, providing self-healing capabilities and 15 nines of durability. Lattus can scale from 500TB to hundreds of petabytes, offers both file system and REST access, and provides lower costs than other solutions through reduced maintenance needs, power/cooling, and high redundancy.
The document provides an overview of the Isilon scale-out storage platform. It summarizes that Isilon provides simple yet powerful storage solutions that allow businesses to manage their data instead of their storage. It can scale virtually unlimited while staying simple. Key features highlighted include accelerating processes, unlimited growth capabilities, high data protection, and low management needs.
The Scality RING is a software-defined storage platform that runs on standard x86 servers and is designed to scale linearly across multiple sites and thousands of servers to store petabytes of data. It provides a single, distributed storage system with unlimited storage capacity and high durability through data replication and erasure coding. The hardware-agnostic RING architecture ensures continuous availability during hardware upgrades and failures.
In this presentation from the DDN User Meeting at SC13, Jeff Denworth provides a product update,
Watch the video presentation: http://insidehpc.com/2013/11/13/ddn-user-meeting-coming-sc13-nov-18/
David Pechon is an IT professional with experience in virtualization, storage, and identity management. He has worked for the US Army, banks, and consulting companies. While RAID provides data redundancy, it does not replace the need for backups, as RAID cannot protect against accidental or intentional deletion of data. Storage technologies have evolved from drum memory and floppy disks to today's hybrid arrays, all-flash arrays, and converged infrastructure solutions. The differences between SANs and NAS depend on factors like latency requirements, data access patterns, distance between servers and storage, and administrative needs. Newer solutions aim to provide the benefits of both SAN and NAS through unified storage and software-defined storage.
The document summarizes a presentation on evolving a new analytical platform. It discusses defining the platform to include tools for the whole research cycle beyond just business intelligence (BI), with SQL Server 2008 R2 as an example of defining the platform. It also discusses what is working with existing platforms and what is still missing, including the need for more scalable data storage and processing.
The HiTiME project aims to develop a system that can recognize entities like people, organizations, locations, dates and professions in historical text documents. The system splits documents into words, recognizes entities using named entity recognition and stores the output in a database. It also aims to integrate with other systems at the International Institute of Social History to improve search, metadata and visualization of historical data. Some planned improvements include using additional natural language processing tools, disambiguating entities, recognizing composite entities, and integrating with applications like the Basic Word Sequence Analysis tool.
The Riverbed Whitewater appliance accelerates access to cloud storage by leveraging deduplication technology. It provides high-performance LAN access to cloud storage, enabling organizations to extend their storage to the cloud without compromising on performance or availability. Whitewater dramatically optimizes and accelerates backup and archiving to cloud storage, allowing unlimited scalability in storage infrastructures. It integrates seamlessly with existing storage management, data protection and archiving tools.
The document discusses a document management solution called Thinking Team that is affordable, easy to use, and addresses common information management problems. It offers flexible and pragmatic document storage, retrieval, and collaboration capabilities. Thinking Team has three pricing tiers and various optional extensions, and claims to put users back in control of information flow with convenience and low effort.
In this deck from the DDN User Group Meeting at SC14, Alex Bouzari presents: DDN Vision.
"Bouzari will discuss DDN’s vision of enabling organizations to maximize the value of their most valuable asset: Information, and how DDN has applied that vision, and a relentless focus on customer satisfaction, to quietly grow into the world’s largest privately held data storage company, named by Inc. magazine in 2010 and 2011 as one of the fastest-growing companies in America. By developing technology which solves the 21st century challenges of massive data creation and complex information analytics — what many are calling the “Big Data Era,” and leveraging an extensive network of go-to-market partners which includes IBM, HP, Dell, and Sony, DDN has successfully deployed thousands of systems in enterprises, universities and government agencies worldwide."
Watch the video presentation:
This document provides an overview of a presentation by Advanced Systems Group on top technology trends for virtualization. It discusses flash storage technologies, the importance of disaster recovery (DR), and architecting for the cloud. The presentation covers various flash storage options and their performance characteristics. It emphasizes the need for DR to address hardware failures, data corruption, and natural disasters. It also discusses best practices for virtualization including cluster sizing, resource allocation, and security considerations for virtual machines.
DDN GS7K - Easy-to-deploy, High Performance Scale-Out Parallel File System Ap...inside-BigData.com
In this deckt, Uday Mohan from DataDirect Networks presents: DDN GS7K - Easy-to-deploy, High Performance Scale-Out Parallel File System Appliance.
High performance computing is critical in commercial markets, spanning a wide range of applications across multiple industries, and this trend is only growing. The GS7K from DDN will help bring the latest high-performance storage technologies to more of these markets, connecting companies to their next innovations faster while satisfying their enterprise standards.”
Watch the video presentation: http://wp.me/p3RLHQ-d99
Drobo storage devices provide affordable storage solutions for small and medium-sized businesses, with features like thin provisioning, automated storage management, and scalability. Models include 8-bay and 12-bay devices for file sharing or iSCSI storage, with capacities up to 24TB. Drobo uses its innovative BeyondRAID technology to protect data like RAID without limitations, allowing mixed drive sizes, instant expansion, and dual disk redundancy.
Webinar: NAS vs. Object Storage: 10 Reasons Why Object Storage Will WinStorage Switzerland
This webinar discusses why object storage is better suited than network attached storage (NAS) for storing unstructured data. The presenters are from Storage Switzerland and Caringo. They will explain the challenges of rapidly growing unstructured data and how object storage addresses them through scalability, cost effectiveness, and continuous data integrity. Specifically, they cite 10 reasons why object storage will replace NAS in enterprises: 1) it ends refresh cycles by using a continuous refreshing architecture, 2) integrates cloud storage, 3) easily verifies data integrity, 4) replacement parts are redundant and easy to maintain, 5) has power/cooling advantages to operate more cost effectively, 6) is more cost effective to manage with one system instead of dozens
The document discusses the history and development of artificial intelligence over several decades. It outlines milestones in AI such as the creation of logic theories, machine learning algorithms, and neural networks. Recent advances in deep learning now allow AI to perform complex tasks like object recognition.
Object Storage promises many things - unlimited scalability, both in terms of capacity and file count, low cost but highly redundant capacity and excellent connectivity to legacy NAS. But, despite these promises object storage has not caught on in the enterprise like it has in the cloud. It seems like, for the enterprise object storage just isn’t a good fit. The problem is that most object storage system’s starting capacity is too large. And while connectivity to legacy NAS systems is available, seamless integration is not. Can object storage be sized so that it is a better fit for the enterprise?
Primend Pilveseminar - Soodne hind + lihtne haldus – pilve minek= ?Primend
Kuidas saada oma andmekeskusesse rohkem pilvele omaseid funktsioone, kui pilve minek pole võimalik? Kuidas saavutada 90% kokkuhoidu andmehoidla ja varukoopia mahult? Kuidas taastada 1 TB mahuga varukoopia vähem kui minutiga? Koostöös Cisco UCS Director automatiseerimise ja juhtimisega pakub SimpliVIty avalikule pilvele omast paindlikkust ja madalat halduskulu.
S3 Server, a Scality product, was born after a hackathon in Paris, France in 2015. What better way to continue with our philosophy of innovation than to host a hackathon of our own?
On October 21st, coders joined us for a weekend of coding, developing new solutions for storage, integrations for S3 and much more!
This event was sponsored by Seagate and hosted at Holberton School.
Even the best system administrator cannot always avoids any and
every disaster that may plague his data center, but he should have a
contingency plan to recover from one - and an administrator that
manages his virtual data centers with oVirt is of course no different.
This session will cover the new features introduced in oVirt 3.5.0 to
handle such scenarios and will showcase how stringing together a set
of building blocks can produce a well rounded solution for disaster
scenarios.
This document summarizes BlueStore, a new storage backend for Ceph that provides faster performance compared to the existing FileStore backend. BlueStore manages metadata and data separately, with metadata stored in a key-value database (RocksDB) and data written directly to block devices. This avoids issues with POSIX filesystem transactions and enables more efficient features like checksumming, compression, and cloning. BlueStore addresses consistency and performance problems that arose with previous approaches like FileStore and NewStore.
This document summarizes a presentation about tuning MySQL performance on Ceph block storage. The presentation covers Ceph architecture, tuning Ceph block devices, and tuning QEMU block virtualization. It then shows benchmarks comparing different configurations for reads, writes, and a 70/30 read/write mix using Sysbench OLTP workloads. Configurations tested include QEMU backends, caching modes, I/O threading, virtio types and queues, and containers versus metal. The benchmarks demonstrate up to 60,000 IOPS can be achieved.
This document provides an overview of storage best practices in oVirt, including oVirt storage domains, manual tiering across different storage types, volume types and allocation policies, and single disk snapshots. It discusses using different storage domains like NFS, iSCSI, Fibre Channel for manual tiering to choose the best storage. It also covers volume types, allocation policies of preallocated vs thin provisioning, and using QCOW2 format for snapshots. Finally, it describes how oVirt implements single disk snapshots using logical volume manager (LVM).
AWS re:Invent 2016 - Scality's Open Source AWS S3 ServerScality
Presented by Giorgio Regni, CTO
Try Scality S3 Server Today!
https://s3.scality.com/
http://www.scality.com/scality-s3-server/
https://hub.docker.com/r/scality/s3server/
Ian Colle, the Ceph Program Manager at Inktank, gave a presentation on Ceph and how it can be used for cloud storage with OpenStack. Ceph is an open source software-defined storage system that provides object, block, and file storage in a single distributed system. It utilizes a CRUSH algorithm for data distribution, thin provisioning for efficient storage of VMs, and is integrated with OpenStack through the Cinder block storage and Swift object storage APIs. Inktank was formed to ensure the long-term success of Ceph through services, support, and helping companies adopt it.
The document discusses a webinar comparing the total cost of ownership (TCO) of hyperconverged infrastructure and public cloud solutions. It presents a TCO model comparing a SimpliVity hyperconverged cluster to an equivalent Amazon EC2 cloud configuration. Over 3 years, the analysis found that the SimpliVity solution had a TCO 22-49% lower than the AWS configuration. Key factors that make hyperconverged infrastructure more cost-effective include lower upfront capital costs through converged hardware, simplified management and expansion, and efficiency features. While public cloud remains viable for some workloads, the economics are no longer a given compared to hyperconverged appliances according to the analysis.
This presentation provides an overview of the Dell PowerEdge R730xd server performance results with Red Hat Ceph Storage. It covers the advantages of using Red Hat Ceph Storage on Dell servers with their proven hardware components that provide high scalability, enhanced ROI cost benefits, and support of unstructured data.
This document discusses how Mellanox networks enable high performance Ceph storage clusters. It notes that Ceph performance and scalability are dictated by the backend cluster network performance. It provides examples of customers deploying Ceph with Mellanox 40GbE and 10GbE interconnects, and highlights how these networks allow building scalable, high performing storage solutions. Specifically, it shows how 40GbE cluster networks and 40GbE client networks provide much higher throughput and IOPS compared to 10GbE. The document concludes by mentioning how RDMA offloads can free CPU for application processing, and how the Accelio library enables high performance RDMA for Ceph.
NoSQL – Back to the Future or Yet Another DB Feature?Martin Scholl
The document argues that NoSQL technology takes a step back from relational databases by complicating data integration and quality assurance, and that NoSQL systems will eventually just become another feature of database management systems and cloud computing services, with future "PostSQL" databases being indistinguishable from general data communication services.
The document discusses the ongoing revolution in database technology driven by factors like increasing data volumes, new workloads, and market forces. It provides a history of databases from the pre-relational era to today's relational and post-relational databases. The discussion covers topics around challenges with existing database concepts, the impedance mismatch between databases and applications, and different types of NoSQL databases and database workloads.
Slides from the Live Webcast on Jan. 18, 2012
The purpose of this event is to allow the Analysts, Robin Bloor and Mark Madsen, to offer their theories on where the database market stands today: What’s new? What’s standard? What is the trajectory of this changing market? Each Analyst will present for 10-15 minutes, then will engage in a dialogue with Host Eric Kavanagh and all attendees.
For more information visit: http://www.databaserevolution.com
Watch this and the entire series at : http://www.youtube.com/playlist?list=PLE1A2D56295866394
This document discusses the challenges of big data and different approaches to structuring data systems. It notes that most enterprise databases are still relatively small, less than 1TB, but that 80% of business information is now unstructured. Big data represents a new class of economic asset. The document examines different approaches taken by companies like Google and eBay and questions whether one is clearly better. It argues that being able to model data is a more significant challenge than raw data size. Making data globally accessible within an organization in a unified way could help make the organization smarter by improving insights.
Beyond The Data Grid: Coherence, Normalisation, Joins and Linear ScalabilityBen Stopford
In 2009 RBS set out to build a single store of trade and risk data that all applications in the bank could use. This talk discusses a number of novel techniques that were developed as part of this work. Based on Oracle Coherence the ODC departs from the trend set by most caching solutions by holding its data in a normalised form making it both memory efficient and easy to change. However it does this in a novel way that supports most arbitrary queries without the usual problems associated with distributed joins. We'll be discussing these patterns as well as others that allow linear scalability, fault tolerance and millisecond latencies.
Where Does Big Data Meet Big Database - QCon 2012Ben Stopford
The document discusses the evolution of big data technologies and databases. It describes how early big data technologies like MapReduce took a simpler approach compared to relational databases. This led to a disruption in the database market as NoSQL systems gained popularity. However, relational databases have also advanced by leveraging new hardware and dropping some traditional constraints. Today, the technologies have converged and many vendors offer integrated suites combining relational and big data approaches. The best solution depends on the specific problem and data characteristics rather than just data size.
Big data management requires large data centers to store the vast amounts of data being generated. Hadoop is a technology used to manage big data across clusters of commodity servers in data centers. It uses a distributed file system and is designed to parallelize data processing. Data centers are facilities that house computer systems and associated components, such as telecommunications and storage systems, that are needed to manage big data. They contain rows of server racks and must have sufficient power and cooling to handle large amounts of structured and unstructured data from sources like the New York Stock Exchange, Google, and Facebook. As data volumes continue increasing rapidly, more advanced technologies will be needed to manage big data in the future.
Debunking "Purpose-Built Data Systems:": Enter the Universal DatabaseStavros Papadopoulos
Purpose-built databases and platforms have actually created more complexity, effort, and unnecessary reinvention. The status quo is a big mess. TileDB took the opposite approach.
In this presentation, Stavros, the original creator of TileDB, shared the underlying principles of the TileDB universal database built on multi-dimensional arrays, making the case for it as a true first in the data management industry.
DAMA, Oregon Chapter, 2012 presentation - an introduction to Data Vault modeling. I will be covering parts of the methodology, comparison and contrast of issues in general for the EDW space. Followed by a brief technical introduction of the Data Vault modeling method.
After the presentation i I will be providing a demonstration of the ETL loading layers, LIVE!
You can find more on-line training at: http://LearnDataVault.com/training
Accelerating big data with ioMemory and Cisco UCS and NOSQLSumeet Bansal
This document discusses accelerating big data solutions using Oracle NoSQL Database, Cisco UCS, and Fusion-io storage. It provides an overview of big data characteristics and business benefits. It then summarizes Oracle NoSQL Database architecture and how its collaboration with Cisco UCS and Fusion-io storage provides optimized, high-performance solutions. The document outlines performance testing using YCSB that showed the solution can achieve over 1 million operations per second. It also discusses how Fusion-io's software-defined storage architecture provides extreme throughput and low latency.
Evolution of the DBA to Data Platform Administrator/SpecialistTony Rogerson
DBA's used to be Relational Database centric for instance managing Microsoft SQL Server or Oracle, in this changing world of polyglot database environments their role has expanded not just into new platforms other than SQL but also new legal governance, modelling techniques, architecture etc. They need to have a base knowledge of Kimball, Inmon, Data Vault, what CAP theorem is, LAMBDA, Big Data, Data Science etc.
The document discusses how SQL and NoSQL databases can work together for big data. It provides an overview of relational databases based on Codd's rules and how NoSQL databases are used for less structured data like documents and graphs. Examples of using MongoDB and Hadoop are provided. The document also discusses using MySQL with memcached to get the benefits of both SQL and NoSQL for accessing data.
Big Data Basic Concepts | Presented in 2014Kenneth Igiri
This document provides an overview of big data concepts and technologies. It discusses the 3 Vs, 4 Vs and 6 Vs frameworks used to describe big data. Key big data technologies mentioned include MapReduce, Hadoop, HDFS, YARN, and NoSQL databases like MongoDB, Cassandra, HBase and Dynamo. The Lambda architecture and CAP theorem concepts are also covered. Large internet companies like Google, Amazon, eBay are discussed as examples of organizations that have pioneered big data solutions to handle massive volumes of dynamic data at high velocity.
Data, information, and knowledge are related but distinct concepts. A database is a collection of structured data that allows for easier access, storage, and retrieval of data. There are many types of databases including relational, NoSQL, hierarchical, distributed, and cloud databases. Each type has advantages like security, scalability, and accessibility as well as disadvantages like complexity, costs, and lack of standardization. Relational databases use tables with rows and columns while NoSQL databases store data in more flexible ways.
Is Software Defined Storage (SDS) getting hijacked? It seems every vendor, old and new, is claiming that their storage is “software defined”. The original intent was to create software-only solutions that could be deployed on the customer’s choice of servers. But that original intent has evolved, and now hardware vendors are providing what they claim to be software defined storage solutions too. In addition, SDS is being combined with an embedded compute function to create hyper-converged solutions as well.
In this webinar we will discuss the differences in these approaches and you will learn what the four key deliverables of a SDS solution should be so you can decide which makes the most sense for your organization.
Our own Sean Doherty was in Madrid this week, presenting at the Red Hat Partner summit on the rise of big data and what it means for the future of the RDBMS in the enterprise. Check out his presentation!
This deck talks about the basic overview of NoSQL technologies, implementation vendors/products, case studies, and some of the core implementation algorithms. The presentation also describes a quick overview of "Polyglot Persistency", "NewSQL" like emerging trends.
The deck is targeted to beginners who wants to get an overview of NoSQL databases.
This document provides an overview of database systems and Oracle database. It defines what a database is, the different types of databases, and why database design is important. It discusses data redundancy, data anomalies, and data normalization. It also describes the functions of a database management system (DBMS), advantages and disadvantages of DBMS, and the different editions of Oracle database.
A Paradigm Shift: The Increasing Dominance of Memory-Oriented Solutions for H...Ben Stopford
This document discusses the increasing dominance of memory-oriented solutions for high-performance data access. It notes that database lookups are around 20 milliseconds while hashmap lookups are around 20 nanoseconds. It then discusses how abstraction improves software but hurts performance. It outlines the traditional database architecture with disk storage and compares it to newer in-memory and distributed in-memory architectures that can provide faster performance by avoiding disk I/O and leveraging memory and distribution.
Similar to Scality presentation cloud Computing Expo NY 2012 v1.0 (20)
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?
Scality presentation cloud Computing Expo NY 2012 v1.0
1. Cloud Storage Made Seamless
Marc Villemade Ranajit Nevatia
Technology Evangelist VP, Marketing
Scality Panzura
Slide 1
2. There are two types of data
(roughly)
Structured
We (sort of) know how to manage this
Unstructured
This is the new beast we have issues with
Slide 2
3. How to define Structured Data?
Structured data is a set of organized pieces of data
Relational databases are a perfect example
Atomic pieces are, on their own, meaningless
Slide 3
4. What about Unstructured Data?
Unstructured data is self-contained pieces of
data
Self-descriptive
Meaningful in and of itself
Typically has metadata attached to it
Email, Videos, Presentations, Spreadsheets, sate
llite images…
An easy way to think about it is anything that can
be stored in one file is unstructured data
Slide 4
5. Some numbers…
In 2012, Humanity will generate 2.7 ZB of data 1
It is estimated that we permanently store ~ 1 ZB of it 2 (~40%)
80% of it is unstructured 1
500 Quadrillion files (500,000 million million files)
Next year and so on, it will grow by 50% y-o-y 1
It will double every 2 years in the next 10 years
Kind of unfathomable, ain’t it?
(1) IDC numbers – (2) University of Southern California (2007)
Slide 5
6. Humans like organized things
Well, some of them at least…
Structured storage systems have been used for Unstructured Data
Organized in file systems, hierarchies, directories
Easier for us
And then new data creation patterns emerged early 2000s
The model doesn’t fit anymore
And here’s why
Slide 6
7. Typical SAN / NAS issues at Scale
Technology refresh and migration necessary to
benefit from larger disks
Scheduled maintenance window nuisance
Limitations on # of files
Volume management is complex
Serial architecture compromises performance
RAID is less efficient for large drives
FC networks are expensive & point-to-point
Cost is prohibitive for large capacity
Slide 7
8. Humans like organized things
Well, some of them at least…
Structured data storage systems are used for Unstructured.
Organized in file systems, hierarchies, directories
Easier for us
And then new data creation patterns emerged early 2000s
The model doesn’t fit anymore
SANs and NASes were not made to handle this
Slide 8
9. So what’s the solution?
We believe it’s Object Storage
Yahoo!, Amazon, Google.. were the pioneers
Main Characteristics
Flat Namespace
Infinite Scalability
Elasticity
Cost-efficiency
Data availability and durability
Slide 9
10. Scality’s Storage Vision
Their DC Their DC YOUR DC
Their App. YOUR App. YOUR App.
YOUR Data YOUR Data YOUR Data
Slide 10
11. What is the Secret Sauce?
Scality has developed a distributed (scale–out) object-based
storage software to turn x86 servers into Petabyte scale
storage for unstructured data (files).
(Scality is NOT designed for VM, VDI, Relational Database)
• Distributed System
• Distributed metadata
• No Single point of failure
• Self healing
• Organic upgrades
Slide 11
12. What’s unique about Scality RING
• Performance
• ESG Lab report: we’re 10x faster than any other object store
• Hardware-agnostic
• Software Vendor
• Mixed hardware (disks, nodes)
• Erasure-Coding with No penalty on read
• With only 60% overhead
• Tiering
• Policy driven
• Automatic, Transparent
Slide 12