Empresas de todos os tamanhos têm desafios constantes de armazenamento, crescimento e proteção de dados. A aquisição de mais armazenamento prolonga o desafio de gerenciar seu ciclo de vida, que inclui compra, operação contínua, falhas de hardware, atualizações e migrações. Nesta sessão, aprenda como usar o AWS Storage Gateway para conectar suas aplicações on-premises aos serviços de armazenamento da AWS usando protocolos de armazenamento padrão. O Storage Gateway permite soluções de armazenamento em nuvem híbrida para compartilhamento de arquivos, data lakes, análise de big data, backup e recuperação de desastres e migração. Discutiremos as melhores práticas e novas abordagens de implementação.
Palestrante: Melissa Ravanini
A apresentação dela durou 1h
O primeiro cliente falou 20 minutos
Storage is more than just the protocol or interface. It’s the lifeblood of application design and renewed architectures. Our customers have taught us that they need two things: scale and trust. 1. Make sure I can grow. 2. Make sure I can access what I need when I need it, (and of course help me keep costs down).
The suite of transfer services that support customers in their migrations means more choice. Large batches, incremental changes, constant streams or seamless integration are all part of the storage offering. Today we’re going to talk about two of the newest ways to do cloud data migration, Snowball and S3 Transfer Acceleration.
Note to presenters: Disk Transfer service is not EOL but has been deprecated out of the transfer services story in favor of Snowball. Snowball has already surpassed the amount of data imported over the lifetime of the disk transfer service.
EFS is in preview and due before the end of the year
Using Amazon EFS with Microsoft Windows Amazon EC2 instances is not supported. Source: http://docs.aws.amazon.com/efs/latest/ug/mounting-fs.html
So at this point, I got you thinking that S3 is great, but how much does it cost. Let’s first discuss what we’re going to charge you for… with traditional storage you pay for raw capacity but after accounting for protection schemes, such as RAID, file system overhead and the need to keep a free storage reserve, you’re left with much less of actual capacity used by data. With S3 you only pay for used capacity, when you use it. So in this example for 400 TBs, you’re really paying for 400 TBs and this is not accounting for DR copies. This drastic difference affects both CAPEX and OPEX costs.
Cost saving vs. engineering savings
Example:
Move object to Amazon Glacier after 30 days
Delete object after 365 days
For more information, see: http://docs.aws.amazon.com/AmazonS3/latest/dev/object-lifecycle-mgmt.html
Nós temos três tipos de família dentro da solução
Arquivo: que oferece uma interface NFS; que permite que você use suas soluções existentes que se baseiam em NFS mas armazenando dados no S3, se aproveitando de todos os seus benefícios, como durabilidade e alta disponibilidade.
Volume expõe interfaces iScsi que permite que você escreva em volumes, e então fazer Snapshots destes volumes na AWS.
O Fita provê uma interface de fita virtual (VTL), e permite que você use suas ferramentas de backup para levar suas fitas para a AWS, e arquivá-las na nossa solução de arquivamento, o Amazon Glacier.
E todas estas famílias são integradas com outros serviços AWS, para monitoramento, segurança, criptografia e controle.
O serviço provê interface NFS v3 e 4, ou seja, funciona com Windows e Linux, iScsi e iScsi VTL.
O serviço é implantando como uma máquina virtual, que pode ser VMWare, Hyper-V ou na própria AWS como um EC2.
(Mentira [Indepentende de qual modelo você esteja usando]), o serviço provê três funcionalidades principais:
Conversão de protocol, NFS ou iScsi, para o protocolo dos serviços de storage da AWS. Ele também provê emulação de dispositivo de fita virtual
Ele prove cache. Para muitas aplicações, é preciso prover baixa latência de acesso aos dados, que é provido pelo cache local controlado pelo serviço
Tráfego de dados otimizado, via “firewall friendly HTTPS” da VM sendo transferido de forma segura e criptografada para a AWS -> Nós investimos muito em tecnologias como multi-thread, multi-part upload, subir só os dados que mudaram, e também low-latency pre-fetching para otimizar as transferências pelo SGW.
S3 armazena objetos de forma imutável. Então você pode até habilitar versionamento, mas put é um objeto diferente, você não consegue editá-lo.
Então vamos supor que você tenha um objeto com ABC dentro dele.
Então este objeto é armazenado no S3
Vamos supor que você edite este arquivo localmente colocando o D nele.
O que a gente faz é fazer o upload apenas do D, e fazemos referência aos dados que você já tem lá para então criarmos um novo objeto.
In the fullness of time we see hybrid cloud storage addressing needs at the edge of your networks. Customers asked for a way to incorporate simple detached cloud storage platform with some computing capability at the edge of their networks, for applications like wind farms, medical devices, shipboard scientific computing and manufacturing shop floors.
AWS Snowball Edge is a petabyte-scale data transfer solution with temporary on-premises storage and compute capabilities. It transports up to 100TB of data with the same embedded cryptography and security as the original Snowball, and may also integrate smoothly with existing workflows, scale local capacity, and process stored data. Snowball Edge hosts a file server and an S3-compatible endpoint that allow you to use the NFS protocol, S3 SDK or S3 CLI to transfer data directly to the device without specialized client software. Multiple units may be clustered together, forming a temporary data collection storage tier in your datacenter so you can work as data is generated without managing copies. As storage needs scale up and down, devices can be easily added or removed from the local cluster and returned to AWS.
Snowball Edge also comes with embedded computing power (equivalent to an EC2 m4.4 xlarge instance) that hosts a platform for general compute tasks. AWS Lambda functions can run on the device to do things like examine a data stream collected from an IoT sensor, search for anomalies, create aggregated metrics or send alarms or control signals. Environments with unstable connectivity but high operational demands can run data processes redundantly on Snowball Edge devices, protecting against connectivity issues and eventually returning the captured and processed results to AWS.
Snowball Edge is designed to keep data and applications secure while on site or in transit to AWS, making it appropriate for even the most sensitive customer data. The hardware and software is cryptographically signed and all data stored is automatically encrypted using 256 bit encryption keys, owned by the customer and managed by AWS Key Management Service (KMS). Customer data stays encrypted in the appliance and is decrypted only at the time when it is copied from the appliance to AWS. Encryption is now performed on the device, instead of on the client, producing higher data throughput rates and reducing overall processing time.
Snowball Edge devices are Amazon-owned and eliminate the need for customers to invest in new hardware. Customers pay $300 plus shipping per device and a $30 per day usage fee, applied after the initial 10 days on site. If more capacity is needed at the edge, multiple devices can be requested and used together in a cluster. Amazon monitors the health and utilization of Snowballs and provides replacement devices when needed. Current Snowball data transport appliances in 50TB and 80TB volumes will continue to be available in addition to the new Snowball Edge. Availability in regions will vary, please check the Snowball product page for additional information.
Philips Healthcare develops technology solutions for consumers, patients, providers and caregivers across the health continuum, from supporting healthy living and prevention to diagnosis, treatment and home care. They embedded Snowball v2 devices in their hospital networks to collect data and initiate real-time analytics. Now the hospital staff no longer waits for answers and they have a local dataset to run on in case of any connectivity issues.
Essess provides a software-as-a-service (SaaS) platform that collects and analyzes the energy-efficiency performance of buildings via thermal-imaging technology. The company deploys vehicles equipped with thermal-imaging systems that create heat maps of thousands of buildings and electric grid assets every hour. The imaging vehicles detect leaks in windows, doors, walls, and foundations to help building owners prevent energy loss. Essess works with North America’s largest electric and gas utilities, building material providers, government agencies, and energy-service companies.
“We had to find a way to easily and cost-effectively store thermal-imaging data from its vehicles’ imaging systems. “Each vehicle, deployed across large geographic areas, collects more than a petabyte of data each year,”
The company also needed to be able to scale its application quickly. “We sometimes need to scale up to run several hundred instances at once, depending on the project and the customer,”
“Previously, we had to invest in and manage up to 40 hard drives in each of our vehicles. That wasn’t optimal. The hard drives often got damaged as we shipped them back and forth, and their lifespan was short. Snowball makes the process more reliable and simple. Our remote teams simply drop the appliance in the mail once it’s full, and we get the data into Amazon S3 within a few days.”
Snowball, S3,
Located in the Blue Ridge and Great Smoky mountains in North Carolina, the City of Asheville attracts both tourists and businesses. Recent disasters like Hurricane Sandy led the city’s IT department to search for an offsite disaster recovery (DR) solution. Working with AWS partner CloudVelox, the city used AWS to build an agile disaster recovery solution without the time and cost of investing in an on-premises data center.