Your SlideShare is downloading. ×
Vss 101 and design considerations in v mware environment
Upcoming SlideShare
Loading in...5
×

Thanks for flagging this SlideShare!

Oops! An error has occurred.

×
Saving this for later? Get the SlideShare app to save on your phone or tablet. Read anywhere, anytime – even offline.
Text the download link to your phone
Standard text messaging rates apply

Vss 101 and design considerations in v mware environment

3,840
views

Published on

Published in: Design, Technology, Education

0 Comments
0 Likes
Statistics
Notes
  • Be the first to comment

  • Be the first to like this

No Downloads
Views
Total Views
3,840
On Slideshare
0
From Embeds
0
Number of Embeds
0
Actions
Shares
0
Downloads
14
Comments
0
Likes
0
Embeds 0
No embeds

Report content
Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
No notes for slide

Transcript

  • 1. VSS 101 and Design Considerations in VMware environment (Part I) Posted on November 2, 2012 by wenjyu We have gotten lots of customer inquiries in learning more about VMDK vs. RDM vs. in-guest attached iSCSI storage, and its implication on data protection with Microsoft VSS framework/integration. Instead of telling you what to do directly, I decided to work with our engineering gurus (Jay Wang, AnaghaBarve, SathyaBhat and Scott Moreland) who think about this day in and day out, to do a joint two part post on how VSS works, follow by design considerations. VSS 101 Microsoft Volume Shadow Service (VSS) framework provides application consistent ―shadow‖ copy of volumes hosting application data. There are three major components in the VSS framework: Requestor (any backup app that requests the service from VSS framework to create consistent shadow copy of the application volume(s)) Provider (manages application data volume, responds to requests from requestor to make shadow copy of the managed volume) Writer (applications that are VSS-framework aware, for example, Exchange, SQL, Sharepoint)
  • 2. When it comes to ―Provider‖, there are three main types: 1. Software-based Provider ->shadow copy of the volume is created in software, at a layer above NTFS filesystem 2. System-based Provider -> like #1 above, the provider is supplied by the Operating System itself. System-based provider typically creates a copy-on-write shadow copy, and does not leverage the capabilities of the underlying hardware storage device. Third party backup vendors typically supplied their own software based Provider that is optimized for their own backup application or storage hardware. 3. Hardware-based Provider -> the work of creating the shadow copy is performed by the storage controller (array)
  • 3. So how does it work? Below is an over-simplified diagram along with explanation: 1. Requestor asks VSS to create a shadow copy of a specific volume 2. VSS instructs the Writer to prepare data for shadow copy (i.e., completing all pending transactions). When finished, the writer will inform VSS that it‘s done prepping the volume for an application consistent shadow copy 3. VSS, upon confirmation that all pending transactions have been completed, instructs Writer to hold all new write requests (reads could still be served), for up to 60 seconds 4. Upon acknowledgement from Writer that the application has been quiesced, all buffer cache is flushed to disk from NTFS filesystem 5. VSS now directs Provider to create a shadow copy of the volume (it has 10 seconds to finish the job) 6. Provider creates a shadow copy of the volume 7. After Provider informs VSS that the shadow copy creation is done, VSS informs the writer to ‗un-quiesce‘ the new writes(NOT shown in diagram above); lastly, VSS will inform the requestor that the shadow copy request is completed NOTE the above steps are greatly simplified – keep in mind that VSS does check back with the Writer to make sure step 4 is completed successfully, meaning the new writes are properly quiesced/paused. If not, it simply simply fails the operation and falls back to step 1 again. Now let‘s dive a bit more in the CLI ―vssadmin‖ In the Windows Command Prompt, you can use ―vssadmin‖ command to find out some native info about the system providers, writers as well as volumes available for shadow copy vssadmin list providers (this will only list the system provider)
  • 4. What about the software providers then? Typically, they are registered as a service that runs in the OS. For example, VMware Tools provides a VSS requestor and provider, the provider service can be found in the Services: This service is stopped by default – it will only get started when the tools VSS requestor is attempting to make a shadow copy call to VSS. If/when VMware snapshot is invoked, the service will start, and you will notice the ―Volume Shadow Copy‖ service getting stopped. That is to be expected, as we are not leveraging the System Provider to do the job for us. vssadmin list volumes (this command returns the list of volumes that are available for app queisced shadow copy)
  • 5. Above is the output from my SQL 2008 VM with three VMDKs, C: for OS, E: for database files, F: for transaction logs. NOTE: if you created your Win2k8 VM prior to vSphere 4.1, then there‘s an extra step you need to do to take, to enable the UUID for the VM to register with the guest (more info can be found here). vssadmin list writers (this command lists all the applications that are VSS-aware; its core functions is to listen to the VSS service for shadow copy service, so it could flash data from memory buffer, commit pending transaction logs, and freeze new writes). This is command is also a good tool for quick spot check on whether the previous application quiescing was successful. If you need to dig deeper on failures, VMware KB 1037376 and KB 1007696 have list of instructions for tools log, vss trace, in addition to this command. Below is the output from my SQL2008 VM: Don‘t bother with the ―vssadmin list shadows command‖, unless you are leveraging the built-in system providers (in most cases, you will NOT be using those).
  • 6. Here‘s a workflow diagram of application quiesced snapshot workflow with VMware tools VSS provider, working in conjunction with Nimble Storage volume collection with ―vCenter Sync‖ snapshot schedule. At this point, you might be curious to see what the workflow looks like when hardware VSS provider is used – let‘s use MS Exchange application for our example:
  • 7. 1. NPM schedule triggers snapshot process; NMP agent request for shadow copy through VSS 2. VSS tells Exchange to ―quiesce‖ mail stores 3. Exchange ―quiesce‖ mail stores and alerts VSS upon completion 4. VSS tells NTFS to flush buffer cache. 5. VSS tells Nimble array to take a snapshot of volume(s) – either in-guest mounted or RDM in passthru mode (since VMware ESX ignores all RDM in PT-mode during vmware snapshot creation task) 6. Nimble array captures snapshots of all volumes in collection; the hardware provider also truncates the Exchange logs (as an app consistent, quiesced copy of the data is now captured) That‘s all there is to VSS! And this wraps up our part I of the post for data protection with VSS. Stay tuned for part II – we will take a deep dive into design considerations based on various disk connection methods (think RDM, direct attach/in-guest mounted, regular VMDK).
  • 8. VSS 101 and Design Considerations in VMware environment (Part II) Posted on November 29, 2012 by wenjyu Nimble engineers(Jay Wang, AnaghaBarve, SathyaBhat, Scott Moreland) and tech marketing strike again with part II of VSS. Now that you have basic understanding of how VSS framework and VMware quiesced snapshot integration work (if you haven’t read the first post, click here). Now let’s jump into the subject of design considerations – for those of you that read my blog regularly, you know this is my favorite subject – highlighting areas of that you should watch out for, when designing your virtualization infrastructure. Here we go, for each disk attachment method available in ESX environment: RDM(Physical Compatibility Mode) for application disks, VMDK for OS disk In this case, VMware will simply ignore the RDM disk during snapshot operation, meaning ESX will only create a VMware snapshot for the O/S VMDK. As for the application disk that is running as RDM, the Nimble VSS hardware provider will be used for snapshots. Therefore, it is imperative to ensure the Volume Collection containing the RDM volume has “Microsoft VSS” synchronization selected. NOTE
  • 9. 1) With “Microsoft VSS” synchronization, there would be NO VMware snapshot taken by ESX servers. The Nimble VSS hardware provider will be leverage for taking the snapshot, after the VSS writer has successfully freeze incoming I/O requests 2) The Nimble Windows Toolkit (Nimble Protection Manager/NMP) needs to be installed on the VM that has RDM storage attached VMDK for both application and OS disks With this configuration, keep in mind that ALL VMsn a given VMFS volume need to be quiesced + VMware snapshot taken before the array takes a volume level snapshot. It is a wise idea to limit the number of virtual machines you have in a given VMFS volume. For example, if you have VMs that are running file share/print/web services, then you are basically wasting the time for taking a ‘quiesced’ snapshot, as the application is stateless in nature. Simply create another volume to host such VMs, and ensure the volume collection contains only VMs that require VSS quiesced snapshot (with appropriate VSS writer).
  • 10. NOTE The current VMware implementation of the software provider does NOT truncate logs for Exchange. If you have an integrated backup application such as Commvault that could be invoked to truncate the logs, be sure to leverage that. If not, you could 1)enable circular logging in Exchange 2)consider inguest/RDM mounted storage 3)build custom script to invoke during backup to truncate the Exchange logs. Direct attached/in-guest mounted storage for application data, VMDK for OS disk With this configuration, the in-guest mounted storage will bypass the ESX VMkernel storage stack, and simply appear as network traffic to ESX. Customers typically use this configuration for 1)MSCS on iSCSI protocol 2)get beyond the 2TB VMDK size limitation. Just like any other method, there are design considerations/tradeoffs. NOTE
  • 11. 1) The “gotcha” with this configuration is SRM (Site Recovery Manager) and other upper level solution (i.e., vCloud Director) interoperability. Let’s start with SRM – it does NOT know about in-guest mounted storage as storage/vmdk, therefore, you’d have to add extra steps to mount these volumes for each VM that uses this type of storage. Refer to my previous post on SRM “gotchas” for further details. For vCloud Director, you will NOT be able to package the VM as a vApp template and deploy without manual intervention to mount the in-guest storage. Additionally, in terms of performance monitoring, esxtop will not display disk level stats for the in-guest storage – you’d have to reply on the network stats (press ‘N’). vCenter Operations will not interpret the in-guest attached storage stats as ‘storage’ info. It will get treated as network traffic 2) The Nimble Windows Toolkit(Nimble Protection Manager/NMP) needs to be installed on the VM that has RDM storage attached Last but not least, with all three virtual machine storage connectivity methods above, the following “gotcha” applies: *ONLY a scheduled snapshot operation in a Volume Collection would invoke VSS quiescing, a manual snapshot that you take from vCenter Server Nimble plugin and/or Nimble UI would NOT trigger the VSS requestor to request for application quiescing* In Summary, here’s a quick table reference on all three connectivity options and considerations/caveats for each: Connectivity Method Snapshot Synchronization NPM Installation Raw Device Mapping(RDM) Microsoft VSS Inside the guest OS (NPM requestor & provider used for snapshot) No VMware snapshot taken VMDK vCenter None (VMware VSS requestor/provider used for snapshot) No Log truncation for MS Exchange Inside the guest OS (NPM requestor & provider used for snapshot) Manual work needed for SRM/vCD Direct Attached (In-guest mounted) Microsoft VSS Considerations Avoid too many VMs in the same datastore No disk level stats from esxtop/vC Ops NPM requestor
  • 12. &provider used for quiescing That’s it for now – happy backing up your VMs! Oh, don’t forget to restore them once in a while – what good is backup when you can’t restore from it, right?
  • 13. “KISS RAS” = Software Defined Storage – Come see Nimble Storage‟s tech previews @ VMworld next week! Posted onAugust 21, 2013bywenjyu If you are in the virtualization space, you must have heard of the term ―Software Defined Storage‖ – probably more so from storage vendors than VMware. Why? Because it is a hot marketing theme to promote storage technologies. More and more customers & partners are asking what is Nimble‘s take on ―Software Defined Storage‖. Here‘s my unofficial response – ―K.I.S.S – R.A.S = Software Defined Storage‖: Keep It Simple Storage Reliability Availability Supportability I personally hope all storage vendors could focus more on these aspects in their solution/offerings so customers could benefit from this ‗software defined‘ era. Seriously, having REST-APIs in your product release does not make it ‗software defined storage‘ – what good do they do if the customer has to get a field engineer to upgrade the software to the next version? Now let‘s expand further on the KISS RAS aspects, and what we will showcase in our tech preview station @ VMworld next week! K.I.S.S (Keep It Simple Storage) VMware has three key storage integration points to keep things simple for end users: 1)vCenter plugin (extensibility in vCenter to allow storage vendor to provide a plugin for storage operations such as datastore provisioning, snapshotting, cloning, expansion) Here‘s a sneak peak @ our new enhanced vCenter plugin:
  • 14. 2)vVOL (a.k.a virtual volume) (simply put, VMware providing metadata to storage vendor to further offload storage operations such as provisioning, cloning, snapshotting, replication. Up leveling that a bit, it is the platform that enables people to have true policy based management for storage) – my good friend Cormac published a very nice blog on this topic in case you haven‘t seen it (http://blogs.vmware.com/vsphere/2012/10/virtual-volumes-vvols-tech-previewwith-video.html) Here‘s a sneak peak @ our vVOL demo:
  • 15. 3) PSA (Pluggable Storage Architecture): this integration point allows storage vendors to plug into the multipathing framework within ESXiVMkernel. What I am most excited about is our upcoming PSP plugin – this new Path Selection Plugin would auto detect newly available paths , and optimize I/O distribution across all auto discovered paths: From the storage software OS side, here‘s a sneak peak of our next release – can you spot what is NEW from looking at this picture? (HINT: look @ green circles!)
  • 16. R. A. S (Reliability, Availability, Supportability) This is an area where storage vendors should invest heavily (aside from creating all the marketing *COUGH* stuff) . We should leverage the ―software defined‖ theme to bring software closer to hardware (let it be commodity or proprietary) – the software should have tight handshake with the underlying hardware, and proactively process hardware state changes to do the right thing to accommodate failures, report it, and show actionable items for end users to take to prevent it in the future. If you are evaluating ―software defined storage‖, be sure to evaluate it with the following R.A.S criteria in mind: 1) Monitoring & alerting (this is pretty self-explanatory – does the solution provide active monitoring of the software & hardware state, and alert you when things go wrong? Do you get actionable information out of the alerts?) 2) Risk determination (doe the solution inform you about things that *might* go wrong in the future, if you don‘t take corrective action? You shouldn‟t have to predict what the actions are, it should tell you!) 3) Efficiency Measurement (does the solution inform you whether things are running efficiently? Just because things are up and running doesn‘t mean they are optimized) vCenter Operations Manager addresses all three requirements above, for all the virtual entities/managed objects in the virtual infrastructure (VM CPU, memory, network, storage) – now, should vCenter Operations Manager report a VM being bound by storage, you‘d definitely want to dive deeper into the ―software defined storage‖ solution to find out which of the three areas to look at. Let‘s look at a real world scenario below – you have a VM that‘s running very active workload, and vCenter Operations indicates to you that the VM is bound by disk I/O, and
  • 17. latency is high…your software defined storage solution should provide a way for you to look deeper and determine what might be a possible cause. From Nimble perspective, that‘s where Infosight comes in to play. Here‘s a sneak preview of what our support engineers and engineering teams look at – especially the supportability aspect. If you haven‘t seen the current release of InfoSight, check out the demo here and Brain (Neil Glick‘s blog here). If you have been using it, come see a tech preview of what is coming at our VMworld booth – teaser screenshot below: Here we go, vCenter Operations Manager reports VM bound by CPU: What might cause the latency? hmm how about a closer look from a new feature being added to InfoSight?
  • 18. Aside from all the new tech preview stuff we will be showing at VMworld – here‘s a shameless self-plug for my presentations next week: vBrownBag „esxtop‟ technical talk (Monday, 8/26 @ 5:15PM) http://professionalvmware.com/2013/08/vbrownbag-tech-talks-vmworld-2013/ BCO5431 DR Panel (Tuesday, 8/27 @5:30PM) Panel session to share tips/tricks/‖gotcheas‖ in DR design for private and hybrid cloud (joining me in this session is the following customers: Jeff Winters from City of Hot Springs, Jerry Yang from Foster Pepper, Andrew Fife from Virtacore, Bryan Bond from Siemens e-meter) That‘s it for now – hope to see you @ VMworld next week!!
  • 19. Red Hat OpenStack up and running in minutes with Nimble Cinder Driver? The “Why”, “What” and “How”…. Posted on August 15, 2013 by wenjyu There are lots of buzz around OpenStack these days – I‘m sure you want to give it a spin and see what the hype is all about. Here‘s a quick way to try it out using your existing resources (seriously, most of us don‘t have the freaking budget to buy new servers for experimental/eval projects; if you do, more power to you!). NOTE: NOTHING below is officially supported by VMware, Red Hat, Nimble Storage. Why am I doing it? Because I can J Wouldn‘t you want to leverage what you have in the infrastructure and check out what the hype is all about? If you answered yes, then read on. (*Special Thanks to Jay Wang and Brady Zuo in our R&D team for making this happen!) OpenStack 101 For those of you that don‘t know what OpenStack is, here‘s a quick 101 introduction, with key buzzwords to use so you sound like a pro: Founded by RackSpace + NASA several years back Got massive traction after the formation of OpenStack Foundation It is cool because its fate is not controlled by a single company/vendor, projects are driven/led by end users, and the community It is a cloud Platform that has extensive API framework (rich API sets that allows for easy automation + built-in extensibility for the portal) API compatibility with AWS EC2 and S3 Consists of the following projects: o Nova (compute)
  • 20. o o o o o o o Cinder (block storage: you probably know this one; volume access based on SCSI standard) Swift (object storage: scalable storage for unstructured data: think pictures, videos, iso images) Neutron (networking; used to be called Quantum; you do have a choice with various vendor‘s SDN implementation) Glance (image repository: where your ISO, VM images reside) Horizon (Dashboard/self service portal: this is where admin comes to create projects, review project resource utilization; this is where end consumers login to deploy instances/VMs) Keystone (identify/authentication) Ceilometer (monitoring/operational mgmt) Here are some very useful links from various folks in the community: https://ask.openstack.org/en/questions/ (great community support; I asked a few questions here and there and folks are very responsive) http://www.openstack.org/summit/portland-2013/session-videos/ (my favorites are: Day 2OpenStack Grizzly Architecture 101, Top 10 Things We‘ve learned about OpenStack by my friend Cloud Czar from NetApp , OpenStack in Production: the Good, the Bad & the Ugly; Day3-OpenStack, SDN, and the Future of Infrastructure-as-a-Service) http://cloudarchitectmusings.com/page/2/ (Blog by Kenneth Hui from RackSpace – very informative/technical blog posts) Now, let‘s jump straight into my setup:
  • 21. Step 0: Download CentOS 6.4 DVD if you haven‘t done so already: http://isoredirect.centos.org/centos/6/isos/x86_64/ Step 1: Controller VM Install Install CentOS (DUH) as a VM, I chose ―Base Server‖ so I am not sure if a minimal one would have all the dependencies needed for OpenStack ―Gotchas‖ -if you bypass networking setup during installation, your eth0 interface would be disabled by default, to get it back, simply do this: #vi /etc/sysconfig/network-scripts/ifcfg-eth0 Change ―ONBOOT=no‖ to ―ONBOOT=yes‖ #service network restart -be sure you connect this to a network with internet access (unless you are going to build your own YUM repository) Step 2: Install Red Hat OpenStack
  • 22. Install RDO packages for OpenStack – I simply follow the instructions from Red Hat‘s RDO page: http://openstack.redhat.com/Quickstart *I actually wanted to try out Neutron so I use ―packstack –allinone –os-quantum-install=y‖ instead* After this command finishes (takes about 10 minutes or so), you will see a screen like this: ―Gotcha‖ -the default admin password is cryptic, and in a non-obvious location; you could find the actual password in the following file: ―keystonerc_admin‖ After you find the password, copy it, and then open up OpenStack Horizon in your browser: http://<your_serverhost/ip>/dashboard -don‘t forget to change the admin password to one that you would actually remember Step 3: Install Nimble Cinder Driver (this step is optional if you don‟t want to use shared storage for Cinder; you could start creating project and stand up instances with the VMDK that‟s attached to the CentOS VM) *Note* if you are a customer of Nimble and want to try out the Cinder driver as a tech preview, give us a shout out; or send me an email wen@nimblestorage.com After you obtain the binary from us: a) #cp nimble.py /usr/lib/python2.6/site-packages/cinder/volume/drivers/
  • 23. b) Add the following lines to /etc/cinder/cinder.conf nimble_username=admin nimble_password=pass nimble_mgmt_ip=your_array_mgmt_ip volume_driver=cinder.volume.drivers.nimble.NimbleISCSIDriver volume_name_template = volume-%s c)restart all related cinder services: cd /etc/init.d/ for i in $( lsopenstack-cinder-* ) do service $i restart done You are now ready to rock n‘ roll! You can perform the following operations with Nimble: create volume snap volume clone volume delete volume Each volume could be tied to an individual instance in OpenStack. So you could enjoy the benefits of per-VM level performance monitoring, snapshot, cloning with zero copy clone. One advantage you have with shared storage for Cinder is you won‘t have to scale compute to get storage capacity/performance. I will post more stuff once I get to play with this setup some more. Stay tuned!
  • 24. Another deeper look at deploying Nimble with Cisco UCS Posted onJuly 30, 2013bywenjyu We continue to get customer inquiries on the specifics of deploying Nimble with Cisco UCS – particularly on what the service profile should look like for iSCSIvNICs. So here we go, we will dive straight into that bad boy: We will start with the Fabric Interconnect, then to the vNICS, then to Nimble array, and last but not least, the vSpherevSwitch 1. Fabric Interconnect Configure cluster for the FI o The FIs should be configured in cluster mode, with a primary and subordinate (clustering of FI does NOT mean data traffic flows between the two – it is an active/passive cluster with management traffic flowing between the pair) o Configure appliance ports  The ports connected with Nimble data interfaces should be configured with appliance port mode – why you may ask? Well, prior to UCSM 1.4 release, the ports on the FI are Ethernet ports that will receive broadcast/multicast traffic
  • 25. from the Ethernet fabric. Appliance ports are designed specifically to accommodate Ethernet-based storage devices such as Nimble so its ports don’t get treated as another host/VM connected to an Ethernet uplink port  Here’s what ours look like for each FI (under “Physical Ports” for each FI in the “Equipment” tab FI-A (connect one 10G port from each controller to the FI-A) FI-B (connect remaining 10G port from each controller to FI-B) 2. vNIC (it’s important to pin the iSCSIvNICs to a specific FI) In our service profile, we have two vNICs defined for iSCSI traffic, and each vNIC is pinned to a specific FI1 Here‘s what the vNIC setting looks like for each vNIC dedicated for iSCSI (under ―General‖ tab):
  • 26. We use VLAN 27 & 28 representing the two subnets we have Why didn‘t we check ―Enable Failover‖? Simply put, we let ESX SATP/PSP to handle failover for us. More on topic is discussed in my joint presentation with Mostafa Khalil from VMware. 3. Nimble Array Notice we have subnet 127 & 128? Why you may ask – that is so we could leverage both FIs for iSCSI data traffic
  • 27. 4. vSpherevSwitch We will need two VMkernel ports for data traffic, each configured on a separate subnet to match our design. You could use either a single vSwitch or two vSwitches. Note if you use a single vSwitch, your NIC teaming policy for each VMKernel port must be overridden like below: How the hell do I know vmnic1 & vmnic2 are the correct vNICs dedicated for iSCSI? Please don‘t share this secret J If you click on ―vNICs‖ under your service profile/service profile template, you get to see the ―Desired Order‖ in which they will show up in ESX – remember, ESX assigns this based on the PCI bus number. Desired order of ―1‖ will show up as vmnic0, so our vNICiSCSI-A with desired order of ―2‖ will show up as vmnic1, so forth with vNICiSCSI-B That‘s it, wish I could make it more complicated. If you missed my post on booting ESXi from SAN on UCS, check out my previous pos
  • 28. iSCSI Booting UCS blade/rack server with Nimble Storage Posted onJune 6, 2013bywenjyu More and more customers are inquiring about how to do boot-from-SAN with Nimble Storage (save a few bucks on having to buy local SAS/SSD disks) – fortunately, we got both in our playground. Here it is, step-by-step checklist/instructions… *NOTE* *The setup is configured to attach Nimble directly to the fabric interconnects with two subnets defined (one for each 10G interface of the controller) – if you attach the Nimble to a pair of access layer switches such as Nexus 5k with vPC, then dual subnets is NOT needed. Remember, even though the FIs are configured as a cluster pair, the cluster interconnect interfaces between the FIs DO NOT carry data traffic, thus, the need for dual subnets for both FI connections to be active to the Nimble array
  • 29. *You have to have a service profile template configured to have iSCSI boot – here is what yours should look like, for the ―iSCSIvNICs‖ and ―Boot Order‖ tabs:
  • 30. ―Boot Order‖ tab should be pretty self-explanatory, first boot from CDROM, then iSCSI VNICs
  • 31. Next up is configuring the IP and target parameters for each iSCSIvNIC (this is so the iSCSIvNIC knows where to find the boot LUN) – remember to configure BOTH vNICs otherwise you‟d have a single point of failure!
  • 32. 1)ensure you have an IQN pool to derive iQN name for the iSCSIvNIC initiator (so an initiator group can be created on the array side to ensure only this blade could have access to the boot volume) *take note of the initiator name so you could add it to the initiator group on the array side 2)set a static IP for the iSCSIvNIC (there‘s a nice little feature here to determine if the address is used by other blades within UCSM) 3)add the iSCSI array target information(at this point, you are going to switch to the Nimble interface to obtain two pieces of required info: create a boot volume for the blade and obtain its UUID obtain the target discovery IP
  • 33. Here‘s how to get this stuff from the Nimble side: First let‘s get the iSCSI discovery IP:
  • 34. Next obtain the iQN for each iSCSIvNIC and add them into an initiator group (you certainly want to do this because you don‘t want the volume to be presented to every and any host that sees it!)…the IQN can be found under the server‘s service profile->―boot order‖->vNIC1 or vNIC2->‖set iSCSI Boot Parameters‖: Remember, you‘d want to get the IQN for both iSCSI VNICs, and then add them to an initiator group from the Nimble side:
  • 35. Once the initiator group has been created, you are now ready to create a volume to serve as the boot LUN for the ESXi host:
  • 36. Notice I have all the boot LUNs on the ESX server configured to be part of a volume collection – it‘s for safety measures in case something were to go wrong with the ESXi install. It‘s a lot quicker to restore from a snap than to reinstall ESX all over again (unless you are fortunate to have stateless ESXi Auto Deploy configured). If you have a DR site, it certainly wouldn‘t hurt to configure replication for the ESXi boot LUN volume collection as well! After the volume has been created, we‘ll then obtain the UUID for the volume so it could get entered in the blade server‘s iSCSI boot target parameter:
  • 37. Here we go again, same screen as before, but this time with the required info for both 1 & 2 below: Now you are ready to boot/install ESXi on a Nimble volume! Power on the blade, and watch for the following screens to make sure the volume is discovered correctly:
  • 38. On Nimble (as the server boots, the iSCSI connection would be made if things are configured properly – if ―Connected Initiators‖ count remains ‗0‘ even when you see the ESXi install prompt, then go back to the previous iSCSI boot parameters to make sure 1)array target IP is entered correctly 2)boot volume UUID is entered correctly for EACH of the iSCSIvNICs: Want to see this in action
  • 39. Notice I have all the boot LUNs on the ESX server configured to be part of a volume collection – it‘s for safety measures in case something were to go wrong with the ESXi install. It‘s a lot quicker to restore from a snap than to reinstall ESX all over again (unless you are fortunate to have stateless ESXi Auto Deploy configured). If you have a DR site, it certainly wouldn‘t hurt to configure replication for the ESXi boot LUN volume collection as well! After the volume has been created, we‘ll then obtain the UUID for the volume so it could get entered in the blade server‘s iSCSI boot target parameter:
  • 40. Here we go again, same screen as before, but this time with the required info for both 1 & 2 below:
  • 41. Now you are ready to boot/install ESXi on a Nimble volume! Power on the blade, and watch for the following screens to make sure the volume is discovered correctly: On Nimble (as the server boots, the iSCSI connection would be made if things are configured properly – if ―Connected Initiators‖ count remains ‗0‘ even when you see the ESXi install prompt, then go back to the previous iSCSI boot parameters to make sure 1)array target IP is entered correctly 2)boot volume UUID is entered correctly for EACH of the iSCSIvNICs: Want to see this in action? Check out Mike Mclaughlin‘s demo video on Nimble Connect
  • 42. Storage for Your Branch Office VDI By Radhika Krishnan, Head of Solutions and AlliancesOn October 10, 2012 · Add Comment ―One size fits all‖ is not a tenet applicable for branch office deployments. Branch office deployments vary based on number of users, type of applications, available WAN bandwidth, and IT infrastructure. A recent ESG survey shows that 92 percent of customers are interested in replacing the laptops/PCs of their remote office employees with virtual desktops running in a central location. VMware, working with partners such as Nimble Storage, provides a variety of approaches to tackle VDI with their Branch Office Desktop solution. One way to deploy is the centrally-hosted desktop environment and the other is using a locally-hosted desktop solution kept in sync using VMware Mirage. Either way, you want to make sure the storage you choose is optimized for your specific deployment model. With the centralized desktop hosting model, the storage not only needs to be cost-effective, but also deliver high performance. This is where Nimble shines with optimizations for both read and write operations. In addition, built-in snapshot-based data protection allows you to dramatically improve business continuity. In the case of the locally-hosted desktop environment, the primary storage factors in play are cost, ease of deployment and use, as well as supportability. Nimble‘s CS200 series offers costeffective, dense performance for these environments in a 3U form factor. Systems come tuned out of the box with only a few easy steps to go from plug-in to operational. Finally, proactive wellness features enable the relay of comprehensive system telemetry data to Nimble HQ that is analyzed in real-time to identify potential problems at the branch location. In case of a support incident, the system can be serviced remotely the majority of the time. Additionally, in the locally-deployed VDI solution, Nimble Storage used for branch office desktops can serve as a replication target for the primary datacenter. That way critical data can
  • 43. be replicated and protected in case of a disaster at the primary site. Similarly, critical data such as local persona and user files can be replicated back to the primary datacenter.
  • 44. re All Hybrid Storage Arrays Created Equal? By Ajay Singh Vice President, Product ManagementOn October 9, 2012 · Add Comment Nimble Storage was founded in early 2008 on the premise that hybrid storage arrays would be the dominant networked storage architecture over the next decade – a premise that is now widely accepted. The interesting question today is, ―Are all hybrid storage arrays are created equal?‖ After all, SSDs and HDDs are commodities, so the only factor setting them apart is the effectiveness of the array software. How does one compare hybrid storage arrays? Here are some key factors: 1. How cost-effectively does the hybrid storage array use SSDs to minimize costs while maximizing performance? 2. How cost-effectively does the hybrid storage array use HDDs to minimize costs while maximizing useable capacity? 3. How responsive and flexible is the hybrid array at handling multiple workloads and workload changes? 4. Aside from price/performance and price/capacity, how efficient is the array data management functionality (such as snapshots, clones, and replication)? This blog will cover the first three. The fourth dimension of efficient data management is a very important factor in evaluating storage arrays, and a topic we‘ll cover in detail in a future blog post. How cost-effectively does the hybrid storage array use SSDs? Most hybrid storage array architectures stage all writes to SSDs first in order to accelerate write performance, allowing data that is deemed less ―hot‖ to be moved to HDDs at a later point. However as explained below, this is an expensive approach.Nimble storage arrays employ a unique architecture in that only data that is deemed to be cache-worthy for subsequent read access is written to SSDs, while all data is written to low-cost HDDs. Nimble‟s unique architecture achieves very high write performance despite writing all data to HDDs by converting random write IOs issued by applications into sequential IOs on the fly, leveraging the fact that HDDs are very good at handling sequential IO. 1. Write endurance issues demand the use of expensive SSDs. When SSDs receive random writes directly, the actual write activity within the physical SSD itself is higher than the number of logical writes issued to the SSD (a phenomenon called write amplification). This eats into the SSD lifespan, i.e. the number of write cycles that the SSD can endure. Consequently, many storage systems are forced to use higher endurance eMLC or SLC SSDs, which are far more expensive. In addition to the selective writing capability mentioned above, the Nimble architecture also optimizes the written data layout on SSDs so as to minimize write amplification. This allows the use of lower cost commodity MLC SSDs, while still delivering a 5 year lifespan.
  • 45. 2. Overheads reduce useable capacity relative to raw capacity of SSDs.Hybrid arrays that can leverage data reduction techniques such as compression and de-duplication can significantly increase useable capacity. On the flip side, RAID parity overheads can significantly reduce useable capacity. Nimble’s architecture eliminates the need for RAID overheads on SSD entirely and further increases useable capacity by using inline compression. 3. Infrequent decision-making about what data to place on SSDs and moving large-sized data chunks wastes SSD capacity. Most hybrid storage arrays determine what data gets placed on SSDs vs. HDDs by analyzing access patterns for (and eventually migrating) large “data chunks”, sometimes called pages or extents. This allows “hot” or more frequently requested data chunks to be promoted into SSDs, while keeping the “cold” or less frequently requested data on HDDs. Infrequent decisions on data placement cause SSD over-provisioning. Many storage systems analyze what data is “hot” on an infrequent basis (every several hours) and move that data into SSDs with no ability to react to workload changes between periods. Consequently, they have to over-provision SSD capacity to optimize performance between periods. Nimble’s architecture optimizes data placement real-time, with every IO operation. Optimizing data placement in large data chunks (many MB or even GB) causes SSD overprovisioning. The amount of meta-data needed to manage placement of data chunks gets larger as the data chunks get smaller. Most storage systems are not designed to manage a large amount of meta-data and they consequently use large-sized data chunks, which wastes SSD capacity. For example, if a storage array were to use data chunks that are 1GB in size, frequent access of a database record that is 8KB in size results in an entire 1GB chunk of data being treated as “hot” and getting moved into SSDs. Nimble’s architecture manages data placement in very small chunks (~4KB), thus avoiding SSD wastage. How cost-effectively does the hybrid storage array use HDDs? This means assessing the ratio of usable to raw HDD capacity, as well as the cost per GB of capacity. Three main areas drive this: 1. Type of HDDs. Many hybrid arrays are forced to use high-RPM (10K or 15K) HDDs to handle performance needs for data that is not on SSDs, because of their (higher) random IO performance. Unfortunately high RPM HDD capacity is about 5x costlier ($/GB) vs. low RPM HDDs. As mentioned earlier, Nimble’s write-optimized architecture coalesces thousands of random writes into a small number of sequential writes.Since low-cost, high-density HDDs are good at handling sequential IO, this allows Nimble storage arrays to deliver very high random write performance with low-cost HDDs. In fact a single shelf of low RPM HDDs with the Nimble layout handily outperforms the random write performance of multiple shelves of high RPM drives. 2. Data Reduction. Most hybrid arrays are unable to compress or de-duplicate data that is resident on HDDs (some may be able to compress or de-duplicate data resident on SSDs). Even among those that do, many recommend that data reduction approaches not be deployed for transactional applications (e.g., databases, mail applications, etc.). The Nimble architecture is able to compress data inline, even for high-performance applications. 3. RAID and Other System Overheads. Arrays can differ significantly in how much capacity is lost due to RAID protection and other system overheads. For example many architectures force the use of mirroring (RAID-10) for performance intensive workloads. Nimble on the other hand uses
  • 46. a very fast version of dual parity RAID that delivers resiliency in the event of dual disk failure, allows high performance, and yet consumes low capacity overhead. This can be assessed by comparing useable capacity relative to raw capacity, while using the vendor’s RAID best practices for your application. How responsive and flexible is the hybrid array at handling multiple workloads? One of the main purposes of a hybrid array is to deliver responsive, high performance at a lower cost than traditional arrays. There are a couple of keys to delivering on the performance promise: 1. Responsiveness to workload changes based on timeliness and granularity of data placement. As discussed earlier, hybrid arrays deliver high performance by ensuring that “hot” randomly accessed data is served out of SSDs. However many hybrid arrays manage this migration process only on a periodic basis (on the order of hours) which results in poor responsiveness if workloads change between intervals. And in most cases hybrid arrays can only manage very large data chunks for SSD migration, on the order of many MB or even GB. Unfortunately, when such large chunks are promoted into SSDs, large fractions of that can be “cold data” that is forced to be promoted because of design limitations. Then because some of the SSD capacity is used up by this cold data, not all the “hot” data that would have been SSD worthy is able to make it into SSDs. Nimble’s architecture optimizes data placement real-time, for every IO that can be as small as a 4 KB in size. 2. The IO penalty of promoting “hot” data and demoting “cold” data. Hybrid arrays that rely on a migration process often find that the very process of migration can actually hurt performance when it is most in need! In a migration based approach, promotion of “hot” data into SSDs requires not just that data be read from HDDs and written to SSDs, but also that to make room for that hot data, some colder data needs to be read from SSDs and written into HDDs – which we already know are slow at handling writes. The Nimble architecture is much more efficient in that promoting hot data only requires that data be read from HDDs and written into SSDs – the reverse process is not necessary since a copy of all data is already stored in HDDs. 3. Flexibly scaling the ratio of SSD to HDD on the fly. Hybrid arrays need to be flexible in that as the attributes of SSDs and HDDs change over time (performance, $/GB, sequential bandwidth, etc.), or as the workloads being consolidated on the array evolve over time, you can vary the ratio of SSD to HDD capacity within the array. A measure of this would be whether a hybrid array can change the SSD capacity on the fly without requiring application disruption, so that you can adapt the flash/disk ratio if and when needed, in the most cost effective manner. We truly believe that storage infrastructure is going through the most significant transformation in over a decade, and that efficient hybrid storage arrays will displace modular storage over that time frame. Every storage vendor will deploy a combination of SSDs and HDDs within their arrays, and argue that they have already embraced hybrid storage architectures. The real winners over this transformation will be those who have truly engineered their product architectures to maximize the best that SSDs and HDDs together can bring to bear for Enterprise applications.
  • 47. NimbleStorage User Interface leave a comment » In my last post related to NimbleStorage it was more focused on the Company, and products they offer. However I did note that I would take a few moments to document the UI and how easy it is to manage. In this post I am going to go over the UI and the different features contained within it. The UI is Adobe Flash based, you will need v9 or higher installed on your management system to view all aspects of the UI, it will still function without Adobe Flash installed, but you will be missing some of the charts. I have ran the interface from Chrome, IE, FireFox, and Safari without any issues. Landing Page Once you open up your browser of choice go to the management IP that you configured on your array. You will be presented with the following page. Home This is the Home Page, and will contain a dashboard related to all the information you will normally want to see from a quick health check.
  • 48. Disk Space: Will show everything you need to know about your usage, as you can see it will show the Volume Usage, Snapshot Usage, and unused reserve. It is very simple to read, and makes for a great screen shot. Space Savings: This is where you get to see your savings and how the Nimble is saving you some space. As you can see from below my space savings is 2.51 TB or 40.55%. Throughput: From here you can see how much data is passing over your Array interfaces. This shows current usage and for the past Hour. IOPS: Again from here you get to see your Current IOPS usage across all your LUNs, this also shows back for an hour. Event Summary: All Events that have happened in the last 24 Hours Recent Events: All recent events, last hour. The interface is controlled from a tabbed driven menu, this is the the menu you will see from any page within the Nimble Management Interface. I will explain the rest of the Interface in order.
  • 49. Manage The Manage tab is where you can configure your Volumes, Replication, Snapshots, Protection and Performance Policies, Array, and Security. Volumes The Volume Menu, is where all the magic happens with the Volumes, you can do anything volume related within this Menu. The first page you will see will give you a full overview of all Volumes that reside on your array, with basic information related to the usage. Click on a volume name, and now you have all information pertaining to that Volume. Here you will get the overview of your volume, you can edit, take a snapshot, set it to offline, Take ownership of the volume with ―Claim‖ (Used if the Volume came from a replicated source) or Delete the Volume. You can also look at the Snapshot, and replication status for the Volume.
  • 50. Protection The Protection Tab offers three choices. 1. Volume Collections: This allows you to protect your Volumes with Snapshot schedules and replicate those snapshots off to a second Nimble Array. 2. Protection Templates: Predefined templates to help protect your critical applications. You can use these to as a baseline for creating your own custom Volume Collection Groups. Nimble was just nice enough to give us something to start with.
  • 51. 3. Replication Partners: A place where you can view your configured replication partners, or define new ones. I don‘t have any configured but you would see them here, you can also setup Bandwidth Policies so you can replicate without having to worry about the bandwidth being affected during busy times within your environment. Array From here you can look at the status of your Hardware, and view all the system information related to your Array and Controllers. You can also edit the Array name, and Network Address information.
  • 52. Performance Policies Here is where you can setup Policies to configure your Volumes with certain parameters. These Policies give you the option to create volume with certain Block size, and if you want to cache, or compress the volume. We all know we have certain workloads or datasets that we don‘t want to waste our precious cache on or they won‘t benefit from compression. Initiator Groups This allows us to create groups of initiators that we can then assign to volumes to allow a host access to the volume. Chap Accounts From here we can create a Chap Account, that we can then assign to Volumes we create to gives us some control over what we allow to connect to our Volume.
  • 53. Monitor This is the place where you can see all the great performance you are getting out of your Nimble Performance You can view performance based off all Volumes, or a selected Volume. The time frame can also be for 5 Minutes, 24 Hours, 7 Days, or 30 Days. This will give you a very good picture of how your system is performing, it will also allow you to pin-point Volumes that are eating up the majority of your system.
  • 54. Replication The place you will want to head to find out how your replication are performing, and if you have any issues with them.
  • 55. Connections This section shows each Volume you have configured, and how many connections are connected to the Volume. From here you can make sure you have the proper amount of connections as you have configured from your host.
  • 56. Events This Menu shows all the events related to your array, it keeps a log for 30 days, and you have the ability to filter the results based off of severity, Category, and Time. Administration As the name depicts this has everything to do with Administration. I will explain the Menu items below. This is the only Menu that once you select one item, the other items will be listed on the left hand side of the window. This gives you quicker access to the other items. Email Alerts Here you can configure the system to send you email alerts. You can also tell it to send alerts back to Nimble support, which will create a support ticket.
  • 57. AutoSupport/HTTP Proxy AutoSupport enables the array to send health updates to Nimble Support, this lets them know what Software you are currently running, and if you have any configuration issues that may exist. This is a very nice feature, Nimble support will contact you regarding Software Updates, related to your system, they will know which version you are running and why you should be running the latest version. It gives you the personal touch when it comes to support. Also from this menu you can enable Secure Tunnel, this allows Nimble Support to directly connect to your array. This can be enabled or disabled at any time, you can leave it disabled until you need a support tech to connect.
  • 58. Software This is where you can update the software for the Array. It will keep two versions of the software on the Array. When you want to check for a new version click Download, and it will connect to the Nimble Support site, and check for any software updates that are compatible with your Array. If your Array is not connected to the Internet on the management interface, you can go ahead and upload a software file to the array.
  • 59. Change Password No need for a picture here, you can change your password. Default Space Reservations When you create a new Volume, these are the defaults settings that are displayed. Date/Timezone You can set the time, and timezone, you can set this manually or with a NTP.
  • 60. Inactivity Timeout The amount of time before you session expires. The default is 30 Minutes. DNS Here you setup the DNS servers, you can have a min of 1 and a max of 5 SNMP Configured your SNMP to communicate with a network management system. Plugins This is new to 1.3.3.0 Software. Right now all it includes is the ability to add the Nimble Array to vCenter for management from the DataStore View. If you are running Software below 1.3.3.0 you need to do this with CLI. Later in this posting I will talk more about the vCenter Plugin and the CLI. While on the topic of 1.3.3.0 Software, another great feature that was enabled was the ability for the Array to support more than one drive failure, that is great news! If you are running a Nimble Array, upgrade your software as soon as possible.
  • 61. Help We all love to look at the Help Menu, right? This one is pretty intuitive, and make‘s life a little easier. Nimble Support Takes you directly to the Nimble Support Site. About CS-Series Array Gives you the Normal About Screen
  • 62. Administrator’s Guide I like this, no PDF to download, always up to date. Just click the menu item and look for the information you need. Hardware Guide Just like the Administrator‘s guide, just click and enjoy. Conclusion As you can see the Interface is very user friendly, but gives you all the information you need to Configure, Monitor, and troubleshoot your Array. In this post I had planned to talk more about configuring your array, updating the Software, and the CLI. But just the walk through of the UI took longer than expected. I plan in the next couple of weeks to post a few different blogs related to NimbleStorage.
  • 63. O Snapshot: What Art Thou? By SachinChheda - Product Marketing On November 16, 2012 · Add Comment This is part one of a two-part blog on snapshots. In the storage world, snapshots are a point-in-time copy of data. They have been around for some time and are increasingly being used by IT to protect stored data. This post recalls some of the popular Nimble Storage blog posts on snapshots. But before I jump into the list of blogs, let me quickly walk you through the different types of storage-based snapshots, including their attributes. A snapshot copies the metadata (think index) of the data instead of copying the data itself. This means taking a snapshot is almost always instantaneous. This is one of the primary advantages of storage-based snapshots—they eliminate backup windows. In traditional backup deployments, applications either have to be taken off-line or suffer from degraded performance during backups (which is why traditional backups typically happen during off-peak hours). This means snapshot-based backups can be taken more frequently, improving recovery point objectives. However not all snapshot implementations are created equal, posing different requirements and restrictions on their use (example: reserve space required, how frequently snapshots can be taken, and number of snapshots that can be retained). In the ‗Copy-on-Write‘ (COW) implementation, the address map related metadata is copied whenever a snapshot is taken. None of the actual data is copied at that time—resulting in an instant snapshot. In almost all implementations this copy is taken to a ‘pre-designated’ space on storage (aka a snapshot reserve). When the data is modified through writes, the original data is copied over to the reserve area. The snapshot‘s metadata is then updated to point to the copied data. Because of this, ‗COW‘ implementation requires two writes and a read when any of the original data is modified for the first time after a snapshot is taken—causing a performance hit when the original data is updated. This gets progressively worse with frequent snapshots. Vendors such as EMC, IBM, and HP have used COW implementations on their traditional storage. The other major implementation of snapshots is ‗Redirect on Write‘ (ROW). Like COW, only the metadata is copied when a snapshot is taken. Unlike COW, whenever original data is being modified after a snapshot, the write is redirected to a new free location on disk. This means ROW snapshots do not suffer the performance impact of COW snapshots as none of the original data is copied. Nimble snapshots are based on the ROW implementation as the write-optimized data layout in the CASL architecture always redirects writes to new free space. A lightweight, background sweeping process in CASL ensures the continued availability of free space and assures consistent performance, addressing a shortcoming of some older ROW implementations. This efficiency allows IT to think of snapshot + replication in a new light—store weeks/months of history versus
  • 64. mere days of backups with traditional, inefficient implementations. This allows virtually all of the operational recoveries to come from snapshots and dramatically improves RTOs. (Umesh‘s blog ‗A Comparison of File System Architectures‘ linked below covers this in detail.) Nimble Storage snapshots are stored (compressed) alongside the primary data on high-capacity disk drives. This allows 1000s of snapshots to be taken and retained on the same system as primary data. A measurement of our install base shows that over 50% of our customers retain their snapshots for over a month. First is the support of universal inline compression for storing data. This ensures data takes up less space on disk, which as discussed earlier makes replication more efficient and allows for more snapshots to be retained in a given storage space. On average, Nimble‘s install base measure compression rates ranging from 30% to 75% for a variety of workloads. Second is the support of cloning, which is fully functional read/writable copy of the original. Cloning is useful in the cases of VDI and test/development where many copies (clones) of a master data set are needed. In the ROW implementations, clones do not take up any additional space. Last but not least is the granularity of the snapshot. This determines how small a snapshot can be for a volume. This is relevant when the data volume being protected has a small rate of daily change. When the extent of a data write is smaller than the snapshot granularity, the snapshot wastes considerable space storing a duplicate copy of unchanged data. Snapshots in Nimble’s CASL architecture can be as granular as a single 4K block.
  • 65. Before going onto the blogs, I wanted to share that Nimble Storage (@NimbleStorage) and CommVault (@CommVault) recently did a joint Twitter Chat on the Nimble Storage integrations through CommVaultIntelliSnap Connect Program. The chat featured experts from Nimble Storage (@wensteryu, @schoonycom& @scnmb… me) and CommVault (@gregorydwhite& @liemnguyen). Here is the edited transcript for your reading pleasure. Blogs: Leveraging Snapshots for Backup: An Expert View (http://www.nimblestorage.com/blog/technology/leveraging-snapshots-for-backup/): Radhika Krishnan interviews Jason Buffington (@JBuff) who is ESG’s Senior Analyst covering Data Protection. According to ESG’s research 55% of the IT shops are looking at snapshots augmenting their backups. Snapshots + Backup Management = the Best of Both Worlds (http://www.nimblestorage.com/blog/technology/snapshots-backup-management-thebest-of-both-worlds/): Another blog talking about need for integration between storage systems and their native snapshot capabilities and backup storage delivering rich data management functionality How Snappy and Skinny Are Your Snapshots? (http://www.nimblestorage.com/blog/technology/how-snappy-and-skinny-are-yoursnapshots/): UmeshMaheshwari (our CTO) talks about concept of COW versus ROW and discusses benefits of variable block support. A Comparison of File System Architectures (http://www.nimblestorage.com/blog/technology/a-comparison-of-filesystemarchitectures/): Another blog by Umesh. This one talks about the concept of keeping data optimized on disk–especially applicable if your want to know how should storage handle deleted snapshots. The comments at the bottom are worth reading.
  • 66. Extended Snapshots and Replication As Backup (http://www.nimblestorage.com/blog/technology/2160/): Ajay Singh discusses using snapshots and replication for deploying Disaster Recovery. Can you have a backup system based solely on snapshots and replication? (http://www.backupcentral.com/mr-backup-blog-mainmenu-47/13-mr-backup-blog/299snapshots-replication-backups.html/): A W. Curtis Preston special calling it as he sees it. The Nightmare of Incremental Backup is Over (http://www.nimblestorage.com/blog/technology/the-nightmare-of-incremental-backupis-over/): Nicholas Schoonover discusses concepts of RPO and RTO with incremental backups. Better Than Dedupe: Unduped! (http://www.nimblestorage.com/blog/technology/better-thandedupe-unduped/): Umesh shows a mathematically comparison of total storage space between different types of storage making the case for optimizing your entire storage environment. Be sure to skim through the comments at the bottom. This is a part one in a two part series. In the second blog, we’ll cover the concept of data integrity, discuss application integration and review the different demos covering data protection and high availability. We would love to hear from you. Follow us on Twitter (@NimbleStorage), send us a tweet (#nimblestorage #hybridstorage) or leave a note below.
  • 67. Targeting EMC and NetApp, Nimble Storage Flashing Towards an IPO Comment Now Follow Comments Image via CrunchBase Suresh Vasudevan is offering a very compelling proposition to companies trying to store and retrieve data. And that offer could make the investors in the venture of which he‘s CEO, Nimble Storage, even richer than they already are. In an August 10 interview, Vasudevan explained that Nimble sells storage that‘s a hybrid of traditional spinning disks and arrays of flash — the memory chips commonly used in digital cameras. Nimble‘s hybrid storage provides a good mix of high performance — rapid storage and retrieval – and lots of usable capacity. Nimble is going after a segment of the $20 billion network storage industry — and the figure excludes add-on software. Different competitors, such as EMC (EMC) and NetApp (NTAP) are the market leaders. Vasudevan says that Nimble‘s segment — including software – is worth $5 billion and it‘s growing at 15% annually. Nimble is winning a big piece of this market — getting ―600 customers in two years and 1,000 deployments.‖ Vasudevan told me that Nimble wins due to five customer benefits it provides over its competitors: More and faster storage for the buck. Two to five times more storage capacity and five to six times greater performance for the same amount of capital spending. More frequent backup. With a tiny increase in stored data, Nimble’s system lets customers back up their networks every 15 minutes — or as often as they wish — far more frequently than competing products. Quicker recovery. Nimble’s storage arrays let companies recover data in minutes compared to an hour for competitors’ products. Simplicity. While it can take four to eight hours to set up competitors’ products, Nimble customers are up and running in 20 to 30 minutes. Service. At five minute intervals, Nimble analyzes the health of its customers’ networks and if it identifies a current or potential problem — such as unusually high temperature in the data center — initiates a support call. Nimble spurs 75% of these calls to nip such problems in the bud.
  • 68. Customers buy from Nimble in two predominant order sizes. About 70% are orders for about $50,000 and the balance range between $100,000 and $150,000, according to Vasudevan. Vasudevan came into Nimble as CEO after a decade at NetApp. He oversaw ―product strategy/direction that helped NetApp triple‖ between 2001 and 2004. In 2006, as the leader of its Decru subsidiary, revenues grew from $6 million to $45 million in a year. He then joined Omneon, as CEO and led it to a successful acquisition by Harmonic. In 2008, Vasudevan was thinking about the next generation of storage and how flash should be incorporated into the product. He was talking with companies working on that problem and met with Nimble — concluding that its approach would be his approach if he were starting a storage company from scratch. Shortly thereafter, he joined Nimble‘s board. The board ended up hiring Vasudevan as CEO, and he‘s happy to report that Nimble‘s founders, Varun Mehta and UmeshMaheshwari, remain as VP of product development and CTO, respectively. Meanwhile, Nimble has had some help from the venture capital community. It has raised $58 million in total — including $25 million in a Series D round in July 2011 from Sequoia Capital and Accel Partners, among others. When Vasudevan joined in January 2011, Nimble had 40 people — it now has 230. And its headcount is growing fast — he reports that Nimble added 43 people in the second quarter of 2012, alone. And Nimble is hiring sales and engineering people who are highly talented, fit with its culture, and have a record of taking risks. In the last couple of quarters, Nimble has been changing gradually the mix of its people. In particular, as its business model became more established, Nimble‘s risk level declined accordingly. Now Nimble is adding people who have experience growing a more established venture quickly. By the end of 2014, Vasudevan expects Nimble to be ready for an IPO. In his view, that means generating trailing revenues of $25 million to $35 million a quarter and sprinting on a path to achieve ―breakeven within one to two quarters.‖ Vasudevan does not expect Nimble to be acquired — at least not before its IPO. Vasudevan is not concerned about macroeconomic factors and believes that its market is big enough to achieve these goals. He is focusing on keeping Nimble on its three-fold year-over-year growth in the number of customers and revenue. In addition to boosting his sales and service staff, Vasudevan recently introduced a new set of capabilities that offer customers more reasons to buy from Nimble. Specifically, Nimble‘s newest product helps customers expand their storage capacity more flexibly than do competitors‘.
  • 69. For example, if they need more storage, say, for massive databases, they can add ‖new expansion shelves;‖ if they need greater performance, they can add ―a new controller or more flash memory.‖ ―Competitors force customers to increase performance and capacity in tandem,‖ Vasudevan says, ―saddling them with a huge capital investment long before they need it.‖ Nimble appears to be outrunning its competitors and if it can keep adding customers while staying close to breakeven, its investors will become even richer.
  • 70. “Gotcha” with VMware Site Recovery Manager Posted onOctober 10, 2012bywenjyu I have recently been playing a lot more with VMware‘s Site Recovery Manager (SRM) product – as the Nimble SRA EA is out to customers, we have gotten quite a few inquiries on WTF moments (in other words, ―gotchas‖)…so here they are, in order of installation, configuration, test failover & failover. Gotcha 1 During installation, you will get prompted for ―Local Host‖ – several folks had entered the ESX host IP address! DO NOTdo that or else you will NOT be able to download the SRM plugin for vCenter! Installation will proceed without any sort of warning that you shot yourself in the foot, so it‘s best to do it right during installation. Select ―localhost‖ or the IP that shows up and don‘t be a smart ass Gotcha 2 Post installation, with ―Inventory Mapping‖ workflow, you will get prompted to configure ―Placeholder‖ datastore for each site. This is the datastore that is needed to temporarily store the .vmx file, so a nice little ―shadow‖ VM could show up in the recovery site, to give you a peace of mind.
  • 71. To make life easier, provision shared storage to the ESX cluster participating in DR test/recovery. Though you could use local datastore, but you‘d have to configure local datastore for each ESX host. Why make it so messy when you could just have it centralized, with a meaningful name. I find it easiest to name it as <placeholder_remotesite_name> because the placeholder datastore for the recovery site is used to hold the shadow VMs from the protected site. This is what my setup looks like‖ it‘s a thin provisioned volume, 1GB in size, with no snapshot/replication schedule (NOT needed), and named in a way that can easily be correlated. Gotcha 3 During Array Manager configuration, don‘t forget to enable the array pairs, after you have configured array manager for each site! If everything is greyed out during ―Protection Group‖ creation, then the most likely cause is the array pairs have not been enabled – so don‘t forget this important step! Oh one more thing, when you have a protection group created, you cannot disable the array pairs. Gotcha 4
  • 72. During Protection Group creation, if you replicated volume does not show up in the list, be sure to ask yourself: 1)is the volume part of a volume collection? (in non-nimble terms, is it part of a consistency group of volumes configured with replication?) 2)is the volume collection configured with snapshot + replication schedule? 3)is the volume provisioned to ESX server cluster, with VMFS volume formatted? Gotcha 5 After new VMs get deployed to the replicated volume used for SRM DR test/recovery, don‘t forget to configure protection (as in walking through a quick wizard check on which folder, resource pool, network, datastore the VM will live, on the recovery site). It‘s basically read only info, as the global settings are set at the site level (in ―Resource Mapping‖ step Gotcha 6 Last but not least, if you use in-guest/direct attached storage for your VM (i.e., SQL server with OS disk on VMDK, DB & log disks on direct mounted iSCSI storage within the guest), then the workflow is NOT very ‗out-of-box‘ with SRM. Below are the high level checklist/steps you have to keep in mind (of course, this can certainly be automated to a script that can be invoked from the SRM server – stay tuned for that one ) For SRM Recovery Plan
  • 73. Below is an example extra step that you could add, to the test and recovery steps. After the array SRA has promoted the replica volume to primary during DR, add a prompt in the recovery plan to ensure the volume ACL is updated with the initiator group for the respective in-guest sw/iscsi initiator: And here‘s the other extra step you could add, after the VM power on step by SRM, to configure the in-guest attached storage for each VM that has direct attached storage:
  • 74. For array & in-guest initiator within guest efore test failover/failover: -ensure the direct attached volumes are on the same consistency group (in Nimble‘s terms, same volume collection) -ensure the initiator group on the array side is created with IQNs for the in-guest iSCSI initiators (this will come in handy during test recovery/recovery) During test failover/failover: -update in-guest initiator with remote site array discovery IP
  • 75. -update direct attached volume ACL with the IQN group containing the IQN for the in-guest initiators -rescan/‘favorite‘ the newly discovered replicated storage
  • 76. That‘s it for now – I will post more findings as I get to do full blown recovery, reprotect, failback scenarios. Stay tuned.