Slide 1 - common.es

  • 470 views
Uploaded on

 

More in: Technology
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Be the first to comment
    Be the first to like this
No Downloads

Views

Total Views
470
On Slideshare
0
From Embeds
0
Number of Embeds
0

Actions

Shares
Downloads
16
Comments
0
Likes
0

Embeds 0

No embeds

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
    No notes for slide
  • The capability of IBM i to providevirtual I/O resources has been used successfully for several years to integrate AIX®, Linux® and Windows® workloads on the same platform. The same virtualization technology, which is part ofthe IBM i operating system, can now be used to host IBM i LPARs. An IBM i 6.1 LPAR can host one or more additional IBM i LPARs, known as virtual client LPARs. Virtual client partitions can have no physical I/O hardware assigned and instead leverage virtual I/O resources from the host IBM i partition. The types of hardware resources that can bevirtualized by the host LPAR are disk, optical and networking.
  • The capability of an IBM i partition to host another IBM i partition involves hardware and virtualization components. The hardware components are the storage, optical and network adapters and devices physically assigned to the host IBM i LPAR. The virtualization components are the system firmware and IBM i operating system objects necessary to virtualize the physical I/O resources to client partitions. Virtual SCSI and Ethernet adapters IBM i hosting IBM i uses an existing function of the system firmware, or Power Hypervisor: the capability to create virtual SCSI and Ethernet adapters in a partition. Virtual adapters are created for each LPAR in the Hardware Management Console (HMC). Virtual SCSI adapters are used for storage and optical virtualization ; virtual Ethernet adapters are used for network virtualization. A mix of virtual and physical hardware in the same partition is supported for IBM i in this environment, by assigning both types of adapters to the partition in the HMC.
  • To virtualize integrated disk (SCSI or SAS) or LUNs from a SAN system to an IBM i client partition, both HMC and IBM i objects must be created. In the HMC, the minimum required configuration is: • One virtual SCSI client adapter in the client partition
  • One virtual SCSI server adapter in the host partition
  • In the host partition, the minimum required IBM i setup consists of the following: • One Network Server Description (NWSD) object • One Network Server Storage Space (NWSSTG) object The NWSD object associates a virtual SCSI server adapter in IBM i (which in turn is connected to a virtual SCSI client adapter in the HMC) with one or more NWSSTG objects. At least one NWSD object must be created in the host for each client, though more are supported. The NWSSTG objects are the virtual disks provided to the client IBM i partition. They are created from available physical storage in the host partition. In the client, they are recognized and managed as standard DDxx disk devices (with a different type and model). The screenshot shows several storage spaces for a client partition in an IBM i 6.1 host partition
  • Storage spaces for an IBM i client partition do not have to match physical disk sizes; they can be created from 160 MB to 1 TB in size , as long as there is available storage in the host. The 160 MB minimum size is a requirement from the storage management Licensed Internal Code (LIC) on the client partition. For an IBM i client partition, up to 16 NWSSTGs can be linked to a single NWSD , and therefore, to a single virtual SCSI connection. Storage planning Because virtual disks for the IBM i client LPAR are NWSSTG objects in the host LPAR, the main prerequisite to installing a new client LPAR is having sufficient capacity in the host to create those objects. Note that the host partition is not capable of detecting what percent of the virtual storage is used in the client. For example, if a 500-GB storage space is created, it will occupy that amount of physical storage in the host IBM i LPAR, even if the disk capacity is only 50% utilized in the client LPAR. It is recommended to closely match the total size of the storage spaces for each client partition to its initial disk requirements. As the storage needs of the client partition grow, additional storage spaces can be dynamically created and linked to it on the host partition. On the client, the new virtual disk will automatically be recognized as a non-configured drive and can be added to any existing ASP. The only restriction to consider in this case is the maximum number of storage spaces allowed per virtual SCSI connection for an IBM i client partition, which is 16. If more than 16 NWSSTGs are needed for a client LPAR, additional virtual SCSI connections can be created dynamically in the HMC.
  • IBM i virtual client partitions can also be hosted by VIOS. VIOS is virtualization software that runs in a separate partition whose purpose is to provide virtual storage, optical and networking resources to one or more client partitions. The most immediate benefit VIOS brings to an IBM i client partition is the ability to expand its storage portfolio and use DS4000 open storage. DS4000 logical units (LUNs) are physically attached to VIOS via a Fibre Channel connection and then made available to IBM i. While IBM i does not directly attach to the SAN in this case, once DS4000 LUNs become available through VIOS, they are managed the same way as integrated disks or LUNs from a DS8000 storage system. VIOS has been used successfully for several years to virtualize storage, optical and networking resources to AIX and Linux client partitions. Now IBM i joins this virtualization environment, gaining the ability to use DS4000 storage. From a VIOS perspective, IBM i is another client partition; the host-side configuration steps are the same as for AIX and Linux clients. While code changes were made to VIOS to accommodate IBM i client partitions, there is no special version of VIOS in use for IBM i. If you have existing skills in attaching DS4000 to VIOS and virtualizing I/O resources to client partitions, they will continue to prove useful when creating a configuration for an IBM i client partition.
  • The Virtual I/O Server is an appliance that provides virtual storage and shared Ethernet adapter capability to client logical partitions on POWER5 systems. It allows a physical adapter with attached disks on the Virtual I/O Server partition to be shared by one or more partitions, enabling clients to consolidate and potentially minimize the number of physical adapters required. What is virtual SCSI(VSCSI)? Virtual SCSI is based on a client and server relationship. The Virtual I/O Server owns the physical resources and acts as server, or target, device. Physical adapters with attached disks on the Virtual I/O Server partition may be shared by one or more partitions. These partitions contain a virtual SCSI client adapter that sees these virtual devices as standard SCSI compliant devices and LUNs. Creating more logical partitions than there are I/O slots or physical devices available What is the shared Ethernet adapter (SEA)? A shared Ethernet adapter is a bridge between a physical Ethernet adapter or link aggregation and one or more virtual Ethernet adapters on the Virtual I/O Server. A shared Ethernet adapter enables logical partitions on the virtual Ethernet to share access to the physical Ethernet and communicate with stand-alone servers and logical partitions on other systems. The shared Ethernet adapter provides this access by connecting the internal VLANs with the VLANs on the external switches.
  • The capability to attach to DS4000 through VIOS extends the IBM i storage portfolio to include 512-byte-per-sector open storage. The existing IBM i storage portfolio includes integrated SCSI or SAS disk, as well as Fibre Channel-attached storage subsystems that support 520 bytes per sector, such as the IBM DS8000 product line. IBM i cannot currently attach directly to DS4000, because its Fibre Channel adapter code requires 520 bytes per sector. Therefore, logical units (LUNs) created on the DS4000 are physically connected through the Fibre Channel fabric to VIOS. Once recognized by VIOS, the LUNs are virtualized to one or more IBM i virtual client partitions.

Transcript

  • 1. i5/OS V6R1 Virtual Client Partitions VIOS, Blades and DS4000 Storage
  • 2. Agenda
    • Overview: Virtualization enhancements for i5/OS
    • i5/OS host & i5/OS client
    • VIOS host, i5/OS client & DS4000 attachment
  • 3. Virtualization Enhancements for i5/OS
  • 4. i5/OS Virtual Client Partitions
    • i5/OS-based Virtualization
      • i5/OS partition uses I/O resources from another i5/OS partition
      • Eliminates requirement to buy adapters and disk drives for each i5/OS partition
      • Supports simple creation of additional partitions …. e.g., for test and development
      • Requires POWER6 systems with i5/OS V6R1
      • PowerVM not required
      • Can mix virtual and direct I/O in client
    • Platform support
      • All POWER6 System i models* ( non-blade )
    • Storage support
      • Determined by host i5/OS partition (SAN, EXP24, integrated disk)
    • LPAR management
      • HMC
    i5/OS Hypervisor i5/OS POWER6 * All statements regarding IBM's future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only.
  • 5. PowerVM Virtualization i5/OS VIOS Hypervisor POWER6
    • VIOS-Based Virtualization
      • i5/OS partition uses I/O resources from Virtual I/O Server (VIOS)
      • VIOS is included with PowerVM Standard and Enterprise Edition
      • VIOS also provides virtualization for AIX and Linux partitions
      • Requires POWER6 systems with i5/OS V6R1
    • Platform support
      • All POWER6 System i models
      • POWER6 Blade models
    • Storage support
      • Enables attachment to DS4700 y DS4800
      • Models DS8000 (via Blade only)
    • LPAR management
      • HMC or IVM*
    * 570 and 595 systems require an HMC
  • 6. Integrated Virtualization Manager (IVM) * 570 and 595 systems require an HMC
    • Tool that runs in VIOS partition for creating and managing partitions
    • Provides an alternative to HMC for LPAR management*
    • Requires VIOS to own i5/OS I/O resources
    • Included with PowerVM Standard Edition
    • Requires POWER6 processor-based systems with i5/OS V6R1
  • 7.
    • Browser-based interface, supports Mozilla Firefox and Internet Explorer
    • Performs LPAR and virtualization management on POWER6 blade
    Integrated Virtualization Manager
  • 8. IVM Example: Create i5/OS Partition
    • Fewer steps than HMC
    • IVM uses several defaults
    • Virtual I/O resources only
  • 9. IVM Example: i5/OS Partition Properties
    • Load source and alternate IPL devices default to virtual
    • Changes are always DLPAR changes (when possible)
    • Virtual disk assigned to i5/OS by selecting available hdisk, then clicking OK (all available hdisks are assigned in this example)
  • 10. Virtualization Scenarios
    • Hardware platform: POWER6 System i (not POWER6 blade)
    • Host LPAR: i5/OS
    • LPAR management: HMC
    • Storage: Any supported by i5/OS host
    • Hardware platform: POWER6 System i
    • Host LPAR: VIOS
    • LPAR management: HMC
    • Storage: DS4000 ( not DS8000 )
    • Hardware platform: POWER6 blade
    • Host LPAR: VIOS
    • LPAR management: IVM (part of VIOS)
    • Storage: Fibre Channel or SAS* (SOD)
    * All statements regarding IBM's future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only. + + + Host LPAR Client LPAR POWER6 system HMC + + + VIOS Host LPAR Client LPAR POWER6 system HMC + DS4000 + VIOS + IVM Client LPAR + POWER6 Blade FC or SAS +
  • 11. i5/OS Client with i5/OS Host
  • 12. i5/OS V6R1 Host and Client Partitions: Overview i5/OS Host i5/OS Client FC 5294 EXP24 DS8000 DDxx NWSSTGs DDxx OPTxx OPTxx IVE CMNxx
    • Requirements
      • POWER6 hardware
      • V6R1 on host and client
      • PowerVM not required
    • DASD
      • Hardware assigned to host LPAR in HMC
      • DASD can be integrated or SAN
      • DASD virtualized as NWSSTG objects
    • Optical
      • DVD drive in host LPAR virtualized directly (OPTxx)
    • Networking
      • Network adapter (such as IVE) and Virtual Ethernet adapter in host LPAR
      • Virtual Ethernet adapter in client LPAR
    • Virtualization of Tape Devices is not supported
    Virtual SCSI connection DVD DVD Virtual LAN connection
  • 13. Host LPAR Configuration – HMC
    • Virtual SCSI server adapter
    • At least 1 required per client LPAR, but more possible
    • Configured to connect to specific adapter ID on client LPAR
  • 14. Client LPAR Configuration – HMC
    • Virtual SCSI client adapter
    • DASD and optical are accessible through the same VSCSI client adapter
    • By using multiple adapters, a virtual client LPAR can use DASD from multiple hosts
  • 15. Client LPAR Configuration – Load Source
    • Both B- and D-mode IPL devices are set to virtual SCSI client adapters
    • When installing virtual client LPAR, D-mode IPL is performed from DVD drive in host LPAR
    • LIC and OS are installed on NWSSTG (Network server storage space) objects
  • 16. Host LPAR Configuration – i5/OS View
    • Virtual SCSI server adapter in i5/OS (290B device)
  • 17. Host LPAR Configuration – Storage Spaces
    • Storage space objects in host LPAR
    • Each NWSSTG is a DDxx in client LPAR
  • 18. Host LPAR Configuration – Storage Spaces
    • Creating a storage space
    • Identical to creating a storage space for AIX or Linux client today
  • 19. Host LPAR Configuration – NWSDs
    • WRKCFGSTS *NWS provides list of NWSD (Network Server Description) objects
    • Each client LPAR has at least 1 NWSD associated with it
    • NWSD provides link between storage space object and VSCSI adapters
  • 20. Installing Client Partition from IMGCLG
    • An image catalog can be used to install multiple virtual client LPARs or PTFs
  • 21. Client LPAR Configuration – Configured DASD
    • Virtual client LPAR’s System ASP with 4 virtual disks (storage spaces)
  • 22. Backups for i5/OS V6R1 Client with i5/OS Host
    • Simplest approach is to use Dynamic LPAR (DLPAR) resource movement and switch physical tape adapter to client LPAR
      • Mixing of virtual and direct resources in client is supported
      • DLPAR movement of resources can be scheduled in the HMC
    • For full-system backup, the client storage spaces can be saved on the host i5/OS partition
      • Similar to AIX and Linux client partitions
      • File-level backup is not supported
      • Storage spaces can be restored on another i5/OS V6R1 host
      • Storage spaces can be located in IASP, Flash Copy can be used on IASP
  • 23. i5/OS Client with VIOS Host and DS4000 Attachment
  • 24. System i VIOS – Virtual I/O Server Partition VIOS Host AIX or Linux I5/OS
    • Pool of physical I/O resources to share between logical partitions on the system
    • Creating more logical partitions than there are I/O slots or physical devices available
      • AIX, Linux Client Partitions
      • Now 5/OS Clients Partitions
    • Dedicated partition for VIOS
    • Virtual SCSI for Disk (VSCSI)
    • Virtual Networking for LAN adapters (SEA = Shared Ethernet adapters)
    • Implemented via SW in a Logical Partition
    • Managed by HMC/IVM
    Physical Disk Physical Disk Physical Disk Physical Disk Physical Disk
  • 25. i5/OS V6R1 Client + VIOS Host + DS4000 VIOS Host i5/OS Client DS4700 DS4800 hdiskX LUNs DDxx /dev/cd0 OPTxx IVE CMNxx
    • DASD
      • Hardware assigned to VIOS LPAR in HMC
      • Each SAN LUN virtualized directly to i5/OS client
      • LUNs are 512B open storage
    • Optical
      • DVD drive in VIOS virtualized directly (OPTxx)
    • Networking
      • i5/OS client LPAR is assigned a Virtual Ethernet adapter
      • VIOS uses virtual Ethernet bridge to connect VLAN and physical LAN
    SEA System i POWER6 platform or JS22 Blade Virtual SCSI connection DVD DVD Virtual LAN connection
  • 26. VIOS Storage Connectivity Big Picture POWER Hypervisor AIX or Linux VIOS Partition Physical Fibre Channel Adapter Virtual SCSI Server Adapter Virtual SCSI Client Adapter i5/OS Virtual SCSI Client Adapter Physical Fiber Channel Adapter Virtual SCSI Server Adapter DS4700DS4800 SVC SAN Fibre Channel SAN Not supported With i5/OS
  • 27. 520 sector to 512 sector change
    • i5/OS traditionally have been 520 byte sectors
    • Now ONLY with i5/OS clients using VIOS we have 512 byte sectors
    • This is not supported for direct FC attach to i5/OS
    • SLIC knows whether the disk is 520 or 512
      • Decision is made as the disk units enlist to the i5/OS
    • 520 sectors fit on 8 pages
    • 512 sectors fit on 9 pages
      • SLIC simply reads 9x 512 byte data sectors for every 8 actual sectors
      • 9 th page needed for the extra 8 bytes
      • This may cause some performance impact
  • 28. Requirements and Support for i5/OS + VIOS + DS4000
    • Server Side
      • i5/OS cannot connect directly to DS4000
      • POWER6 hardware is required
      • VIOS 1.5 with Fix Pack 10.1 or later is required
      • i5/OS V6R1 or later is required
    • Attached Storage
      • Only DS4700 and DS4800
        • Only Fibre Channel drives are supported
        • For both POWER6 Blade and other POWER6 System i models
      • DS8100 and DS8300 are supported only in the POWER blade environment
        • Not for other POWER6 System i models
      • SAN Volume Controller (SVC) is not supported by i5/OS
        • Directly or through VIOS
      • DS4000 FlashCopy
        • Not currently supported, testing underway
    • System Storage Manager
      • Initially: System Manager 6.6.0.x is supported
      • Later: System Manager 10.10.x will be supported 
  • 29. The DS4000 Family … Control Unit (no drives) + Up to 14 expansion units … Control Unit + up to 6 expansion units … Control Unit + up to 6 expansion units Supported on i5/OS on Power 6 via VIOS Control Unit + 14 EXP810 Control Unit + 6 EXP810 Control Unit + 6 EXP810 Enclosures 1814-80x,-82x,-84x,-86x 1814-70x,-72x 1814-7Vx Machine/Models DS4800 (follow-on to DS4500) DS4700 (follow-on to DS4300) DS4200 (follow-on to DS4100)
  • 30. The DS4000 Family Supported on i5/OS on Power 6 via VIOS 0,1,3,5,7 0,1,3,5,7 0,1,3,5,7 RAID levels 1240-1600 MB/sec 990 MB/sec 990 MB/sec
    • Performance - MB/sec
    • - Disk IO/s
    • - Cache Disk IO/s
    62,000-86,000 io/sec 44,000 io/sec 11,200 io/sec Mdl 80/82: 4 GB Mdl 70: 2 GB, Mdl 72: 4 GB 2 GB Cache 168-224 TB 84-112 TB 84-112 TB Total Disk - SATA - Fibre Mdl 84: 8 GB, Mdl 86 – 16 GB 512 128 128 Max DS4K Partitions (no RAID-6) RAID 6 with SM 10.10 RAID 6 with SM 10.10 375,000-575,000 io/sec 120,000 io/sec 120,000 io/sec 67-224 TB 33-112 TB n/a 2048 1024 1024 Max Logical Drives (LUNs) 224 drives 112 drives 112 drives Max Physical Drives SATA or FC ** only Fibre Channel for i5/OS SATA or FC ** only Fibre Channel for i5/OS SATA only ** not supported for i5/OS Drive Types 1814-80x,-82x,-84x,-86x 1814-70x,-72x 1814-7Vx Machine/Models DS4800 (follow-on to DS4500) DS4700 (follow-on to DS4300) DS4200 (follow-on to DS4100)
  • 31. The System Manager GUI Interface – Create a LUN 1) Choose the DS4000 to Use 2) Choose to Create LUNs 3) Choose automatic or manual configuration 4) Check the parameters 5) Tadaa!
  • 32. Planning i5/OS clients via VIOS to DS4700/DS4800
    • Configurations
      • Fiber channel adapters are 4GB and 2GB
        • 4GB is recommended
    • Modeling and sizing
      • You must use Disk Magic
        • VIOS disk information is taken into account within Disk magic
      • Performance Considerations
    • Planning
      • Check the i5/OS Virtualization and DS4000 Read-me First
      • http://www-3.ibm.com/systems/i/os/i_virtualization_and_ds4000_readme.pdf
  • 33. DS4000 Sizing with Disk Magic Performance Tools Reports Workload Estimator Input Manual Input of Disk Workload 1) Gather Input Data 2) Start Disk Magic 3) Read in or Key in Base System Data 4) Create the Base System Model
    • To get your Copy of Disk Magic:
    • Go to System Sales or Partnerworld Website
    • Search on “Disk Magic”
  • 34. DS4000 Sizing with Disk Magic Performance Tools Reports Workload Estimator Input Manual Input of Disk Workload 1) Gather Input Data 2) Start Disk Magic 3) Read in or Key in Base System Data 4) Create the Base System Model 5) Describe Proposed Hardware 6) “Solve” the Model 7) Print Results – Reports + Graphs showing Response Time + Growth Projections
    • To get your Copy of Disk Magic:
    • Go to System Sales or Partnerworld Website
    • Search on “Disk Magic”
  • 35. DS4000 Performance on System i – DS4000 Setup
    • DS4000 Setup
    • Set up RAID as close to the physical drive as possible (ie on the DS4000)
    • Make a reasonable number of LUNs, sized for 1-2 LUNs per DDM
    • Make symmetrical configurations
      • Uniform LUN size
      • RAID sets all the same size
      • RAID sets balanced across enclosures
    • Avoid sharing physical disks between i5/OS hosts
    • Create a separate VIOS for every 3 Power 6 processors or so
    For details, see Chapter 14.5 of the i5/OS Performance Capabilities Reference Manual (PCRM) at: http://publib.boulder.ibm.com/infocenter/systems/scope/i5os/index.jsp?topic=/books/sc410607.pdf DS4000 System Manager Remote Support Manager
  • 36. DS4000 Performance on System i – VIOS Setup
    • VIOS Setup
    • Avoid VIOS volume groups – they add overhead without any benefit for our platform. Use commands to set the volume indentifier instead
    • Max 16 virtual disks on a virtual SCSI adapter
    • Use dedicated vs shared processors on VIOS if possible
    • VIOS appears to need 1-2 GB of memory for a 1-CPU VIOS partition, more if virtual LAN is in use
    For details, see Chapter 14.5 of the i5/OS Performance Capabilities Reference Manual (PCRM) at: http://publib.boulder.ibm.com/infocenter/systems/scope/i5os/index.jsp?topic=/books/sc410607.pdf
  • 37. DS4000 Performance on System i – System i Setup
    • If you need the best performance possible:
    • Consider
      • Integrated disk first
      • Native external disk next (DS6K/DS8K attached via 4 Gbit fibre)
      • VIOS configurations last – throughput is OK but response time can be hard to control
    • Choose RAID-1 or RAID-10 over RAID-5
    • i5/OS Setup
    • Separate Journal Receivers from Databases – put them on separate DDMs and separate DS4K controllers if possible
    For details, see Chapter 14.5 of the i5/OS Performance Capabilities Reference Manual (PCRM) at: http://publib.boulder.ibm.com/infocenter/systems/scope/i5os/index.jsp?topic=/books/sc410607.pdf
  • 38. VIOS Configuration
    • Partition type is VIOS (not AIX)
    • Physical storage adapter assigned to VIOS
    • At least one Virtual SCSI adapter pair (server-client) exists per i5/OS client
    • Detailed VIOS information in APV Operations Guide ( https://www.ibm.com/servers/resourcelink/lib03030.nsf/pages/AdvancedPowerVirtualizationOperationsGuide )
  • 39. VIOS Installation
    • Install similar to AIX
    • Use HMC console and DVD drive or Network Install Manager (NIM) server
    • Detailed VIOS information in APV Operations Guide ( https://www.ibm.com/servers/resourcelink/lib03030.nsf/pages/AdvancedPowerVirtualizationOperationsGuide )
  • 40. VIOS Virtual Resource Configuration
    • Once physical devices are available in VIOS, they can be configured as virtual resources for client partitions
      • Example: physical LUN hdiskX can be mapped as virtual disk for i5/OS by creating a virtual target device ( vtscsiX )
    • VIOS command line is used
      • Telnet to VIOS; default administrator ID is padmin
    • VIOS command examples
      • lsdev: shows physical and virtual devices
      • cfgdev: discovers newly added devices (such as new LUNs)
      • mkvdev: creates virtual devices to make resources available to client partitions
    • VIOS commands to virtualize resources are well documented
      • IBM Systems Information Center ( http://publib.boulder.ibm.com/infocenter/systems/scope/hw/index.jsp?topic=/iphb1/iphb1_vios_commandslist.htm)
      • APV Operations Guide ( https://www.ibm.com/servers/resourcelink/lib03030.nsf/pages/AdvancedPowerVirtualizationOperationsGuide
  • 41. i5/OS Client HMC Configuration
    • Virtual SCSI client adapter
    • DASD and optical are accessible through the same VSCSI client adapter
    • By using multiple adapters, a virtual client LPAR can use DASD from multiple hosts
  • 42. i5/OS View of VIOS Resources
    • Virtual storage IOA is the VSCSI client adapter
    • Disk units are virtual target devices mapped to LUNs in VIOS
  • 43. i5/OS View of VIOS/DS4K LUNs Serial # comes from VIOS Type is always “6B22” Any LUN Size is acceptable – no need to match integrated drive sizes Virtual disks (LUNs) from VIOS initially show as non-configured drives (DDxx) Virtual disks have been added to ASP1 in this example
  • 44. Backups for i5/OS as Client of VIOS
    • Simplest approach is to use Dynamic LPAR (DLPAR) resource movement and switch physical tape adapter to i5/OS client
      • Mixing of virtual and direct resources in i5/OS client is supported
        • Does not apply to POWER blade environment
      • DLPAR movement of resources can be scheduled in the HMC
    • Other backup solutions exist
      • i5/OS save to virtual tape IMGCLG; FTP IMGCLG files to another i5/OS partition
      • i5/OS save to virtual optical IMGCLG, FTP images to VIOS or another i5/OS
      • i5/OS save to file-backed virtual optical in VIOS; VIOS TSM client save over network
      • i5/OS save to file-backed virtual optical in VIOS; FlashCopy VIOS virtual media LUNs and attach to 2nd VIOS
      • FlashCopy (full copy) VIOS LUNs assigned to i5/OS; bring up separate i5/OS partition under VIOS from copy
      • Replicate VIOS LUNs assigned to i5/OS; bring up separate i5/OS partition under VIOS from copy
  • 45. VIOS HA Considerations
    • NO High Availability solution
    • I5/OS partition depend on VIOS partition availability & DS4000
    vSCSI Target Device FC or SCSI vSCSI Server Adapter vSCSI Client Adapter disk Power6 Hypervisor I5 partition Virtual IO server Adapter Disk driver DS4700/4800
  • 46. VIOS HA Considerations
    • I5/OS disk in mirroring for i5/OS protection
    • I5/OS partition not dependent from a VIOS
    • Single Point of failure is system i & DS4000
    FC or SCSI Device vSCSI Target Device vSCSI Server vSCSI Client disk Power6 Hypervisor I5 partition Virtual IO server Adapter Disk driver DS4700/4800 Virtual IO server vSCSI Target Device Adapter Disk driver disk vSCSI Client vSCSI Server i5Mirorring
  • 47. VIOS HA Considerations
    • I5/OS mirroring with one disk of the disk pair on a DS4000 and the other disk on another DS4000
    • No replication between DS4000
    • Avoid Single point of Failure for the DS4000
    FC or SCSI vSCSI Target Device vSCSI Server vSCSI Client disk Power6 Hypervisor I5 partition Virtual IO server Adapter Disk driver DS4700/4800 Virtual IO server vSCSI Target Device Adapter Disk driver disk vSCSI Client vSCSI Server FC or SCSI DS4700/4800 Adapter Disk driver Adapter Disk driver Multi-path Driver Multi-path Driver i5Mirorring
  • 48. More Information
    • APV Operations Guide (VIOS information)
      • https://www.ibm.com/servers/resourcelink/lib03030.nsf/pages/AdvancedPowerVirtualizationOperationsGuide
    • i5/OS Virtual Client Partitions and DS4000 Storage Read-me First document
      • Scheduled for availability by 3/21, URL not available yet
    • Performance Capability Reference manual (Chapter 14)
      • http://publib.boulder.ibm.com/infocenter/systems/scope/i5os/index.jsp?topic=/books/sc410607.pdf
    • Redbook IBM System Storage DS4000 and Storage Manager V10.10
      • http://www.redbooks.ibm.com/redpieces/abstracts/sg247010.html?Open
    • VIOS datasheet (VIOS support only, refer to this presentation for i5/OS + VIOS support)
      • http://www14.software.ibm.com/webapp/set2/sas/f/vios/documentation/datasheet.html
  • 49. Special Notices This document was developed for IBM offerings in the United States as of the date of publication. IBM may not make these offerings available in other countries, and the information is subject to change without notice. Consult your local IBM business contact for information on the IBM offerings available in your area. Information in this document concerning non-IBM products was obtained from the suppliers of these products or other public sources. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. IBM may have patents or pending patent applications covering subject matter in this document. The furnishing of this document does not give you any license to these patents. Send license inquires, in writing, to IBM Director of Licensing, IBM Corporation, New Castle Drive, Armonk, NY 10504-1785 USA. All statements regarding IBM future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only. The information contained in this document has not been submitted to any formal IBM test and is provided "AS IS" with no warranties or guarantees either expressed or implied. All examples cited or described in this document are presented as illustrations of the manner in which some IBM products can be used and the results that may be achieved. Actual environmental costs and performance characteristics will vary depending on individual client configurations and conditions. IBM Global Financing offerings are provided through IBM Credit Corporation in the United States and other IBM subsidiaries and divisions worldwide to qualified commercial and government clients. Rates are based on a client's credit rating, financing terms, offering type, equipment type and options, and may vary by country. Other restrictions may apply. Rates and offerings are subject to change, extension or withdrawal without notice. IBM is not responsible for printing errors in this document that result in pricing or information inaccuracies. All prices shown are IBM's United States suggested list prices and are subject to change without notice; reseller prices may vary. IBM hardware products are manufactured from new parts, or new and serviceable used parts. Regardless, our warranty terms apply. Any performance data contained in this document was determined in a controlled environment. Actual results may vary significantly and are dependent on many factors including system hardware configuration and software design and configuration. Some measurements quoted in this document may have been made on development-level systems. There is no guarantee these measurements will be the same on generally-available systems. Some measurements quoted in this document may have been estimated through extrapolation. Users of this document should verify the applicable data for their specific environment. Revised September 26, 2006
  • 50. Special Notices (Cont.) The following terms are registered trademarks of International Business Machines Corporation in the United States and/or other countries: AIX, AIX/L, AIX/L (logo), AIX 6 (logo), alphaWorks, AS/400, BladeCenter, Blue Gene, Blue Lightning, C Set++, CICS, CICS/6000, ClusterProven, CT/2, DataHub, DataJoiner, DB2, DEEP BLUE, developerWorks, DirectTalk, Domino, DYNIX, DYNIX/ptx, e business (logo), e(logo)business, e(logo)server, Enterprise Storage Server, ESCON, FlashCopy, GDDM, i5/OS, i5/OS (logo), IBM, IBM (logo), ibm.com, IBM Business Partner (logo), Informix, IntelliStation, IQ-Link, LANStreamer, LoadLeveler, Lotus, Lotus Notes, Lotusphere, Magstar, MediaStreamer, Micro Channel, MQSeries, Net.Data, Netfinity, NetView, Network Station, Notes, NUMA-Q, OpenPower, Operating System/2, Operating System/400, OS/2, OS/390, OS/400, Parallel Sysplex, PartnerLink, PartnerWorld, Passport Advantage, POWERparallel, Power PC 603, Power PC 604, PowerPC, PowerPC (logo), Predictive Failure Analysis, pSeries, PTX, ptx/ADMIN, Quick Place, Rational, RETAIN, RISC System/6000, RS/6000, RT Personal Computer, S/390, Sametime, Scalable POWERparallel Systems, SecureWay, Sequent, ServerProven, SpaceBall, System/390, The Engines of e-business, THINK, Tivoli, Tivoli (logo), Tivoli Management Environment, Tivoli Ready (logo), TME, TotalStorage, TURBOWAYS, VisualAge, WebSphere, xSeries, z/OS, zSeries. The following terms are trademarks of International Business Machines Corporation in the United States and/or other countries: Advanced Micro-Partitioning, AIX 5L, AIX PVMe, AS/400e, Calibrated Vectored Cooling, Chiphopper, Chipkill, Cloudscape, DataPower, DB2 OLAP Server, DB2 Universal Database, DFDSM, DFSORT, DS4000, DS6000, DS8000, e-business (logo), e-business on demand, EnergyScale, Enterprise Workload Manager, eServer, Express Middleware, Express Portfolio, Express Servers, Express Servers and Storage, General Purpose File System, GigaProcessor, GPFS, HACMP, HACMP/6000, IBM Systems Director Active Energy Manager, IBM TotalStorage Proven, IBMLink, IMS, Intelligent Miner, iSeries, Micro-Partitioning, NUMACenter, On Demand Business logo, POWER, PowerExecutive, PowerVM, PowerVM (logo), Power Architecture, Power Everywhere, Power Family, POWER Hypervisor, Power PC, Power Systems, Power Systems (logo), Power Systems Software, Power Systems Software (logo), PowerPC Architecture, PowerPC 603, PowerPC 603e, PowerPC 604, PowerPC 750, POWER2, POWER2 Architecture, POWER3, POWER4, POWER4+, POWER5, POWER5+, POWER6, POWER6+, pure XML, Quickr, Redbooks, Sequent (logo), SequentLINK, Server Advantage, ServeRAID, Service Director, SmoothStart, SP, System i, System i5, System p, System p5, System Storage, System z, System z9, S/390 Parallel Enterprise Server, Tivoli Enterprise, TME 10, TotalStorage Proven, Ultramedia, VideoCharger, Virtualization Engine, Visualization Data Explorer, Workload Partitions Manager, X-Architecture, z/Architecture, z/9. A full list of U.S. trademarks owned by IBM may be found at: http://www.ibm.com/legal/copytrade.shtml . The Power Architecture and Power.org wordmarks and the Power and Power.org logos and related marks are trademarks and service marks licensed by Power.org. UNIX is a registered trademark of The Open Group in the United States, other countries or both. Linux is a trademark of Linus Torvalds in the United States, other countries or both. Microsoft, Windows, Windows NT and the Windows logo are registered trademarks of Microsoft Corporation in the United States, other countries or both. Intel, Itanium, Pentium are registered trademarks and Xeon is a trademark of Intel Corporation or its subsidiaries in the United States, other countries or both. AMD Opteron is a trademark of Advanced Micro Devices, Inc. Java and all Java-based trademarks and logos are trademarks of Sun Microsystems, Inc. in the United States, other countries or both. TPC-C and TPC-H are trademarks of the Transaction Performance Processing Council (TPPC). SPECint, SPECfp, SPECjbb, SPECweb, SPECjAppServer, SPEC OMP, SPECviewperf, SPECapc, SPEChpc, SPECjvm, SPECmail, SPECimap and SPECsfs are trademarks of the Standard Performance Evaluation Corp (SPEC). NetBench is a registered trademark of Ziff Davis Media in the United States, other countries or both. AltiVec is a trademark of Freescale Semiconductor, Inc. Cell Broadband Engine is a trademark of Sony Computer Entertainment Inc. InfiniBand, InfiniBand Trade Association and the InfiniBand design marks are trademarks and/or service marks of the InfiniBand Trade Association. Other company, product and service names may be trademarks or service marks of others. Revised January 15, 2008