BladeCenter 101
Upcoming SlideShare
Loading in...5
×
 

BladeCenter 101

on

  • 2,360 views

 

Statistics

Views

Total Views
2,360
Views on SlideShare
2,346
Embed Views
14

Actions

Likes
1
Downloads
240
Comments
0

4 Embeds 14

http://lj-toys.com 7
http://www.slideshare.net 3
http://l.lj-toys.com 3
http://xss.yandex.net 1

Accessibility

Categories

Upload Details

Uploaded via as Adobe PDF

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

BladeCenter 101 BladeCenter 101 Presentation Transcript

  • BladeCenter 101 Mike Schambureck with help from Janus Hertz schambur@us.ibm.com IBM Systems Lab Services and Training STG Technical Conferences 2009 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Agenda IBM BladeCenter concepts BladeCenter chassis and blade server options IBM i and BladeCenter solutions BladeCenter components Meet the BladeCenter S Meet the BladeCenter H Management modules and interface Media trays I/O module and expansion adapter portfolio Power domains and management 2 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 What Is IBM BladeCenter technology? IBM BladeCenter is a simple integration of servers, storage and networking. Its innovative, open design offers a true alternative to sprawling racks and overheated server rooms Built on the IBM X-Architecture strategy Five different chassis to choose from Many different blades, including solid-state design and a range of expansion modules An expansive I/O portfolio with simplified deployment and failover capability Software for systems, energy and virtualization management IBM services and support 3 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Smarter, More Efficient Infrastructure IBM BladeCenter is a simple integration of servers, storage and networking. Its innovative, open design offers a true alternative to sprawling racks and overheated server rooms Power, cooling and floor space reduced Easy to add new servers to LAN and SAN fabric Hardware management with a single interface – AMM LAN, SAN switches & cables integrated into chassis KVM costs eliminated PDU costs drastically reduced 4 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter Chassis iSCSI IBM BladeCenter E Best energy efficiency, IBM BladeCenter T iSCSI best density Ruggedized IBM BladeCenter S iSCSI Distributed, small office, IBM BladeCenter HT easy to configure IBM BladeCenter H Ruggedized, High performance high performance A common set of blades A common set of industry-standard switches and I/O fabrics A common management infrastructure = IBM i on Power Blade support iSCSI = IBM i iSCSI integration support 5 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter chassis comparison BladeCenter S BladeCenter E BladeCenter H (M/T 8886/7779) (M/T 8678) (M/T 8852/7989) Chassis height 7U / 12.0” (305mm) 7U / 12.0” (305mm) 9U / 15.75” (400mm) Chassis depth 28 “ (711mm) 28 “ (711mm) 28.0” (711mm) Max # of blade servers 6 14 14 Max # of I/O Modules 4 4 10 (4 x horizontal, 4 x (Switch, and bridge) vertical, 2 x bridge) Power supply size 950W at 110V 2000 Watts AC 2900 Watts AC (standard) 1450W at 220V # of Power Supplies 2/4 2/ 4 2/ 4 (standard/ maximum) Number of Blowers 4/4 2/ 2 2/ 2 (standard/ maximum) Diskette Drives (standard) None 1x 1.44 MB diskette drive None (in Media Tray) DVD/CD drives standard 1x DVD-ROM (in Media 1x DVD-ROM (in Media 1x DVD-ROM (in Media Tray) Tray) Tray) Max # of Disk Drives in 12 0 0 Chassis 6 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 What Is a Blade Server? • A blade server contains the core components of a server • Each blade server has: • Processor(s) • Memory Blade server with cover off • Internal storage (optional) • Embedded Network Interface Cards (NIC) • I/O expansion adapters (Fibre Channel, SAS) • The blade server plugs into the midplane of a chassis that provides common functions as: • KVM • Power supplies Blade server with its cover on ready for • Cooling fans insertion into the BladeCenter chassis • I/O connectivity to network, storage • Shared media devices (optical, USB and diskette (BCE)) • Optional modules to support additional functions IBM BladeCenter S chassis 7 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM Blade Servers JS12 QS21 HS12 High- JS23 HS21 XM High performance Entry LS21 performance, Scalable Extended-memory High performance virtualization performance, virtualization QS22 JS22 HS21 High performance Scalable General-purpose HS22 LS41 performance, enterprise Scalable Scalable, enterprise virtualization performance performance enterprise JS43 Scalable performance, Multiple processor architectures and OS options for virtualization, different applications CPU, memory, I/O expansion Unique value proposition from IBM 8 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM i Solutions with IBM BladeCenter iSCSI Integration IBM i and BladeCenter H IBM i and BladeCenter S BladeCenter E BladeCenter H BladeCenter S Power System Run IBM i, Windows, VMware and Run IBM i, Windows, Linux and Run IBM i, Windows, VMware and Linux Linux VMware IBM i runs on Power based rack/tower IBM i runs on Power blade IBM i runs on Power blade systems Windows/VMware/Linux runs on x86 Windows/Linux/VMware runs on x86 Windows/VMware/Linux runs on x86 blades blades blades IBM FC or SAS storage required Up to 12 drives in BladeCenter chassis IBM i provides storage for x86 blades Diverse set of skills required: or DS3200 SAS storage Skills required are IBM i and BladeCenter BladeCenter, FC/SAS, Unix and IBM i Skills required are BladeCenter, Unix IBM i centric operations and IBM i BladeCenter- & SAN-centric operations BladeCenter centric operations 9 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM i Integration with BladeCenter via iSCSI Power server with IBM i integrates with x86 Initiator Blades running Windows or VMware – iSCSI adapters connect via Ethernet network Target IBM i provides the storage for BladeCenter – Consolidates storage management – Simplifies storage management – Improves resiliency of x86 environment New Software Initiator support – i 6.1 supports the Microsoft software initiator service with select models of Power System BladeCenter and System x servers – Provides same level of integration, while saving the expense of the iSCSI hardware adapter on the blades & servers – Supported with Microsoft Windows 2008 Server and Windows 2003 Server http://www.ibm.com/systems/i/advantages/integratedserver/index.html 10 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM i and BladeCenter H Overview • JS12, JS22, JS23, JS43 supported • VIOS virtualizes all I/O up to IBM i • Mix of POWER6-architecture and x86 blades • Fibre Channel or SAS storage supported • IVM used for Partition Creation and Resource assignments • Virtual tape supported for backups 11 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM i and BladeCenter S Overview • JS12, JS22, JS23, JS43 supported • RAID in BladeCenter S supported • VIOS virtualizes all I/O up to IBM i • Mix of POWER6-architecture and x86 blades • SAS storage supported (in chassis or DS3200) • IVM used for Partition Creation and Resource assignments • Virtual tape supported for backups 12 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Components of IBM BladeCenter There are several components that make up an IBM BladeCenter: – Ethernet Components – Fibre Channel Components – High Speed Solution Components Horizontal I/O Module – Blade Servers Vertical I/O Modules Blade Servers Multi-switch Interconnect Module 13 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Meet the BladeCenter S – Front View Service label cards slot enable quick and easy reference to BladeCenter S SAS and SATA disks can be mixed RAID 0, 1, 5, 0+1 supported with RAID SAS Switch Module (RSSM) 7U Supports up to 6 BladeServers Shared USB ports and CD-RW / DVD-ROM Combo Battery Backup Units for use only with RAID SAS Switch Module 14 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Meet the BladeCenter S – Rear View Hot-swap Power Supplies 3 & 4 are optional, Hot-swap Power Supplies 1 & 2 are Auto-sensing b/w 950W / 1450W standard, Auto-sensing b/w 950W / 1450W 7U Top: AMM standard Four Blower modules Bottom: Serial Pass-thru Module optional standard Top(SW1) & Bottom(SW2) left: Ethernet Top(SW3) & Bottom(SW4) right: SAS 15 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 SAS RAID Controller Switch Module RAID controller support provides additional protection options for BladeCenter S storage SAS RAID Controller Switch Module – High-performance, fully duplex, 3Gbps speeds – Support for RAID 0, 1, 5, & 10 – Supports 2 disk storage modules with up to 12 SAS drives – Supports external SAS tape drive – Supports existing #8250 CFFv SAS adapter on JS12 and JS22 – Supports #8246 CIOv SAS adapter on JS23 and JS43 – 1GB of battery-backed write cache between the 2 modules – Two SAS RAID Controller Switch Modules (#3734) required Supports Power and x86 Blades – Recommend separate RAID sets – For each IBM i partition – For IBM i and Windows storage – Requirements – Firmware update for SAS RAID Controller Switch Modules – VIOS 2.1.1, eFW 3.4.2 Note: Does not support connection to DS3200 IBM i is not pre-installed with RSSM configurations 16 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter S Office Enablement Kit Ideal way to deploy BladeCenter S into office environments Includes Acoustical Module built into the back Optional Air Filter on the front Locking door for security Mobile with rollers 33% (4U) extra room to grow 17 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Meet the BladeCenter H – Front View Power Module 3 Power Filler Module 1 and Fan pack Front System HS20 Panel Blade # 1 9U CD DVD- drive Blade Filler Front USB Power Power Module 2 Module 4 Filler and Fan pack 18 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Meet the BladeCenter H – Rear View I/O module bay 7 and 8 Power Power Connector 2 Connector 1 I/O Module I/O Module bay 1 and bay 3 5 Advanced Management Blower Module 1 Module 1 and 2 Advanced Management I/O Module bay Module 2 slot 2 and 6 Rear LED panel and Serial I/O Module connector bay 4 I/O module bay 9 and 10 Left Shuttle release lever Right Shuttle release lever 19 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter® Management Modules 20 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Advanced Management Module (AMM) BladeCenter systems management controller – the heart beat of a BladeCenter installation – Ethernet port for web interface – USB port for keyboard – USB interface for mouse – Serial connection – Allows up to 5 simultaneous IBM Director Connections – Connection is used for iSCSI attached blade control – Use in BCS, BCE (newer) and BCH – Backward compatible with BladeCenter as well as the standard MM for BladeCenter H – Hot swap, removal of the MM does not effect server operation – Default IP address: 192.168.70.125 – Default Userid: USERID (all caps) – Default Password: PASSW0RD (all caps and zero for letter o) 21 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 AMM UI, Monitors 22 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter Management Module (MM) Interfaces to midplane via 10/100 Mb Ethernet The Management Module has the following features: – KVM Switching Technology Video Keyboard and Mouse USB – Local RS-485 interface – Remote – Service Processor I2C interfaces – Hot Swap Optional redundant Management Module Maybe found on BCE, it was the original MM 10/100 MB Ethernet port Fault LED Active LED Power-on LED IP Reset switch access Video port Mouse Keyboard port port 23 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter® Media Tray 24 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter S Media Tray 5 System LEDs CD-RW / DVD-ROM Combo • Battery backup modules/fillers • Used only with RAID Controller SAS Module 2 USB 2.0 ports 25 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter H - front view – media tray Customer Serviceable Media Tray – New half blade design media tray slides in and is serviceable similar to a blade – Direct wired to the mid-plane – Can be removed without impacting operation of chassis – Tray contains – 2 External 2.0 USB connectors, 1 Internal connector – Full Light Path Diagnostic Panel – DVD-ROM or DVD-RAM Media Tray Release Latch 26 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Blade Server LEDs Power on LED – When a Power blade is inserted, the power LED fast blinks for about 90 seconds while the FSP on the blade initializes the system – When the blade is discovered by the AMM the frequency of the blinking slows down, allowing the power switch to function Media Tray Select Button – Select this button to associate the DVD and USB port with a blade server – This button lights when the ownership of the DVD and USB port transfers to a blade server 27 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter® I/O Modules & Expansion Cards 28 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter Ethernet Networking Portfolio Nortel Layer 2/3 Gb Nortel L2-7 Gb Copper Nortel L2/3 10Gb Nortel Layer 2/3 Gb Copper Ethernet Ethernet Switch Ethernet Uplink Switch Fiber Ethernet Switch Module Module Module Switch Module Cisco Systems Cisco Systems Fiber Cisco Catalyst Cisco Catalyst Intelligent Gb Copper Intelligent Gb Switch Module 3012 Switch Module Ethernet Switch Ethernet Switch 3110G/3110X Module Module 29 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter Ethernet Networking & SAS Portfolio Server Connectivity Nortel 10Gb Ethernet Intelligent Copper Module for IBM Switch Module Pass-Through Module BladeCenter SAS Connectivity SAS RAID Controller Module Module 30 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter Fibre Channel I/O Modules Cisco 4Gb 10 and 20 Brocade 4Gb 10 and QLogic 8Gb 20 port QLogic 4Gb 10 and 20 port Fibre Channel 20 port Fibre Channel Fibre Channel Switch port Fibre Channel Switch Modules Switch Modules Module Switch Module Brocade Intelligent 8Gb Brocade Intelligent 4Gb Pass-Thru Fibre Channel Pass-Thru Fibre Channel Switch Module Switch Module Note: See IBM i on Power Blade Supported Environments for hardware supported by IBM i: 31 BladeCenter 101 http://www.ibm.com/systems/power/hardware/blades/ibmi.html IBM Corporation © 2009
  • STG Technical Conferences 2009 BladeCenter SAS I/O Modules BladeCenter S SAS RAID Controller Module • Supported only in BladeCenter S • RAID support for SAS drives in chassis • Supports SAS tape attachment • No support for attaching DS3200 • 2 are always required BladeCenter SAS Connectivity Module • Supported in BladeCenter S and BladeCenter H • No RAID support • Supports SAS tape attachment • Supports DS3200 attachment Note: See IBM i on Power Blade Supported Environments for hardware supported by IBM i: 32 BladeCenter 101 http://www.ibm.com/systems/power/hardware/blades/ibmi.html IBM Corporation © 2009
  • STG Technical Conferences 2009 Multi-switch Interconnect Module for BCH • Installed in high-speed bays 7 & 8 and/or 9 & 10 • Allows a “vertical” switch to be MSIM installed and use the “horizontal” high- speed fabric (bays 7 – 10) • High-speed fabric is used by CFFh expansion adapters • Fibre Channel switch module must be installed in right I/O module bay (switch bay 8 or 10) • If additional Ethernet networking required additional Ethernet switch module can be installed in left I/O module bay (switch bay 7 or 9) 33 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 I/O Expansion Adapters QLogic Ethernet and 4Gb Fibre Channel 3Gb SAS Expansion Card Expansion Card (CFFh) (CFFv) 3Gb SAS Connectivity Emulex 8Gb Fibre QLogic 8Gb Fibre QLogic 4Gb Fibre Card (CIOv) Channel Expansion Channel Expansion Channel Expansion Card (CIOv) Card (CIOv) Card (CIOv) Note: See IBM i on Power Blade Supported Environments for hardware supported by IBM i: http://www.ibm.com/systems/power/hardware/blades/ibmi.html 34 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 CFFv and CFFh I/O Expansion Adapters Combination Form Factor (CFF) allows for 2 different expansion adapters on the HSSM1 HSSM3 same blade CFFv (Combo Form Factor – Vertical) CFFX CFFv Connects to PCI-X bus to provide access to SerDes SM3 SM4 switch modules in bays 3 & 4 PCI-X Vertical switch form factor SAS for IBM i HSSM2 HSSM4 CFFh (Combo Form Factor – Horizontal) CFFE CFFh Connects to PCIe bus to provide access to the switch modules in bays 7 – 10 PCI-Express Horizontal switch form factor, unless MSIM used Fibre Channel and Ethernet for IBM i Note: See IBM i on Power Blade Supported Environments for hardware supported by IBM i: 35 BladeCenter 101 http://www.ibm.com/systems/power/hardware/blades/ibmi.html © 2009 IBM Corporation
  • STG Technical Conferences 2009 CIOv and CFFh I/O Expansion Adapters Combination I/O Form Factor – Vertical is available only on JS23 and JS43 CFFv adapters not supported on JS23 and JS43 CIOv Connects to new PCIe bus to provide access to switch modules in bays 3 & 4 Vertical switch form factor SAS, Fibre Channel for IBM i CFFh Connects to PCIe bus to provide access to the switch modules in bays 7 – 10 Horizontal switch form factor, unless MSIM used Note: See IBM i on Power Blade Supported Environments for hardware supported by IBM i: Fibre Channel and Ethernet for IBM i http://www.ibm.com/systems/power/hardware/blades/ibmi.html 36 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 CFFh CFFv or CIOv Connectors CFFh 37 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter S server connections to I/O bays Both Ethernet ports wired to I/O bay 1 SAS, Ethernet or iSCSI adapter, wired to I/O bays 3 & 4 Redundant Ethernet, wired to I/O bay 2, CFFh adapter required 38 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter S Midplane - Blade to I/O Bay Mapping AMM Bay Blade “A” I/O Bay 1 #1 Ethernet Bay “B” Blade #2 Blade #3 Blade #4 Blade #5 I/O Bay 3 Blade #6 ENet Switch Fibre SAS Switch Bay SAS “A” “B” RAID Battery Bay PCI-X (CFFv) or PCIe (CIOv) D.C. Blade D.C. Blade #1 Blade Daughter Card D.C. Blade #2 eNet, Fibre, SAS, SAS RAID D.C. Blade #3 I/O Bay 4 #4 D.C. Blade ENet Switch #5 D.C. Blade Fibre #6 SAS SAS Switch Bay “A” PCI-E (CFFh) RAID Battery Bay Blade Daughter Card “B” C.C. Blade I/O Bay 2 C.C. Blade #1 C.C. Blade #2 Option Bay C.C. Blade #3 C.C. Blade #4 C.C. Blade #5 #6 BC-S Mid-Plane 39 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BCH: CFFv and CFFh I/O Connections Blade #N On-Board Dual Switch #1 Gbit Ethernet Ethernet On-Board Dual Gbit Ethernet M I Switch #2 POWER SAS CFFv Ethernet Expansion Card D Blade Server #1 P L Switch #3 QLogic CFFh A Expansion Card N Switch #4 QLogic CFFh Expansion Card: E • Provides 2 x 4Gb Fibre Channel connections to SAN Switch #7 • 2 Fibre Channel ports externalized via Switch 8 & 10 • Provides 2 x 1 Gb Ethernet ports for additional networking Switch #8 • 2 Ethernet ports externalized via Switch 7 & 9 SAS CFFv Expansion Card: Switch #9 • Provides 2 SAS ports for connection to SAS tape drive • 2 SAS ports externalized via Switch 3 & 4 Switch #10 40 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BCH: CIOv and CFFh I/O Connections Blade #N On-Board Dual Switch #1 Gbit Ethernet Ethernet On-Board Dual Gbit Ethernet M I Switch #2 POWER CIOv Expansion Ethernet Card D Blade Server #1 P L Switch #3 QLogic CFFh A Expansion Card N Switch #4 CIOv Expansion Card: E Switch #7 • 2 x 8Gb or 2 x 4Gb Fibre Channel • OR, 2 x 3Gb SAS passthrough • Uses 4Gb or 8Gb FC vertical switches in bays 3 & 4 Switch #8 • OR, 3Gb SAS vertical switches in bays 3 & 4 • Redundant FC storage connection option for IBM i CFFh Expansion Card: Switch #9 • 2 x 4Gb and 2 x 1Gb Ethernet Switch #10 41 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter® S Power Configuration 42 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 When to Use the Optional Power Supplies 3 & 4 One Single Power Domain Standard Power Supplies supported only for: 1 x blade server 1 x Disk Storage Module with disks 1 & 3 I/O bays Optional Power Supplies 3 & 4 needed to power all slots: 2 or more blade servers 2 x Disk Storage Modules with disks 4 x I/O bays Use Power Configurator and Planning Guide to determine specific blade and configuration power usage and power supply requirements. http://www.ibm.com/systems/bladecenter/powerconfig 43 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter® H Power Configuration 44 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter H – power modules Power Module Bays – Maximum of four per chassis – Two ship standard with the chassis – Other two come as a single option part number Power domains – BladeCenter H chassis deploys same ‘domain style’ power topology – Having two domains reduces the chance that any catastrophic failure can take out all four power supplies as might be seen in a topology where all supplies are on same bus 45 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter H – 2900W AC power module detail Power Modules – Power subsystem is redundant – VPD available via I2C – Rated at 2900W maximum DC output – Each Module contains a Fan Pack – 60mm fans (3), and is a CRU – Fan Pack control logic is via I2C bus thru supply – AC OK, DC OK, and Fan Fail LED’s Fan Pack – PS Handle must be rotated up to 90 degrees – Release spring on each side of PS must be depressed to remove Fan Pack – There will be a spring between the Pack and the PS to push the Pack out when the side latches are depressed. – PS will not power when Fan pack is off – PS must be removed from chassis far enough to rotate handle before Fan pack can be removed. 46 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter H AC power distribution rear view 47 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter H - power domains 48 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter H - AC power topology New power simplifies power inputs for BladeCenter – Allows several power cord input options – Solution will vary based on number of chassis being installed Connector on the back of the BladeCenter assures that the cable can not be installed incorrectly These cables work in the same fashion as the connectors on many of the IBM PDU family – Customer serviceable – Easy to install and remove – Same chassis WW This end of the cable allow a single chassis to work WW. Different voltages Different ratings Power Power connector Connector WW Safety 2 1 certifications 49 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 BladeCenter H - AC power input cables TRIPLE 16A IEC 320-C20 (200-240V) 2.8M DOUBLE 30A KSC 8305 (S. KOREA) (220V) 4.3M DOUBLE 30A NEMA L6-30P (208V) 4.3M DUAL 32A IEC 309 P+N+G/16A IEC 320-C20 OR DUAL 32A AS/NZS 3112/16A IEC 320-C20 (AUS/NZ) Both (230V) 4.3M 50 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 IBM BladeCenter® is the RIGHT choice. OPEN. EASY. GREEN. The RIGHT choice, tailored to fit your diverse needs so you can drive innovation with technology advancements OPEN and innovative for a flexible business foundation. EASY to deploy, integrate and manage so you can improve the responsiveness of systems and people GREEN today for a better tomorrow helping to maximize return on your investments 51 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Further Reading IBM i on Blade Read-me First: http://www.ibm.com/systems/power/hardware/blades/ibmi.html IBM i on Blade Supported Environments: http://www.ibm.com/systems/power/hardware/blades/ibmi.html IBM i on Blade Performance Information: http://www.ibm.com/systems/i/advantages/perfmgmt/resource.html Service vouchers: http://www.ibm.com/systems/i/hardware/editions/services.html IBM i on Blade Training: http://www.ibm.com/systems/i/support/itc/educ.html 52 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Trademarks and Disclaimers 8 IBM Corporation 1994-2007. All rights reserved. References in this document to IBM products or services do not imply that IBM intends to make them available in every country. Trademarks of International Business Machines Corporation in the United States, other countries, or both can be found on the World Wide Web at http://www.ibm.com/legal/copytrade.shtml. Intel, Intel logo, Intel Inside, Intel Inside logo, Intel Centrino, Intel Centrino logo, Celeron, Intel Xeon, Intel SpeedStep, Itanium, and Pentium are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries. Linux is a registered trademark of Linus Torvalds in the United States, other countries, or both. Microsoft, Windows, Windows NT, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. IT Infrastructure Library is a registered trademark of the Central Computer and Telecommunications Agency which is now part of the Office of Government Commerce. ITIL is a registered trademark, and a registered community trademark of the Office of Government Commerce, and is registered in the U.S. Patent and Trademark Office. UNIX is a registered trademark of The Open Group in the United States and other countries. Java and all Java-based trademarks are trademarks of Sun Microsystems, Inc. in the United States, other countries, or both. Other company, product, or service names may be trademarks or service marks of others. Information is provided "AS IS" without warranty of any kind. The customer examples described are presented as illustrations of how those customers have used IBM products and the results they may have achieved. Actual environmental costs and performance characteristics may vary by customer. Information concerning non-IBM products was obtained from a supplier of these products, published announcement material, or other publicly available sources and does not constitute an endorsement of such products by IBM. Sources for non-IBM list prices and performance numbers are taken from publicly available information, including vendor announcements and vendor worldwide homepages. IBM has not tested these products and cannot confirm the accuracy of performance, capability, or any other claims related to non-IBM products. Questions on the capability of non-IBM products should be addressed to the supplier of those products. All statements regarding IBM future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only. Some information addresses anticipated future capabilities. Such information is not intended as a definitive statement of a commitment to specific levels of performance, function or delivery schedules with respect to any future products. Such commitments are only made in IBM product announcements. The information is presented here to communicate IBM's current investment and development activities as a good faith effort to help with our customers' future planning. Performance is based on measurements and projections using standard IBM benchmarks in a controlled environment. The actual throughput or performance that any user will experience will vary depending upon considerations such as the amount of multiprogramming in the user's job stream, the I/O configuration, the storage configuration, and the workload processed. Therefore, no assurance can be given that an individual user will achieve throughput or performance improvements equivalent to the ratios stated here. Prices are suggested U.S. list prices and are subject to change without notice. Starting price may not include a hard drive, operating system or other features. Contact your IBM representative or Business Partner for the most current pricing in your geography. Photographs shown may be engineering prototypes. Changes may be incorporated in production models. 53 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Special notices This document was developed for IBM offerings in the United States as of the date of publication. IBM may not make these offerings available in other countries, and the information is subject to change without notice. Consult your local IBM business contact for information on the IBM offerings available in your area. Information in this document concerning non-IBM products was obtained from the suppliers of these products or other public sources. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. IBM may have patents or pending patent applications covering subject matter in this document. The furnishing of this document does not give you any license to these patents. Send license inquires, in writing, to IBM Director of Licensing, IBM Corporation, New Castle Drive, Armonk, NY 10504-1785 USA. All statements regarding IBM future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only. The information contained in this document has not been submitted to any formal IBM test and is provided "AS IS" with no warranties or guarantees either expressed or implied. All examples cited or described in this document are presented as illustrations of the manner in which some IBM products can be used and the results that may be achieved. Actual environmental costs and performance characteristics will vary depending on individual client configurations and conditions. IBM Global Financing offerings are provided through IBM Credit Corporation in the United States and other IBM subsidiaries and divisions worldwide to qualified commercial and government clients. Rates are based on a client's credit rating, financing terms, offering type, equipment type and options, and may vary by country. Other restrictions may apply. Rates and offerings are subject to change, extension or withdrawal without notice. IBM is not responsible for printing errors in this document that result in pricing or information inaccuracies. All prices shown are IBM's United States suggested list prices and are subject to change without notice; reseller prices may vary. IBM hardware products are manufactured from new parts, or new and serviceable used parts. Regardless, our warranty terms apply. Any performance data contained in this document was determined in a controlled environment. Actual results may vary significantly and are dependent on many factors including system hardware configuration and software design and configuration. Some measurements quoted in this document may have been made on development-level systems. There is no guarantee these measurements will be the same on generally- available systems. Some measurements quoted in this document may have been estimated through extrapolation. Users of this document should verify the applicable data for their specific environment. Revised September 26, 2006 54 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Special notices (cont.) IBM, the IBM logo, ibm.com AIX, AIX (logo), AIX 6 (logo), AS/400, BladeCenter, Blue Gene, ClusterProven, DB2, ESCON, i5/OS, i5/OS (logo), IBM Business Partner (logo), IntelliStation, LoadLeveler, Lotus, Lotus Notes, Notes, Operating System/400, OS/400, PartnerLink, PartnerWorld, PowerPC, pSeries, Rational, RISC System/6000, RS/6000, THINK, Tivoli, Tivoli (logo), Tivoli Management Environment, WebSphere, xSeries, z/OS, zSeries, AIX 5L, Chiphopper, Chipkill, Cloudscape, DB2 Universal Database, DS4000, DS6000, DS8000, EnergyScale, Enterprise Workload Manager, General Purpose File System, , GPFS, HACMP, HACMP/6000, HASM, IBM Systems Director Active Energy Manager, iSeries, Micro-Partitioning, POWER, PowerExecutive, PowerVM, PowerVM (logo), PowerHA, Power Architecture, Power Everywhere, Power Family, POWER Hypervisor, Power Systems, Power Systems (logo), Power Systems Software, Power Systems Software (logo), POWER2, POWER3, POWER4, POWER4+, POWER5, POWER5+, POWER6, System i, System p, System p5, System Storage, System z, Tivoli Enterprise, TME 10, Workload Partitions Manager and X-Architecture are trademarks or registered trademarks of International Business Machines Corporation in the United States, other countries, or both. If these and other IBM trademarked terms are marked on their first occurrence in this information with a trademark symbol (® or ™), these symbols indicate U.S. registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law trademarks in other countries. A current list of IBM trademarks is available on the Web at "Copyright and trademark information" at www.ibm.com/legal/copytrade.shtml The Power Architecture and Power.org wordmarks and the Power and Power.org logos and related marks are trademarks and service marks licensed by Power.org. UNIX is a registered trademark of The Open Group in the United States, other countries or both. Linux is a registered trademark of Linus Torvalds in the United States, other countries or both. Microsoft, Windows and the Windows logo are registered trademarks of Microsoft Corporation in the United States, other countries or both. Intel, Itanium, Pentium are registered trademarks and Xeon is a trademark of Intel Corporation or its subsidiaries in the United States, other countries or both. AMD Opteron is a trademark of Advanced Micro Devices, Inc. Java and all Java-based trademarks and logos are trademarks of Sun Microsystems, Inc. in the United States, other countries or both. TPC-C and TPC-H are trademarks of the Transaction Performance Processing Council (TPPC). SPECint, SPECfp, SPECjbb, SPECweb, SPECjAppServer, SPEC OMP, SPECviewperf, SPECapc, SPEChpc, SPECjvm, SPECmail, SPECimap and SPECsfs are trademarks of the Standard Performance Evaluation Corp (SPEC). NetBench is a registered trademark of Ziff Davis Media in the United States, other countries or both. AltiVec is a trademark of Freescale Semiconductor, Inc. Cell Broadband Engine is a trademark of Sony Computer Entertainment Inc. InfiniBand, InfiniBand Trade Association and the InfiniBand design marks are trademarks and/or service marks of the InfiniBand Trade Association. Other company, product and service names may be trademarks or service marks of others. Revised April 24, 2008 55 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Notes on benchmarks and values The IBM benchmarks results shown herein were derived using particular, well configured, development-level and generally-available computer systems. Buyers should consult other sources of information to evaluate the performance of systems they are considering buying and should consider conducting application oriented testing. For additional information about the benchmarks, values and systems tested, contact your local IBM office or IBM authorized reseller or access the Web site of the benchmark consortium or benchmark vendor. IBM benchmark results can be found in the IBM Power Systems Performance Report at http://www.ibm.com/systems/p/hardware/system_perf.html. All performance measurements were made with AIX or AIX 5L operating systems unless otherwise indicated to have used Linux. For new and upgraded systems, AIX Version 4.3, AIX 5L or AIX 6 were used. All other systems used previous versions of AIX. The SPEC CPU2006, SPEC2000, LINPACK, and Technical Computing benchmarks were compiled using IBM's high performance C, C++, and FORTRAN compilers for AIX 5L and Linux. For new and upgraded systems, the latest versions of these compilers were used: XL C Enterprise Edition V7.0 for AIX, XL C/C++ Enterprise Edition V7.0 for AIX, XL FORTRAN Enterprise Edition V9.1 for AIX, XL C/C++ Advanced Edition V7.0 for Linux, and XL FORTRAN Advanced Edition V9.1 for Linux. The SPEC CPU95 (retired in 2000) tests used preprocessors, KAP 3.2 for FORTRAN and KAP/C 1.4.2 from Kuck & Associates and VAST-2 v4.01X8 from Pacific-Sierra Research. The preprocessors were purchased separately from these vendors. Other software packages like IBM ESSL for AIX, MASS for AIX and Kazushige Goto’s BLAS Library for Linux were also used in some benchmarks. For a definition/explanation of each benchmark and the full list of detailed results, visit the Web site of the benchmark consortium or benchmark vendor. TPC http://www.tpc.org SPEC http://www.spec.org LINPACK http://www.netlib.org/benchmark/performance.pdf Pro/E http://www.proe.com GPC http://www.spec.org/gpc NotesBench http://www.notesbench.org VolanoMark http://www.volano.com STREAM http://www.cs.virginia.edu/stream/ SAP http://www.sap.com/benchmark/ Oracle Applications http://www.oracle.com/apps_benchmark/ PeopleSoft - To get information on PeopleSoft benchmarks, contact PeopleSoft directly Siebel http://www.siebel.com/crm/performance_benchmark/index.shtm Baan http://www.ssaglobal.com Microsoft Exchange http://www.microsoft.com/exchange/evaluation/performance/default.asp Veritest http://www.veritest.com/clients/reports Fluent http://www.fluent.com/software/fluent/index.htm TOP500 Supercomputers http://www.top500.org/ Ideas International http://www.ideasinternational.com/benchmark/bench.html Storage Performance Council http://www.storageperformance.org/results Revised January 15, 2008 56 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Notes on HPC benchmarks and values The IBM benchmarks results shown herein were derived using particular, well configured, development-level and generally-available computer systems. Buyers should consult other sources of information to evaluate the performance of systems they are considering buying and should consider conducting application oriented testing. For additional information about the benchmarks, values and systems tested, contact your local IBM office or IBM authorized reseller or access the Web site of the benchmark consortium or benchmark vendor. IBM benchmark results can be found in the IBM Power Systems Performance Report at http://www.ibm.com/systems/p/hardware/system_perf.html. All performance measurements were made with AIX or AIX 5L operating systems unless otherwise indicated to have used Linux. For new and upgraded systems, AIX Version 4.3 or AIX 5L were used. All other systems used previous versions of AIX. The SPEC CPU2000, LINPACK, and Technical Computing benchmarks were compiled using IBM's high performance C, C++, and FORTRAN compilers for AIX 5L and Linux. For new and upgraded systems, the latest versions of these compilers were used: XL C Enterprise Edition V7.0 for AIX, XL C/C++ Enterprise Edition V7.0 for AIX, XL FORTRAN Enterprise Edition V9.1 for AIX, XL C/C++ Advanced Edition V7.0 for Linux, and XL FORTRAN Advanced Edition V9.1 for Linux. The SPEC CPU95 (retired in 2000) tests used preprocessors, KAP 3.2 for FORTRAN and KAP/C 1.4.2 from Kuck & Associates and VAST-2 v4.01X8 from Pacific-Sierra Research. The preprocessors were purchased separately from these vendors. Other software packages like IBM ESSL for AIX, MASS for AIX and Kazushige Goto’s BLAS Library for Linux were also used in some benchmarks. For a definition/explanation of each benchmark and the full list of detailed results, visit the Web site of the benchmark consortium or benchmark vendor. SPEC http://www.spec.org LINPACK http://www.netlib.org/benchmark/performance.pdf Pro/E http://www.proe.com GPC http://www.spec.org/gpc STREAM http://www.cs.virginia.edu/stream/ Veritest http://www.veritest.com/clients/reports Fluent http://www.fluent.com/software/fluent/index.htm TOP500 Supercomputers http://www.top500.org/ AMBER http://amber.scripps.edu/ FLUENT http://www.fluent.com/software/fluent/fl5bench/index.htm GAMESS http://www.msg.chem.iastate.edu/gamess GAUSSIAN http://www.gaussian.com ABAQUS http://www.abaqus.com/support/sup_tech_notes64.html select Abaqus v6.4 Performance Data ANSYS http://www.ansys.com/services/hardware_support/index.htm select “Hardware Support Database”, then benchmarks. ECLIPSE http://www.sis.slb.com/content/software/simulation/index.asp?seg=geoquest& MM5 http://www.mmm.ucar.edu/mm5/ MSC.NASTRAN http://www.mscsoftware.com/support/prod%5Fsupport/nastran/performance/v04_sngl.cfm STAR-CD www.cd-adapco.com/products/STAR-CD/performance/320/index/html NAMD http://www.ks.uiuc.edu/Research/namd HMMER http://hmmer.janelia.org/ Revised January 15, 2008 http://powerdev.osuosl.org/project/hmmerAltivecGen2mod 57 BladeCenter 101 © 2009 IBM Corporation
  • STG Technical Conferences 2009 Notes on performance estimates rPerf for AIX rPerf (Relative Performance) is an estimate of commercial processing performance relative to other IBM UNIX systems. It is derived from an IBM analytical model which uses characteristics from IBM internal workloads, TPC and SPEC benchmarks. The rPerf model is not intended to represent any specific public benchmark results and should not be reasonably used in that way. The model simulates some of the system operations such as CPU, cache and memory. However, the model does not simulate disk or network I/O operations. rPerf estimates are calculated based on systems with the latest levels of AIX and other pertinent software at the time of system announcement. Actual performance will vary based on application and configuration specifics. The IBM eServer pSeries 640 is the baseline reference system and has a value of 1.0. Although rPerf may be used to approximate relative IBM UNIX commercial processing performance, actual system performance may vary and is dependent upon many factors including system hardware configuration and software design and configuration. Note that the rPerf methodology used for the POWER6 systems is identical to that used for the POWER5 systems. Variations in incremental system performance may be observed in commercial workloads due to changes in the underlying system architecture. All performance estimates are provided "AS IS" and no warranties or guarantees are expressed or implied by IBM. Buyers should consult other sources of information, including system benchmarks, and application sizing guides to evaluate the performance of a system they are considering buying. For additional information about rPerf, contact your local IBM office or IBM authorized reseller. ======================================================================== CPW for IBM i Commercial Processing Workload (CPW) is a relative measure of performance of processors running the IBM i operating system. Performance in customer environments may vary. The value is based on maximum configurations. More performance information is available in the Performance Capabilities Reference at: www.ibm.com/systems/i/solutions/perfmgmt/resource.html Revised April 2, 2007 58 BladeCenter 101 © 2009 IBM Corporation