IBM Ex5 Implementation Guide
IBM Ex5 Implementation Guide
IBM Ex5 Implementation Guide
David Watts Aaron Belisle Duncan Furniss Scott Haddow Michael Hurman Jeneea Jervay Eric Kern Cynthia Knight Miroslav Peic Tom Sorcic Evans Tanurdin
ibm.com/redbooks
International Technical Support Organization IBM eX5 Implementation Guide May 2011
SG24-7909-00
Note: Before using this information and the product it supports, read the information in Notices on page xi.
First Edition (May 2011) This edition applies to the following servers: IBM System x3850 X5, machine type 7145 IBM System x3950 X5, machine type 7145 IBM System x3690 X5, machine type 7148 IBM BladeCenter HX5, machine type 7872
Copyright International Business Machines Corporation 2011. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xii Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii The team who wrote this book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii Chapter 1. Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.1 eX5 systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 1.2 Model summary. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.2.1 IBM System x3850 X5 models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.2.2 Workload-optimized x3950 X5 models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.2.3 x3850 X5 models with MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.2.4 Base x3690 X5 models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.2.5 Workload-optimized x3690 X5 models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 1.2.6 BladeCenter HX5 models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 1.3 Positioning. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 1.3.1 IBM System x3850 X5 and x3950 X5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 1.3.2 IBM System x3690 X5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 1.3.3 IBM BladeCenter HX5. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 1.4 Energy efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 1.5 Services offerings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 1.6 What this book contains . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 Part 1. Product overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 Chapter 2. IBM eX5 technology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1 eX5 chip set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Intel Xeon 6500 and 7500 family processors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.1 Intel Virtualization Technology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2 Hyper-Threading Technology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.3 Turbo Boost Technology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.4 QuickPath Interconnect (QPI) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.5 Processor performance in a green world . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3 Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.1 Memory speed . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.2 Memory DIMM placement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.3 Memory ranking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.4 Nonuniform memory architecture (NUMA) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.5 Hemisphere Mode. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.6 Reliability, availability, and serviceability (RAS) features . . . . . . . . . . . . . . . . . . . 2.3.7 I/O hubs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4 MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5 Scalability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.6 Partitioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7 UEFI system settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7.1 System power operating modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Copyright IBM Corp. 2011. All rights reserved.
15 16 16 17 17 18 18 21 22 22 23 24 26 26 28 30 31 33 34 36 38 iii
2.7.2 System power settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7.3 Performance-related individual system settings . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8 IBM eXFlash . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8.1 IBM eXFlash price-performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9 Integrated virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.1 VMware ESXi . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.2 Red Hat RHEV-H (KVM) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.3 Windows 2008 R2 Hyper-V. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.10 Changes in technology demand changes in implementation . . . . . . . . . . . . . . . . . . . 2.10.1 Using swap files . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.10.2 SSD drives and battery backup cache on RAID controllers . . . . . . . . . . . . . . . . 2.10.3 Increased resources for virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.10.4 Virtualized Memcached distributed memory caching . . . . . . . . . . . . . . . . . . . . .
42 43 47 49 50 50 50 51 51 51 52 52 52
Chapter 3. IBM System x3850 X5 and x3950 X5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 3.1 Product features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 3.1.1 IBM System x3850 X5 product features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 3.1.2 IBM System x3950 X5 product features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 3.1.3 IBM MAX5 memory expansion unit. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 3.1.4 Comparing the x3850 X5 to the x3850 M2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 3.2 Target workloads. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 3.3 Models. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 3.4 System architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 3.4.1 System board . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 3.4.2 QPI Wrap Card . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 3.5 MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68 3.6 Scalability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70 3.6.1 Memory scalability with MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 3.6.2 Two-node scalability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 3.7 Processor options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74 3.8 Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 3.8.1 Memory cards and DIMMs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 3.8.2 DIMM population sequence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79 3.8.3 Maximizing memory performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84 3.8.4 Memory mirroring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 3.8.5 Memory sparing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 3.8.6 Effect on performance by using mirroring or sparing . . . . . . . . . . . . . . . . . . . . . . 89 3.9 Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 3.9.1 Internal disks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 3.9.2 SAS and SSD 2.5-inch disk support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 3.9.3 IBM eXFlash and 1.8-inch SSD support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 3.9.4 SAS and SSD controllers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96 3.9.5 Dedicated controller slot . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100 3.9.6 External storage connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 3.10 Optical drives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102 3.11 PCIe slots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 3.12 I/O cards . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104 3.12.1 Standard Emulex 10Gb Ethernet Adapter . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104 3.12.2 Optional adapters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 3.13 Standard onboard features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 3.13.1 Onboard Ethernet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 3.13.2 Environmental data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 3.13.3 Integrated Management Module (IMM). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
iv
3.13.4 UEFI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.13.5 Integrated Trusted Platform Module (TPM). . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.13.6 Light path diagnostics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.14 Power supplies and fans of the x3850 X5 and MAX5 . . . . . . . . . . . . . . . . . . . . . . . . 3.14.1 x3850 X5 power supplies and fans . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.14.2 MAX5 power supplies and fans . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.15 Integrated virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.16 Operating system support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.17 Rack considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 4. IBM System x3690 X5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1 Product features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.1 System components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.2 IBM MAX5 memory expansion unit. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2 Target workloads. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3 Models. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.4 System architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.5 MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.6 Scalability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7 Processor options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8 Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.1 Memory DIMM options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.2 x3690 X5 memory population order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.3 MAX5 memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.4 Memory balance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.5 Mixing DIMMs and the performance effect . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.6 Memory mirroring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.7 Memory sparing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.8 Effect on performance of using mirroring or sparing . . . . . . . . . . . . . . . . . . . . . . 4.9 Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.1 2.5-inch SAS drive support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.2 IBM eXFlash and SSD disk support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.3 SAS and SSD controller summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.4 Battery backup placement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.5 ServeRAID Expansion Adapter. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.6 Drive combinations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.7 External SAS storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.8 Optical drives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10 PCIe slots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10.1 Riser 1. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10.2 Riser 2. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10.3 Emulex 10Gb Ethernet Adapter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10.4 I/O adapters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11 Standard features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.1 Integrated management module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.2 Ethernet subsystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.3 USB subsystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.4 Integrated Trusted Platform Module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.5 Light path diagnostics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.6 Cooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.12 Power supplies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.12.1 x3690 X5 power subsystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.12.2 MAX5 power subsystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
110 111 111 112 112 113 114 114 115 117 118 119 121 123 124 126 127 128 130 131 133 133 136 139 140 141 143 144 145 145 149 152 155 157 158 162 163 164 165 165 166 168 169 169 170 170 170 170 171 173 173 174
Contents
4.13 Integrated virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 174 4.14 Supported operating systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175 4.15 Rack mounting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176 Chapter 5. IBM BladeCenter HX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.1 Comparison to the HS22 and HS22V . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2 Target workloads. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.3 Chassis support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4 Models. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5 System architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6 Speed Burst Card . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7 IBM MAX5 for BladeCenter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.8 Scalability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.8.1 Single HX5 configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.8.2 Double-wide HX5 configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.8.3 HX5 with MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.9 Processor options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10 Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10.1 Memory options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10.2 DIMM population order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10.3 Memory balance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10.4 Memory mirroring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10.5 Memory sparing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.11 Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.11.1 Solid-state drives (SSDs) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.11.2 LSI configuration utility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.11.3 Determining which SSD RAID configuration to choose . . . . . . . . . . . . . . . . . . 5.11.4 Connecting to external SAS storage devices . . . . . . . . . . . . . . . . . . . . . . . . . . 5.12 BladeCenter PCI Express Gen 2 Expansion Blade . . . . . . . . . . . . . . . . . . . . . . . . . 5.13 I/O expansion cards . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.13.1 CIOv . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.13.2 CFFh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.14 Standard onboard features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.14.1 UEFI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.14.2 Onboard network adapters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.14.3 Integrated Management Module (IMM). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.14.4 Video controller . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.14.5 Trusted Platform Module (TPM) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.15 Integrated virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.16 Partitioning capabilities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.17 Operating system support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177 178 180 181 182 183 184 185 186 188 188 188 190 192 194 194 196 199 200 202 203 204 205 207 207 208 209 209 210 212 212 212 213 213 213 214 214 215
Part 2. Implementing scalability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 217 Chapter 6. IBM System x3850 X5 and x3950 X5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.1 Before you apply power for the first time after shipping . . . . . . . . . . . . . . . . . . . . . . . 6.1.1 Verify that the components are securely installed. . . . . . . . . . . . . . . . . . . . . . . . 6.1.2 Clear CMOS memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.1.3 Verify that the server completes POST before adding options . . . . . . . . . . . . . . 6.2 Processor considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.1 Minimum processors required. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.2 Processor operating characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.3 Processor installation order. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi
IBM eX5 Implementation Guide
6.2.4 Processor installation tool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3 Local memory configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.1 Testing the memory DIMMs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.2 Memory fault tolerance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4 Attaching the MAX5 memory expansion unit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.1 Before you attach the MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.2 Installing in a rack . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.3 MAX5 cables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.4 Accessing the DIMMs in the MAX5. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.5 Forming a 2-node x3850 X5 complex . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.5.1 Firmware requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.5.2 Processor requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.5.3 Memory requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.5.4 Cabling the servers together . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.6 PCIe adapters and riser card options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.6.1 Generation 2 and Generation 1 PCIe adapters . . . . . . . . . . . . . . . . . . . . . . . . . 6.6.2 PCIe adapters: Slot selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.6.3 Cleaning up the boot sequence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.7 Power supply considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8 Using the Integrated Management Module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.1 IMM network access . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.2 Configuring the IMM network interface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.3 IMM communications troubleshooting. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.4 IMM functions to help you perform problem determination . . . . . . . . . . . . . . . . . 6.9 UEFI settings. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.9.1 Settings needed for 1-node, 2-node, and MAX5 configurations . . . . . . . . . . . . . 6.9.2 UEFI performance tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.10 Installing an OS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.10.1 Installing without a local optical drive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.10.2 Use of embedded VMware ESXi. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.10.3 Installing the ESX 4.1 or ESXi 4.1 Installable onto x3850 X5 . . . . . . . . . . . . . . 6.10.4 OS installation tips and instructions on the web . . . . . . . . . . . . . . . . . . . . . . . . 6.10.5 Downloads and fixes for x3850 X5 and MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . 6.10.6 SAN storage reference and considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.11 Failure detection and recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.11.1 What happens when a node fails or the MAX5 fails . . . . . . . . . . . . . . . . . . . . . 6.11.2 Reinserting the QPI wrap cards for extended outages . . . . . . . . . . . . . . . . . . . 6.11.3 Tools to aid hardware troubleshooting for x3850 X5. . . . . . . . . . . . . . . . . . . . . 6.11.4 Recovery process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 7. IBM System x3690 X5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.1 Before you apply power for the first time after shipping . . . . . . . . . . . . . . . . . . . . . . . 7.1.1 Verify that the components are securely installed. . . . . . . . . . . . . . . . . . . . . . . . 7.1.2 Clear CMOS memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.1.3 Verify that the server will complete POST before adding options . . . . . . . . . . . . 7.2 Processor considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.1 Minimum processors required. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.2 Processor operating characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3 Memory considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.1 Local memory installation considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.2 Testing the memory DIMMs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.3 Memory fault tolerance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4 MAX5 considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
224 225 226 229 230 230 231 231 233 235 235 236 236 236 238 239 244 245 249 250 251 251 253 253 259 261 262 263 263 271 275 288 293 294 297 297 297 297 299 301 302 302 302 304 304 304 305 306 306 307 310 311
Contents
vii
7.4.1 Before you attach the MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.2 Installing in a rack . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.3 MAX5 cables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.4 Accessing the DIMMs in the MAX5. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5 PCIe adapters and riser card options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5.1 Generation 2 and Generation 1 PCIe adapters . . . . . . . . . . . . . . . . . . . . . . . . . 7.5.2 PCIe adapters: Slot selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5.3 Cleaning up the boot sequence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.6 Power supply considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.7 Using the Integrated Management Module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.7.1 IMM network access . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.7.2 Configuring the IMM network interface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.7.3 IMM communications troubleshooting. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.7.4 IMM functions to help you perform problem determination . . . . . . . . . . . . . . . . . 7.8 UEFI settings. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.8.1 Scaled system settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.8.2 Operating system-specific settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.8.3 Power and performance system settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.8.4 Optimizing boot options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9 Operating system installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9.1 Installation media . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9.2 Integrated virtualization hypervisor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9.3 Windows Server 2008 R2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9.4 Red Hat Enterprise Linux 6 and SUSE Linux Enterprise Server 11 . . . . . . . . . . 7.9.5 VMware vSphere ESXi 4.1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9.6 VMware vSphere ESX 4.1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.9.7 Downloads and fixes for the x3690 X5 and MAX5 . . . . . . . . . . . . . . . . . . . . . . . 7.9.8 SAN storage reference and considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.10 Failure detection and recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.10.1 System alerts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.10.2 System recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 8. IBM BladeCenter HX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1 Before you apply power for the first time after shipping . . . . . . . . . . . . . . . . . . . . . . . 8.1.1 Verifying that the components are securely installed . . . . . . . . . . . . . . . . . . . . . 8.1.2 Clearing CMOS memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1.3 Verifying the server boots before adding options . . . . . . . . . . . . . . . . . . . . . . . . 8.2 Planning to scale: Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.1 Processors supported and requirements to scale. . . . . . . . . . . . . . . . . . . . . . . . 8.2.2 Minimum memory requirement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.3 Required firmware of each blade and the AMM . . . . . . . . . . . . . . . . . . . . . . . . . 8.3 Recommendations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.1 Power sharing cap . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.2 BladeCenter H considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4 Local storage considerations and array setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.1 Launching the LSI Setup Utility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.2 Creating a RAID-1 mirror using the LSI Setup Utility . . . . . . . . . . . . . . . . . . . . . 8.4.3 Using IBM ServerGuide to configure the LSI controller . . . . . . . . . . . . . . . . . . . 8.4.4 Speed Burst Card reinstallation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5 UEFI settings. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.1 UEFI performance tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.2 Start-up parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.3 HX5 single-node UEFI settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
311 312 312 314 316 316 321 322 326 327 328 328 330 331 337 338 339 340 343 346 346 355 356 358 358 362 365 367 369 369 371 373 374 374 375 376 377 377 377 379 382 382 383 385 386 389 392 394 396 397 398 400
viii
8.5.4 HX5 2-node UEFI settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.5 HX5 with MAX5 attached . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.6 Operating system-specific settings in UEFI . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.6 Creating an HX5 scalable complex . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.6.1 Troubleshooting HX5 problems. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7 Operating system installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.1 Operating system installation media . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.2 VMware ESXi on a USB key . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.3 Installing ESX 4.1 or ESXi 4.1 Installable onto HX5 . . . . . . . . . . . . . . . . . . . . . . 8.7.4 Windows installation tips and settings. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.5 Red Hat Enterprise Linux installation tips and settings . . . . . . . . . . . . . . . . . . . . 8.7.6 SUSE Linux Enterprise Server installation tips and settings. . . . . . . . . . . . . . . . 8.7.7 Downloads and fixes for HX5 and MAX5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.8 SAN storage reference and considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.8 Failure detection and recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.8.1 Tools to aid hardware troubleshooting for the HX5. . . . . . . . . . . . . . . . . . . . . . . 8.8.2 Reinserting the Speed Burst card for extended outages . . . . . . . . . . . . . . . . . . 8.8.3 Effects of power loss on HX5 2-node or MAX5 configurations . . . . . . . . . . . . . .
401 401 402 402 406 407 407 415 421 434 436 437 438 440 442 443 444 444
Chapter 9. Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 447 9.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 448 9.2 Integrated Management Module (IMM) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 449 9.2.1 IMM out-of-band configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 449 9.2.2 IMM in-band configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 453 9.2.3 Updating firmware using the IMM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 454 9.3 Advanced Management Module (AMM) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 454 9.3.1 Accessing the Advanced Management Module . . . . . . . . . . . . . . . . . . . . . . . . . 456 9.3.2 Service Advisor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 458 9.3.3 Updating firmware using the AMM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 461 9.4 Remote control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 462 9.4.1 Accessing the Remote Control feature on the x3690 X5 and the x3850 X5 . . . . 462 9.4.2 Accessing the Remote Control feature for the HX5 . . . . . . . . . . . . . . . . . . . . . . 465 9.5 IBM Systems Director 6.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 467 9.5.1 Discovering the IMM of a single-node x3690 X5 or x3850 X5 out-of-band via IBM Systems Director . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 468 9.5.2 Discovering a 2-node x3850 X5 via IBM Systems Director 6.2.x . . . . . . . . . . . . 472 9.5.3 Discovering a single-node HX5 via IBM Systems Director . . . . . . . . . . . . . . . . . 477 9.5.4 Discovering a 2-node HX5 via IBM Systems Director 6.2.x . . . . . . . . . . . . . . . . 478 9.5.5 Service and Support Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 481 9.5.6 Performing tasks against a 2-node system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 488 9.6 IBM Electronic Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 493 9.7 Advanced Settings Utility (ASU) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 495 9.7.1 Using ASU to configure settings in IMM-based servers . . . . . . . . . . . . . . . . . . . 495 9.7.2 Common problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 498 9.7.3 Command examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 499 9.8 IBM ServerGuide. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 501 9.9 IBM ServerGuide Scripting Toolkit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 507 9.10 Firmware update tools and methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 509 9.10.1 Configuring UEFI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 509 9.10.2 Requirements for updating scalable systems . . . . . . . . . . . . . . . . . . . . . . . . . . 510 9.10.3 IBM Systems Director . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 511 9.11 UpdateXpress System Pack Installer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 511 9.12 Bootable Media Creator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 514
Contents
ix
9.13 MegaRAID Storage Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.13.1 Installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.13.2 Drive states . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.13.3 Virtual drive states . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.13.4 MegaCLI utility for storage management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.14 Serial over LAN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.14.1 Enabling SoL in UEFI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.14.2 BladeCenter requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.14.3 Enabling SoL in the operating system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.14.4 How to start a SoL connection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
521 521 522 523 523 525 526 527 529 533
Abbreviations and acronyms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 537 Related publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . IBM Redbooks publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Other publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Online resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Help from IBM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 541 541 542 543 543
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 545
Notices
This information was developed for products and services offered in the U.S.A. IBM may not offer the products, services, or features discussed in this document in other countries. Consult your local IBM representative for information on the products and services currently available in your area. Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to evaluate and verify the operation of any non-IBM product, program, or service. IBM may have patents or pending patent applications covering subject matter described in this document. The furnishing of this document does not give you any license to these patents. You can send license inquiries, in writing, to: IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A. The following paragraph does not apply to the United Kingdom or any other country where such provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you. This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice. Any references in this information to non-IBM websites are provided for convenience only and do not in any manner serve as an endorsement of those websites. The materials at those websites are not part of the materials for this IBM product and use of those websites is at your own risk. IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you. Information concerning non-IBM products was obtained from the suppliers of those products, their published announcements or other publicly available sources. IBM has not tested those products and cannot confirm the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. This information contains examples of data and reports used in daily business operations. To illustrate them as completely as possible, the examples include the names of individuals, companies, brands, and products. All of these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is entirely coincidental. COPYRIGHT LICENSE: This information contains sample application programs in source language, which illustrate programming techniques on various operating platforms. You may copy, modify, and distribute these sample programs in any form without payment to IBM, for the purposes of developing, using, marketing or distributing application programs conforming to the application programming interface for the operating platform for which the sample programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or imply reliability, serviceability, or function of these programs.
xi
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines Corporation in the United States, other countries, or both. These and other IBM trademarked terms are marked on their first occurrence in this information with the appropriate symbol ( or ), indicating US registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law trademarks in other countries. A current list of IBM trademarks is available on the Web at http://www.ibm.com/legal/copytrade.shtml The following terms are trademarks of the International Business Machines Corporation in the United States, other countries, or both:
AIX BladeCenter Calibrated Vectored Cooling DS4000 Dynamic Infrastructure Electronic Service Agent eServer IBM Systems Director Active Energy Manager IBM iDataPlex Netfinity PowerPC POWER Redbooks Redpaper Redbooks (logo) RETAIN ServerProven Smarter Planet System Storage System x System z Tivoli X-Architecture XIV xSeries
The following terms are trademarks of other companies: Java, and all Java-based trademarks are trademarks of Sun Microsystems, Inc. in the United States, other countries, or both. Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. Intel Xeon, Intel, Itanium, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries. Linux is a trademark of Linus Torvalds in the United States, other countries, or both. Other company, product, or service names may be trademarks or service marks of others.
xii
Preface
High-end workloads drive ever-increasing and ever-changing constraints. In addition to requiring greater memory capacity, these workloads challenge you to do more with less and to find new ways to simplify deployment and ownership. And although higher system availability and comprehensive systems management have always been critical, they have become even more important in recent years. Difficult challenges, such as these, create new opportunities for innovation. The IBM eX5 portfolio delivers this innovation. This family of high-end computing introduces the fifth generation of IBM X-Architecture technology. The family includes the IBM System x3850 X5, x3690 X5, and the IBM BladeCenter HX5. These servers are the culmination of more than a decade of x86 innovation and firsts that have changed the expectations of the industry. With this latest generation, eX5 is again leading the way as the shift toward virtualization, platform management, and energy efficiency accelerates. This book is divided into two parts. In the first part, we provide detailed technical information about the servers in the eX5 portfolio. This information is most useful in designing, configuring, and planning to order a server solution. In the second part of the book, we provide detailed configuration and setup information to get your servers operational. We focus particularly on setting up MAX5 configurations of all three eX5 servers as well as 2-node configurations of the x3850 X5 and HX5. This book is aimed at clients, IBM Business Partners, and IBM employees that want to understand the features and capabilities of the IBM eX5 portfolio of servers and want to learn how to install and configure the servers for use in production.
xiii
Rack, Power and Cooling. He is an IBM Certified IT Specialist and member of the IT Specialist Certification Review Board. Scott Haddow is a Presales Technical Support Specialist for IBM in the UK. He has 12 years of experience working with servers and storage. He has worked at IBM for six years, his experience spanning IBM Netfinity, xSeries, and now the System x brand. His areas of expertise include Fibre Channel fabrics. Michael Hurman is a Senior IT Specialist for IBM STG Lab Services in South Africa. He has more than 12 years of international experience in IT and has co-authored previous IBM Redbooks publications including Implementing the IBM BladeCenter S Chassis, SG24-7682. His areas of expertise include assisting clients to design and implement System x, BladeCenter, IBM Systems Director, midrange storage and storage area networks (SAN_-based solutions. He started his career at IBM in 2006. Jeneea Jervay (JJ) was a Technical Support Management Specialist in Raleigh at the time of writing this publication. She provided presales technical support to IBM Business Partners, clients, IBM Advanced Technical Support specialists, and IBM Field Technical Sales Support Specialists globally for the BladeCenter portfolio. She authored the IBM BladeCenter Interoperability Guide from 2007 to early 2010. She is a PMI Certified Project Manager and former System x and BladeCenter Top Gun instructor. She was the lead for the System x and BladeCenter Demand Acceleration Units (DAU) program. Previously, she was a member of the Americas System x and BladeCenter Brand team and the Sales Solution Center, which focused exclusively on IBM Business Partners. She started her career at IBM in 1995. Eric Kern is a Senior Managing Consultant for IBM STG Lab Services. He currently provides technical consulting services for System x, BladeCenter, System Storage, and Systems Software. Since 2007, he has helped clients design and implement x86 server and systems management software solutions. Prior to joining Lab Services, he developed software for the BladeCenters Advanced Management Module and for the Remote Supervisor Adapter II. He is a VMware Certified Professional and a Red Hat Certified Technician. Cynthia Knight is an IBM Hardware Design Engineer in Raleigh and has worked for IBM for 11 years. She is currently a member of the IBM eX5 design team. Previous designs include the Ethernet add-in cards for the IBM Network Processor Reference Platform and the Chassis Management Module for BladeCenter T. She was also the lead designer for the IBM BladeCenter PCI Expansion Units. Miroslav Peic is a System x Support Specialist in IBM Austria. He has a graduate degree in applied computer science and has many industry certifications, including the Microsoft Certified Systems Administrator 2003. He trains other IBM professionals and provides technical support to them, as well as to IBM Business Partners and clients. He has 10 years experience in IT and has worked at IBM since 2008. Tom Sorcic is an IT specialist and technical trainer for BladeCenter and System x support. He is part of Global Technology Enterprise Services at the Intel Smart Center in Atlanta, Georgia, where he started working for IBM in 2001. He has 37 years of international experience with IT in banking, manufacturing, and technical support. An author of hundreds of web pages, he continues his original role as core team member for the Global System x Skills Exchange (GLOSSE) website, assisting in the site design and providing technical content on a wide variety of topics since 2008. He is a subject matter expert in all forms of IBM ServeRAID hardware, Ethernet networking, storage area networks, and Microsoft high availability clusters.
xiv
Evans Tanurdin is an IT Specialist at IBM Global Technology Services in Indonesia. He provides technical support and services on the IBM System x, BladeCenter, and System Storage product lines. His technology focus areas include the design, operation, and maintenance services of enterprise x86 server infrastructure. Other significant experiences include application development, system analysis, and database design. Evans holds a degree in Nuclear Engineering from Gadjah Mada University (Indonesia), and certifications from Microsoft, Red Hat, and Juniper. The authors of this book were divided into two teams, Part 1 of the book is based on the IBM Redpaper IBM eX5 Portfolio Overview: IBM System x3850 X5, x3950 X5, x3690 X5, and BladeCenter HX5, REDP-4650, and written by one team of subject matter experts.
The team that wrote Part 1 (left to right): David, Duncan, JJ, Scott, Cynthia, and Eric
Part 2 of the book was written by a second team of subject matter experts. This team also provided updates to the first part of the book.
The team that wrote Part 2 (left to right): David, Evans, Aaron, Miro, Tom, and Mike
Thanks to the following people for their contributions to this project: From IBM Marketing: Mark Chapman Michelle Gottschalk Harsh Kachhy
Preface
xv
Richard Mancini Tim Martin Kevin Powell Heather Richardson David Tareen From IBM Development: Justin Bandholz Ralph Begun Jon Bitner Charles Clifton Candice Coletrane-Pagan David Drez Royce Espy Dustin Fredrickson Larry Grasso Dan Kelaher Randy Kolvick Chris LeBlanc Mike Schiskey Greg Sellman Mehul Shah Matthew Trzyna Matt Weber From the IBM Redbooks organization: Mary Comianos Linda Robinson Stephen Smith From other IBM employees throughout the world: Randall Davis, IBM Australia John Encizo, IBM U.S. Shannon Meier, IBM U.S. Keith Ott, IBM U.S. Andrew Spurgeon, IBM New Zealand Xiao Jun Wu, IBM China
xvi
Comments welcome
Your comments are important to us! We want our books to be as helpful as possible. Send us your comments about this book or other IBM Redbooks publications in one of the following ways: Use the online Contact us review Redbooks form found at: ibm.com/redbooks Send your comments in an email to: redbooks@us.ibm.com Mail your comments to: IBM Corporation, International Technical Support Organization Dept. HYTD Mail Station P099 2455 South Road Poughkeepsie, NY 12601-5400
Preface
xvii
xviii
Chapter 1.
Introduction
The IBM eX5 product portfolio represents the fifth generation of servers built upon Enterprise X-Architecture. Enterprise X-Architecture is the culmination of bringing generations of IBM technology and innovation derived from our experience in high-end enterprise servers. Now with eX5, IBM scalable systems technology for Intel processor-based servers has also been delivered to blades. These servers can be expanded on demand and configured by using a building block approach that optimizes system design servers for your workload requirements. As a part of the IBM Smarter Planet initiative, our Dynamic Infrastructure charter guides us to provide servers that improve service, reduce cost, and manage risk. These servers scale to more CPU cores, memory, and I/O than previous systems, enabling them to handle greater workloads than the systems they supersede. Power efficiency and machine density are optimized, making them affordable to own and operate. The ability to increase the memory capacity independently of the processors means that these systems can be highly utilized, yielding the best return from your application investment. These systems allow your enterprise to grow in processing, I/O, and memory dimensions, so that you can provision what you need now, and expand the system to meet future requirements. System redundancy and availability technologies are more advanced than the technologies that were previously available in the x86 systems. This chapter contains the following topics: 1.1, eX5 systems on page 2 1.2, Model summary on page 3 1.3, Positioning on page 7 1.4, Energy efficiency on page 10 1.5, Services offerings on page 11 1.6, What this book contains on page 11
Figure 1-1 eX5 family (top to bottom): BladeCenter HX5 (2-node), System x3690 X5, and System x3850 X5 (the System x3950 X5 looks the same as the x3850 X5)
The IBM System x3850 X5 and x3950 X5 are 4U highly rack-optimized servers. The x3850 X5 and the workload-optimized x3950 X5 are the new flagship servers of the IBM x86 server family. These systems are designed for maximum utilization, reliability, and performance for computer-intensive and memory-intensive workloads. The IBM System x3690 X5 is a new 2U rack-optimized server. This machine brings new features and performance to the middle tier, as well as a memory scalability option with MAX5. The IBM BladeCenter HX5 is a single-wide (30 mm) blade server that follows the same design as all previous IBM blades. The HX5 brings unprecedented levels of capacity to high-density environments. The HX5 is expandable to form either a 2-node system with four processors, or a single-node system with the MAX5 memory expansion blade. When compared to other machines in the System x portfolio, these systems represent the upper end of the spectrum, are suited for the most demanding x86 tasks, and can handle jobs which previously might have been run on other platforms. To assist with selecting the ideal system for a given workload, we have designed workload-specific models for virtualization and database needs.
Intel Xeon processors (two standard; maximum of four) E7520 4C 1.86 GHz, 18 MB L3, 95W c E7520 4C 1.86 GHz, 18 MB L3, 95W c E7530 6C 1.86 GHz,12 MB L3, 105W c E7540 6C 2.0 GHz, 18 MB L3, 105W X7550 8C 2.0 GHz, 18 MB L3, 130W X7560 8C 2.27 GHz, 24 MB L3, 130W
Memory speed (MHz) 800 MHz 800 MHz 978 MHz 1066 MHz 1066 MHz 1066 MHz
a. The x character in the seventh position of the machine model denotes the region-specific character. For example, U indicates US, and G indicates EMEA. b. Emulex 10Gb Ethernet Adapter is installed in PCIe slot 7. c. Any model using the E7520 or E7530 CPU cannot scale beyond single-node 4-way, even with the addition of MAX5.
Model 5Dx
Model 5Dx is designed for database applications and uses solid-state drives (SSDs) for the best I/O performance. Backplane connections for eight 1.8-inch SSDs are standard and there is space for an additional eight SSDs. The SSDs themselves must be ordered separately. Because no SAS controllers are standard, you can select from the available cards as described in 3.9, Storage on page 90.
Model 4Dx
Model 4Dx is designed for virtualization and is fully populated with 4 GB memory dual inline memory modules (DIMMs), including in an attached MAX5 memory expansion unit, a total of 384 GB of memory. Backplane connections for four 2.5-inch serial-attached SCSI (SAS) hard disk drives (HDDs) are standard; however, the SAS HDDs themselves must be ordered separately. A ServeRAID BR10i SAS controller is standard in this model.
Chapter 1. Introduction
Table 1-2 Models of the x3950 X5: Workload-optimized models Power supplies (std/max) 2/2 2/2 10Gb Ethernet standardb Memory cards (std/max)
Modela
Memory speed
MAX5
Standard memory
Database workload-optimized models 7145-5Dx X7560 8C 2.27 GHz, 24 MB L3, 130W 1066 MHz Opt Server: 8x 4GB 4/8 No Yes 8x 1.8/16c
Virtualization workload-optimized models 7145-4Dx 4x X7550 8C 2.0 GHz, 18 MB L3, 130W 1066 MHz Std Server: 64x 4GB MAX5: 32x 4GB 8/8 Yes Yes 4x 2.5/8
a. The x character in the seventh position of the machine model denotes the region-specific character. For example, U indicates US, and G indicates EMEA. b. Emulex 10Gb Ethernet Adapter is installed in PCIe slot 7. c. Includes, as standard, one 8-bay eXFlash SSD backplane; one additional eXFlash backplane is optional.
Intel Xeon processors (four standard and max) 4x E7530 6C 1.86 GHz, 12 MB L3, 105W c 4x X7550 8C 2.0 GHz, 18 MB L3, 130W 4x X7560 8C 2.27 GHz, 24 MB L3, 130W
Standard memory (MAX5 is standard) Server: 8x 4 GB MAX5: 2x 4 GB Server: 8x 4 GB MAX5: 2x 4 GB Server: 8x 4 GB MAX5: 2x 4 GB
a. The x character in the seventh position of the machine model denotes the region-specific character. For example, U indicates US, and G indicates EMEA. b. Emulex 10Gb Ethernet Adapter is installed in PCIe slot 7. c. Any model using the E7520 or E7530 CPU cannot scale beyond single-node 4-way, even with the addition of MAX5.
Table 1-4 x3690 X5 models ServeRAID M1015 standard 10Gb Ethernet standardb
Intel Xeon processors (two maximum) 1x E7520 4C, 1.86 GHz, 95W 1x E7520 4C, 1.86 GHz, 95W 1x E6540 6C, 2.00 GHz, 105W 1x X6550 8C, 2.00 GHz, 130W 1x X6550 8C, 2.00 GHz, 130W 1x X7560 8C, 2.26 GHz, 130W 1x X7550 8C, 2.00GHz, 130W 1x X7560 8C, 2.26GHz, 130W
Memory tray
Memory speed 800 MHz 800 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz
Standard memorya Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB MAX5: 2x 4GB Server: 2x 4GB MAX5: 2x 4GB
Power supplies std/max 1/4 1/4 1/4 1/4 1/4 1/4 Server: 2/4 MAX5: 1/2 Server: 2/4 MAX5: 1/2
Drive bays std/max None 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16
a. Up to 64 DIMM sockets: Each server has 16 DIMM sockets standard or 32 sockets with the addition of the internal memory tray (mezzanine). With the addition of the MAX5 memory expansion unit, 64 DIMM sockets in total are available. b. Emulex 10Gb Ethernet Adapter.
Chapter 1. Introduction
Table 1-5 x3690 X5 workload-optimized models Memory tray ServeRAID M1015 std 10Gb Eth standardb
Model
Memory speed
MAX5
Standard memorya
Database workload-optimized models 7148-3Dx 2x X6550 8C, 2.00 GHz, 130W 1066 MHz Opt Server: 4x 4 GB Std Opt Opt Server: 4/4 16x 1.8/32
Virtualization workload-optimized models 7148-2Dx 2x E6540 6C, 2.00 GHz, 105W 1066 MHz Std Server: 32x 4GB MAX5: 32x 4GB Std Opt Std Server: 4/4 MAX5: 2/2 4x 2.5/16
a. Up to 64 DIMM sockets: Each server has 16 DIMM sockets standard or 32 sockets with the addition of the internal memory tray (mezzanine). With the addition of the MAX5 memory expansion unit, a total of 64 DIMM sockets are available. b. Emulex 10Gb Ethernet Adapter.
1.86 GHz 1.86 GHz 1.86 GHz 2.00 GHz 2.00 GHz 2.00 GHz
2x 4 GB 2x 4 GB 2x 4 GB 2x 4 GB 2x 4 GB HX5: 4x 4GB MAX5: None HX5: 4x 4GB MAX5: None HX5: 4x 4GB MAX5: None HX5: 4x 4GB MAX5: None 2x 4 GB
7872-6Dx
2x E6540 6C/2
2.00 GHz
105W
978 MHz
1066 MHz
Std
No
Std
7872-83x
2x X6550 8C/2
2.00 GHz
130W
978 MHz
1066 MHz
Std
No
Opt
7872-84x 7872-86x
130W 130W
Std Opt
No Yes
Opt Std
a. This column lists worldwide, generally available variant (GAV) model numbers. They are not orderable as listed and must be modified by country. The US GAV model numbers use the following nomenclature: xxU. For example, the US orderable part number for 7870-A2x is 7870-A2U. See the product-specific official IBM announcement letter for other country-specific GAV model numbers. b. Emulex Virtual Fabric Adapter Expansion Card (CFFh) c. The HX5 has 16 DIMM sockets and can hold 128 GB using 8 GB memory DIMMs. The MAX5 has 24 DIMM sockets and can hold 192 GB using 8 GB memory DIMMs. A 1-node HX5 + MAX5 supports 320 GB total using 8 GB DIMMs.
Also available is a virtualization workload-optimized model of these HX5s. This is a pre-configured, pre-tested model targeted at large-scale consolidation. Table 1-7 shows the model.
Table 1-7 Workload-optimized models of the HX5 Model Intel Xeon model and cores/max Clock speed TDP HX5 max memory speeda MAX5 Scalable to four socket 10GbE cardb Standard memory (max 320 GB)c
Virtualization workload-optimized models (includes VMware ESXi 4.1 on a USB memory key) 7872-68x 2x E6540 6C/2 2.00 GHz 105 W 978 MHz Std No Std 160 GB HX5: 16x 4GB MAX5: 24x 4GB
a. Memory speed of the HX5 is dependent on the processor installed; however, the memory speed of the MAX5 is up to 1066 MHz irrespective of the processor installed in the attached HX5. b. Emulex Virtual Fabric Adapter Expansion Card (CFFh). c. HX5 has 16 DIMM sockets and can hold 128 GB using 8 GB memory DIMMs. MAX5 has 24 DIMM sockets and can hold 192 GB using 8 GB memory DIMMs. A 1-node HX5 + MAX5 supports 320 GB total using 8 GB DIMMs.
Model 7872-68x is a virtualization-optimized model and includes the following features in addition to standard HX5 and MAX5 features: Forty DIMM sockets, all containing 4 GB memory DIMMs for a total of 160 GB of available memory. VMware ESXi 4.1 on a USB memory key is installed internally in the server. See 5.15, Integrated virtualization on page 214 for details. Emulex Virtual Fabric Adapter Expansion Card (CFFh).
1.3 Positioning
Table 1-8 gives an overview of the features of the systems that are described in this book.
Table 1-8 Maximum configurations for the eX5 systems Maximum configurations Processors 1-node 2-node Memory 1-node 1-node with MAX5 2-node Disk drives (non-SSD)c SSDs 1-node 2-node 1-node 2-node Standard 1 Gb Ethernet interfaces 1-node 2-node x3850 X5/x3950 X5 4 8 1024 GB (64 DIMMs)a 1536 GB (96 DIMMs)a 2048 GB (128 DIMMsa 8 16 16 32 2d 4 x3690 X5 2 Not available 512 GB (32 DIMMs)b 1024 GB (64 DIMMs)b Not available 16 Not available 24 Not available 2 Not available HX5 2 4 128 GB (16 DIMMs) 320 GB (40 DIMMs) 256 GB (32 DIMMs) Not available Not available 2 4 2 4
Chapter 1. Introduction
x3850 X5/x3950 X5 2 4
HX5 0 0
a. Requires full processors in order to install and use all memory. b. Requires that the memory mezzanine board is installed along with processor 2. c. For the x3690 X5 and x3850 X5, additional backplanes might be needed to support these numbers of drives. d. Depends on the model. See Table 3-2 on page 64 for the IBM System x3850 X5.
Figure 1-2 IBM System x3850/x3950 X5 with the MAX5 memory expansion unit attached
Two x3850/x3950 X5 servers can be connected for a single system image with a max of eight processors and 2 TB of RAM. Table 1-9 compares the number of processor sockets, cores, and memory capacity of the eX4 and eX5 systems.
Table 1-9 Comparing the x3850 M2 and x3950 M2 with the eX5 servers Processor sockets Previous generation servers (eX4) x3850 M2 x3950 M2 4 4 24 24 256 GB 256 GB Processor cores Maximum memory
Processor sockets x3950 M2 2-node Next generation server (eX5) x3850/x3950 X5 x3850/x3950 X5 2-node x3850/x3950 X5 with MAX5 4 8 4 8
Processor cores 48
32 64 32
a. You must install two processors and the memory mezzazine to use the full memory capacity.
Next generation server (eX5) HX5 (30 mm) HX5 2-node (60 mm) HX5 with MAX5 2 4 2 16 32 16 128 GB 256 GB 320 GB
10
Dynamic fan speeds: In the event of a fan failure, the other fans run faster to compensate until the failing fan is replaced. Regular fans must run faster at all times, just in case, thereby wasting power. Although these systems provide incremental gains at the individual server level, the eX5 systems can have an even greater green effect in your data center. The gain in computational power and memory capacity allows for application performance, application consolidation, and server virtualization at greater degrees than previously available in x86 servers.
11
12
Part 1
Part
Product overview
In this first part of the book, we provide detailed technical information about the servers in the eX5 portfolio. This information is most useful in designing, configuring, and planning to order a server solution. This part consists of the following chapters: Chapter 2, IBM eX5 technology on page 15 Chapter 3, IBM System x3850 X5 and x3950 X5 on page 55 Chapter 4, IBM System x3690 X5 on page 117 Chapter 5, IBM BladeCenter HX5 on page 177
13
14
Chapter 2.
15
16
Table 2-1 Two-socket, 2-socket scalable, and 4-socket scalable Intel processors minimum configuration Xeon 5500 Used in x3400 M2 x3500 M2 x3550 M2 x3650 M2 HS22 HS22V Nehalem-EP 2 2 or 4 4 or 8 MB
a
Intel development name Maximum processors per server CPU cores per processor Last level cache (MB) Memory DIMMs per processor (maximum)
Nehalem-EX 2 4, 6, or 8 12 or 18 MB 16
a
a. Requires that the memory mezzanine board is installed along with processor two on x3690 X5
For more information about processor options and the installation order of the processors, see the following links: IBM System x3850 X5: 3.7, Processor options on page 74 IBM System x3690 X5: 4.7, Processor options on page 130 IBM BladeCenter HX5: 5.9, Processor options on page 192
17
Hyper-Threading Technology is designed to improve server performance by exploiting the multi-threading capability of operating systems and server applications in such a way as to increase the use of the on-chip execution resources available on these processors. Applications types that make the best use of Hyper-Threading are virtualization, databases, email, Java, and web servers. For more information about Hyper-Threading Technology, go to the following website: http://www.intel.com/technology/platform-technology/hyper-threading/
18
Processor
Processor
Processor
Processor
Memory
I/O
Figure 2-1 Shared front-side bus, in the IBM x360 and x440; with snoop filter in the x365 and x445
The front-side bus carries all reads and writes to the I/O devices, and all reads and writes to memory. Also, before a processor can use the contents of its own cache, it must know whether another processor has the same data stored in its cache. This process is described as snooping the other processors caches, and it puts a lot of traffic on the front-side bus. To reduce the amount of cache snooping on the front-side bus, the core chip set can include a
snoop filter, which is also referred to as a cache coherency filter. This filter is a table that
keeps track of the starting memory locations of the 64-byte chunks of data that are read into cache, called cache lines, or the actual cache line itself, and one of four states: modified, exclusive, shared, or invalid (MESI). The next step in the evolution was to divide the load between a pair of front-side buses, as shown in Figure 2-2. Servers that implemented this design include the IBM System x3850 and x3950 (the M1 version).
Processor
Processor
Processor
Processor
Memory
I/O
Figure 2-2 Dual independent buses, as in the x366 and x460 (later called the x3850 and x3950)
This approach had the effect of reducing congestion on each front-side bus, when used with a snoop filter. It was followed by independent processor buses, shown in Figure 2-3 on page 20. Servers implementing this design included the IBM System x3850 M2 and x3950 M2.
19
Processor
Processor
Processor
Processor
Memory
I/O
Instead of a parallel bus connecting the processors to a core chip set, which functions as both a memory and I/O controller, the Xeon 6500 and 7500 family processors implemented in IBM eX5 servers include a separate memory controller to each processor. Processor-to-processor communications are carried over shared-clock, or coherent QPI links, and I/O is transported over non-coherent QPI links through I/O hubs. Figure 2-4 shows this information.
I/O Hub
I/O
Memory
Processor
Processor
Memory
Memory
Processor
Processor
Memory
I/O Hub
I/O
In previous designs, the entire range of memory was accessible through the core chip set by each processor, a shared memory architecture. This design creates a non-uniform memory access (NUMA) system, in which part of the memory is directly connected to the processor where a given thread is running, and the rest must be accessed over a QPI link through another processor. Similarly, I/O can be local to a processor, or remote through another processor. For QPI use, Intel has modified the MESI cache coherence protocol to include a forwarding state, so when a processor asks to copy a shared cache line, only one other processor responds. For more information about QPI, go to the following website: http://www.intel.com/technology/quickpath/ 20
IBM eX5 Implementation Guide
Figure 2-5 Setup (F1) System Settings Operating Modes to set Performance Mode
21
2.3 Memory
In this section, we describe the major features of the memory subsystem in eX5 systems. We describe the following topics in this section: 2.3.1, Memory speed on page 22 2.3.2, Memory DIMM placement on page 23 2.3.3, Memory ranking on page 24 2.3.4, Nonuniform memory architecture (NUMA) on page 26 2.3.5, Hemisphere Mode on page 26 2.3.6, Reliability, availability, and serviceability (RAS) features on page 28 2.3.7, I/O hubs on page 30
22
What about 1333 MHz? The maximum memory speed that is supported by Xeon 7500 and 6500 processors is 1066 MHz (1333 MHz is not supported). Although the 1333 MHz DIMMs are still supported, they can operate at a maximum speed of 1066 MHz.
23
Memory configurations
Relative performance
Each processor: 2 memory controllers 2 DIMMs per channel 8 DIMMs per MC Each processor: 2 memory controllers 1 DIMM per channel 4 DIMMs per MC Each processor: 2 memory controllers 2 DIMMs per channel 4 DIMMs per MC Each processor: 2 memory controllers 1 DIMM per channel 2 DIMMs per MC Each processor: 1 memory controller 2 DIMMs per channel 8 DIMMs per MC Each processor: 1 memory controller 1 DIMM per channel 4 DIMMs per MC Each processor: 1 memory controller 2 DIMMs per channel 4 DIMMs per MC Each processor: 1 memory controller 1 DIMM per channel 2 DIMMs per MC
1 2 3 4 5 6 7 8
1.0
Memory card DIMMs Channel Memory buffer SMI link Memory controller
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
0.94
Mem Ctrl 1
0.61
1
1 0.94
Mem Ctrl 1
Mem Ctrl 2
0.58 0.51
0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 1 2 3 4 5 6 7 8 Configuration
0.61 0.58 0.51 0.47 0.31
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
0.47 0.31
0.29
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
0.29
Figure 2-6 Relative memory performance based on DIMM placement (one processor and two memory cards shown)
24
extend the pages across ranks, thereby making the pages effectively larger and therefore creating more page-hit cycles. We used three types of memory DIMMs for this analysis: Four GB 4Rx8 (four ranks using x8 DRAM technology) Two GB 2Rx8 (two ranks) One GB 1Rx8 (one rank) We used the following memory configurations: Fully populated memory: Two DIMMs on each memory channel Eight DIMMs per memory card Half-populated memory: One DIMM on each memory channel Four DIMMs per memory card (slots 1, 3, 6, and 8; see Figure 3-16 on page 76) Quarter-populated memory: One DIMM on just half of the memory channels Two DIMMs per memory card Although several benchmarks were conducted, this section focuses on the results gathered using the industry-standard STREAM benchmark, as shown in Figure 2-7.
Figure 2-7 Comparing the performance of memory DIMM configurations using STREAM
Taking the top performance result of 16x 4 GB quad-rank DIMMs as the baseline, we see how the performance drops to 95% of the top performance with 16x 2 GB dual-rank DIMMs, and 89% of the top performance with 16x 1 GB single-rank DIMMs. You can see similar effects across the three configurations based on eight DIMMs per processor and four DIMMs per processor. These results also emphasize the same effect that is shown in 3.8.3, Maximizing memory performance on page 84 for the x3850 X5, where performance drops away dramatically when all eight memory channels per CPU are not used.
25
Tip: Additional ranks increase the memory bus loading, which is why on Xeon 5500 (Nehalem EP) platforms, the opposite effect can occur: memory slows down if too many rank loads are attached. The use of scalable memory buffers in the x3850 X5 with Xeon 7500/6500 processors avoids this slowdown.
on each of the attached memory buffers), DIMMs must be installed in quantities of four per processor to enable Hemisphere Mode. In addition, because eight DIMMs per processor are required for using all memory channels, eight DIMMs per processor need to be installed at a time for optimized memory performance. Failure to populate all eight channels on a processor can result in a performance reduction of approximately 50%. Hemisphere Mode does not require that the memory configuration of each CPU is identical. For example, Hemisphere Mode is still enabled if CPU 0 is configured with 8x 4 GB DIMMs, and processor 1 is configured with 8x 2 GB DIMMs. Depending on the application characteristics, however, an unbalanced memory configuration can cause reduced performance because it forces a larger number of remote memory requests over the inter-CPU QPI links to the processors with more memory. We summarize these points: There are two memory buffers per memory channel, two channels per memory controller, and two controllers per processor. Each memory channel must contain at least one DIMM to enable Hemisphere Mode. Within a processor, both memory controllers need to contain identical DIMM configurations to enable Hemisphere Mode. Therefore, for best results, install at least eight DIMMs per processor. Industry-standard tests run on one Xeon 7500 processor with various memory configurations have shown that there are performance implications if Hemisphere Mode is not enabled. For example, for a configuration with eight DIMMs installed and spread across both memory controllers in a processor and all memory buffers (see Figure 2-8), there is a drop in performance of 16% if Hemisphere Mode is not enabled.
Memory controller
Memory controller
Buffer
Buffer
Buffer
Buffer
DIMM DIMM
DIMM DIMM
For more information about Hemisphere Mode installation options, see the following sections: IBM System x3850 X5: 3.8.2, DIMM population sequence on page 79 IBM System x3690 X5: Two processors with memory mezzanine installed on page 135 IBM BladeCenter HX5: 5.10.2, DIMM population order on page 196
27
Memory mirroring
To further improve memory reliability and availability beyond error checking and correcting (ECC) and Chipkill, the chip set can mirror memory data to two memory ports. To successfully enable mirroring, you must have both memory cards per processor installed, and populate the same amount of memory in both memory cards. Partial mirroring (mirroring of part but not all of the installed memory) is not supported. Memory mirroring, or full array memory mirroring (FAMM) redundancy, provides the user with a redundant copy of all code and data addressable in the configured memory map. Memory mirroring works within the chip set by writing data to two memory ports on every memory-write cycle. Two copies of the data are kept, similar to the way RAID-1 writes to disk. Reads are interleaved between memory ports. The system automatically uses the most reliable memory port as determined by error logging and monitoring. If errors occur, only the alternate memory port is used, until bad memory is replaced. Because a redundant copy is kept, mirroring results in having only half the installed memory available to the operating system. FAMM does not support asymmetrical memory configurations and requires that each port is populated in identical fashion. For example, you must install 2 GB of identical memory equally and symmetrically across the two memory ports to achieve 1 GB of mirrored memory. FAMM enables other enhanced memory features, such as Unrecoverable Error (UE) recovery, and is required for support of memory hot replace. Memory mirroring is independent of the operating system. For more information about system-specific memory mirroring installation options, see the following sections: x3850 X5: 3.8.4, Memory mirroring on page 87 x3690 X5: 4.8.6, Memory mirroring on page 141 BladeCenter HX5: 5.10.4, Memory mirroring on page 200
28
Memory sparing
Sparing provides a degree of redundancy in the memory subsystem, but not to the extent of mirroring. In contrast to mirroring, sparing leaves more memory for the operating system. In sparing mode, the trigger for failover is a preset threshold of correctable errors. Depending on the type of sparing (DIMM or rank), when this threshold is reached, the content is copied to its spare. The failed DIMM or rank is then taken offline, with the spare counterpart activated for use. There are two sparing options: DIMM sparing Two unused DIMMs are spared per memory card. These DIMMs must have the same rank and capacity as the largest DIMMs that we are sparing. The size of the two unused DIMMs for sparing is subtracted from the usable capacity that is presented to the operating system. DIMM sparing is applied on all memory cards in the system. Rank sparing Two ranks per memory card are configured as spares. The ranks have to be as large as the rank relative to the highest capacity DIMM that we are sparing. The size of the two unused ranks for sparing is subtracted from the usable capacity that is presented to the operating system. Rank sparing is applied on all memory cards in the system. You configure these options by using the UEFI during start-up. For more information about system-specific memory sparing installation options, see the following sections: IBM System x3850 X5: 3.8.5, Memory sparing on page 89 IBM System x3690 X5: 4.8.7, Memory sparing on page 143 IBM BladeCenter HX5: 5.10.5, Memory sparing on page 202
Chipkill
Chipkill memory technology, an advanced form of error checking and correcting (ECC) from IBM, is available for the eX5 blade. Chipkill protects the memory in the system from any single memory chip failure. It also protects against multi-bit errors from any portion of a single memory chip.
29
RBS operates automatically without issuing a Predictive Failure Analysis (PFA) or light path diagnostics alert to the administrator, although an event is logged to the service processor log. After the second DIMM failure, PFA and light path diagnostics alerts occur on that DIMM normally.
Lock step
IBM eX5 memory operates in lock step mode. Lock step is a memory protection feature that involves the pairing of two memory DIMMs. The paired DIMMs can perform the same operations and the results are compared. If any discrepancies exist between the results, a memory error is signaled. Lock step mode gives a maximum of 64 GB of usable memory with one CPU installed, and 128 GB of usable memory with two CPUs installed (using 8 GB DIMMs). Memory must be installed in pairs of two identical DIMMs per processor. Although the size of the DIMM pairs installed can differ, the pairs must be of the same speed.
30
For more information regarding each of the eX5 systems and the available I/O adapters, see the following sections: IBM System x3850 X5: 3.12, I/O cards on page 104. IBM System x3690 X5: 4.10.4, I/O adapters on page 168. IBM BladeCenter HX5: 5.13, I/O expansion cards on page 209.
2.4 MAX5
Memory Access for eX5 (MAX5) is the name give to the memory and scalability subsystem that can be added to eX5 servers. In the Intel QPI specification, the MAX5 is a node controller. MAX5 for the rack-mounted systems (x3850 X5, x3950 X5, and x3690 X5) is in the form of a 1U device that attaches beneath the server. For the BladeCenter HX5, MAX5 is implemented in the form of an expansion blade that adds 30 mm to the width of the blade (the width of one extra blade bay). Figure 2-9 shows the HX5 with the MAX5 attached.
31
Figure 2-10 IBM System x3850 X5 with MAX5 (the MAX5 is the 1U unit beneath the main system)
MAX5 connects to these systems through QPI links and provides the EXA scalability interfaces. The eX5 chip set, which is described in 2.1, eX5 chip set on page 16, is contained in the MAX5 units.
32
Table 2-2 through Table 2-4 show the memory capacity and bandwidth increases that are possible with MAX5 for the HX5, x3690 X5, and x3850 X5.
Table 2-2 HX5 compared to HX5 with MAX5 HX5 Memory bandwidth Memory capacity 16 DDR3 channels at 978 MHz 128 GB using 8 GB DIMMs HX5 with MAX5 16 DDR3 channels at 978 MHz + 12 DDR3 channels at 1066 MHz 320 GB using 8 GB DIMMs
Table 2-3 x3690 X5 compared to x3690 X5 with MAX5 x3690 X5 Memory bandwidth Memory capacity 32 DDR3 channels at 1066 MHza 512 GB using 16 GB DIMMs x3690 with MAX5 64 DDR3 channels at 1066 MHz 1 TB using 16 GB DIMMs
a. Must install optional mezzanine board Table 2-4 x3850 X5 compared to x3850 X5 with MAX5 x3850 X5 Memory bandwidth Memory capacity 32 DDR3 channels at 1066 MHz 1 TB using 16 GB DIMMs x3850 X5 with MAX5 48 DDR3 channels at 1066 MHz 1.5 TB using 16 GB DIMMs
For more information about system-specific MAX5 installation options, see the following sections: IBM System x3850 X5: MAX5 memory on page 79 IBM System x3690 X5: 4.8.3, MAX5 memory on page 136 IBM BladeCenter HX5: MAX5 memory population order on page 198
2.5 Scalability
The architecture of the eX5 servers permits system scaling of up to two nodes on HX5 and x3850 X5. The architecture also supports memory scaling. Figure 2-12 shows these types of scaling.
Server
QPI Scaling
Server
QPI Scaling
MAX5
Server
Memory scaling
Figure 2-12 Types of scaling with eX5 systems
System scaling
The x3850 X5 and HX5 both support 2-node system scaling or memory scaling. The x3690 X5 supports memory scaling only.
33
As shown in Figure 2-12 on page 33, the following scalability is possible: Memory scaling: A MAX5 unit can attach to an eX5 server through QPI link cables. This method provides the server with additional memory DIMM slots. We refer to this combination as a memory-enhanced system. All eX5 systems support this scaling. System scaling: Two servers can connect to form a single system image. The connections are formed by using QPI link cables. The x3850 X5 and HX5 support this type of scaling. For more information about system-specific scaling options, see the following sections: IBM System x3850 X5: 3.6, Scalability on page 70 IBM System x3690 X5: 4.6, Scalability on page 128 BladeCenter HX5: 5.8, Scalability on page 188
2.6 Partitioning
You can operate the HX5 scaled system as two independent systems or as a single system, without removing the blade and taking off the side-scale connector. This capability is called partitioning and is referred to as IBM FlexNode technology. You partition by using the Advanced Management Module (AMM) in the IBM BladeCenter chassis for the HX5. Figure 2-13 depicts an HX5 system that is scaled to two nodes and an HX5 system that is partitioned into two independent servers.
x3690 X5 and x3850 X5: The x3690 X5 and x3850 X5 do not support partitioning. Figure 2-14 on page 35 and Figure 2-15 on page 35 show the scalable complex configuration options for stand-alone mode through the Advanced Management Module of the BladeCenter chassis.
34
Figure 2-14 Shows option for putting a partition into stand-alone mode
The AMM can be accessed remotely, so partitioning can be done without physically touching the systems. Partitioning can allow you to qualify two system types with little additional work, and it allows you more flexibility in system types for better workload optimization. The HX5 blade, when scaled as a 2-node (4-socket) system, supports FlexNode partitioning as standard. Before a 2-node HX5 solution can be used, you must create a partition. When the scalability card is added, the two blades still act as single nodes until a partition is made. For more information about creating a scalable complex, see 8.6, Creating an HX5 scalable complex on page 402.
35
Figure 2-17 on page 37 shows the UEFI System Configuration and Boot Management window.
36
Choose System Settings to access the system settings options that we will describe here, as shown in Figure 2-18 on page 38.
37
For more information about system-specific UEFI options, see the following sections: IBM System x3850 X5: 6.9, UEFI settings on page 259 IBM System x3690 X5: 7.8, UEFI settings on page 337 IBM BladeCenter HX5: 8.5, UEFI settings on page 396
38
Acoustic Mode
Figure 2-20 shows the Acoustic Mode predetermined values. They emphasize power-saving server operation for generating less heat and noise. In turn, the system is able to lower the fan speed of the power supplies and the blowers by setting the processors, QPI link, and memory subsystem to a lower working frequency. Acoustic Mode provides lower system acoustics, less heat, and the lowest power consumption at the expense of performance.
Efficiency Mode
Figure 2-21 on page 40 shows the Efficiency Mode predetermined values. This operating mode provides the best balance between server performance and power consumption. In short, Efficiency Mode gives the highest performance per watt ratio.
39
Performance Mode
Figure 2-22 shows the Performance Mode predetermined values. The server is set to use the maximum performance limits within UEFI. These values include turning off several power management features of the processor to provide the maximum performance from the processors and memory subsystem. Performance Mode provides the best system performance at the expense of power efficiency.
Performance Mode is also a good choice when the server runs virtualization workloads. Servers that are used as physical hosts in virtualization clusters often have similar power consumption management at the operating system level. In this environment, the operating system makes decisions about moving and balancing virtual servers across an array of physical host servers. Each virtualized guest operating system reports to a single cluster controller about the resources usage and demand on that physical server. The cluster
40
controller makes decisions to move virtual servers between physical hosts to cater to each guest OS resource requirement and, when possible, shuts down unneeded physical hosts to save power. Aggressive power management at the hardware level can interfere with the OS-level power management, resulting in a common occurrence where virtual servers move back and forth across the same set of physical host servers. This situation results in an inefficient virtualization environment that responds slowly and consumes more power than necessary. Running the server in Performance Mode prevents this occurrence, in most cases.
Custom Mode
The default value that is set in new eX5 systems is Custom Mode, as shown in Figure 2-23. It is the recommended factory default setting. The values are set to provide optimal performance with reasonable power consumption. However, this mode allows the user to individually set the power-related and performance-related options. See 2.7.3, Performance-related individual system settings on page 43 for a description of individual settings.
Table 2-5 shows comparisons of the available operating modes of IBM eX5 servers. Using the Custom Mode, it is possible to run the system using properties that are in-between the predetermined operating modes.
Table 2-5 Operating modes comparison Settings Memory Speed CKE Low Power Proc Performance States C1 Enhanced Mode CPU C-States Efficiency Power Efficiency Enabled Enabled Enabled Enabled Acoustics Minimal Power Enabled Enabled Enabled Enabled Performance Max Performance Disabled Enabled Disabled Disabled Custom (Default) Max Performance Disable Enable Enable Enable
41
Settings QPI Link Frequency Turbo Mode Turbo Boost Power Optimization
Additional Settings
In addition to the Operating Mode selection, the UEFI settings under Operating Modes include these additional settings: Quiet Boot (Default: Enable) This mode enables system booting with less information displayed. Halt On Severe Error (Default: Disable, only available in System x3690 X5) This mode enables system boot halt when a severe error event is logged.
42
Processors
Processor settings control the various performance and power features that are available on the installed Xeon processor. Figure 2-25 on page 44 shows the UEFI Processor system settings window with the default values.
43
The following processor feature options are available: Turbo Mode (Default: Enable) This mode enables the processor to increase its clock speed dynamically as long as the CPU does not exceed the Thermal Design Power (TDP) for which it was designed. See 2.2.3, Turbo Boost Technology on page 18 for more information. Turbo Boost Power Optimization (Default: Power Optimized) This option specifies which algorithm to use when determining whether to overclock the processor cores in Turbo Mode: Power Optimized provides reasonable Turbo Mode in relation to power consumption. Turbo Mode does not engage unless additional performance has been requested by the operating system for a period of 2 seconds. Traditional provides a more aggressive Turbo Mode operation. Turbo Mode engages as more performance is requested by the operating system. Processor Performance States (Default: Enable) This option enables Intel Enhanced SpeedStep Technology that controls dynamic processor frequency and voltage changes, depending on operation. CPU C-States (Default: Enable) This option enables dynamic processor frequency and voltage changes in the idle state, providing potentially better power savings. C1 Enhanced Mode (Default: Enable) This option enables processor cores to enter an enhanced halt state to lower the voltage requirement, and it provides better power savings.
44
Hyper-Threading (Default: Enable) This option enables logical multithreading in the processor, so that the operating system can execute two threads simultaneously for each physical core. Execute Disable Bit (Default: Enable) This option enables the processor to disable the execution of certain memory areas, therefore preventing buffer overflow attacks. Intel Virtualization Technology (Default: Enable) This option enables the processor hardware acceleration feature for virtualization. Processor Data Prefetch (Default: Enable) This option enables the memory data access prediction feature to be stored in the processor cache. Cores in CPU Package (Default: All) This option sets the number of processor cores to be activated. QPI Link Frequency (Default: Max Performance) This option sets the operating frequency of the processors QPI link: Minimal Power provides less performance for better power savings. The QPI link operates at the lowest frequency, which, in the eX5 systems, is 4.8 GT/s. Power Efficiency provides best performance per watt ratio. The QPI link operates 1 step under the rated frequency, that is, 5.86 GT/s for processors rated at 6.4 GT/s. Max Performance provides the best system performance. The QPI link operates at the rated frequency, that is, 6.4 GT/s for processors rated at 6.4 GT/s.
Memory
The Memory settings window provides the available memory operation options, as shown in Figure 2-26 on page 46.
45
The following memory feature options are available: Memory Spare Mode (Default: Disable) This option enables memory sparing mode, as described in Memory sparing on page 29. Memory Mirror Mode (Default: Non-Mirrored) This option enables memory mirroring mode, as described in Memory mirroring on page 28. Memory Mirror Mode: Memory Mirror Mode cannot be used in conjunction with Memory Spare Mode. Memory Speed (Default: Max Performance) This option sets the operating frequency of the installed DIMMs: Minimal Power provides less performance for better power savings. The memory operates at the lowest supported frequency, which, in the eX5 systems, is 800 MHz. Power Efficiency provides the best performance per watt ratio. The memory operates one step under the rated frequency, that is, 977 MHz for DIMMs that are rated at 1066 MHz or higher. Max Performance provides the best system performance. The memory operates at the rated frequency, that is, 1066 MHz for DIMMs rated at 1066 MHz or higher. Tip: Although memory DIMMs rated at 1333MHz are supported on eX5 servers, the currently supported maximum memory operating frequency is 1066 MHz.
46
CKE Low Power (Default: Disable) This option enables the memory to enter a low-power state for power savings by reducing the signal frequency. Patrol Scrub (Default: Disable) This option enables scheduled background memory scrubbing before any error is reported, as opposed to default demand scrubbing on an error event. This option provides better memory subsystem resiliency at the expense of a small performance loss. Memory Data Scrambling (Default: Disable) This option enables a memory data scrambling feature to further minimize bit-data errors. Spread Spectrum (Default: Enable) This option enables the memory spread spectrum feature to minimize electromagnetic signal interference in the system. Page Policy (Default: Closed) This option determines the Page Manager Policy in evaluating memory access: Closed: Memory pages are closed immediately after each transaction. Open: Memory pages are left open for a finite time after each transaction for possible recurring access. Adaptive: Use Adaptive Page Policy to decide the memory page state. Multi-CAS Widget: The widget allows multiple consecutive column address strobes (CAS) to the same memory ranks and banks in the Open Page Policy. Mapper Policy (Default: Closed) This option determines how memory pages are mapped to the DIMM subsystem. Closed: Memory is mapped closed to prevent DIMMs from being excessively addressed. Open: Memory is mapped open to decrease latency. Scheduler Policy (Default: Adaptive) This option determines the scheduling mode optimization based on memory operation: Static Trade Off: Equal trade-off between read/write operation latency Static Read Primary: Minimize read latency and consider reads as primary operation Static Write Primary: Minimize write latency and consider writes as primary operation Adaptive: Memory scheduling adaptive to system operation
MAX5 Memory Scaling Affinity (Default: Non-Pooled) The Non-Pooled option splits the memory in the MAX5 and assigns it to each of the installed processors. The Pooled option presents the additional memory in the MAX5 as a pool of memory that is not assigned to any particular processor.
47
Each eXFlash can replace four 2.5-inch serial-attached SCSIs (SAS) or (SSDs). You can install the following number of eXFlash units: The x3850 X5 can have either of the following configurations: Up to four SAS or Serial Advanced Technology Attachment (SATA) drives, plus the eight SSDs in one eXFlash unit Sixteen SSDs in two eXFlash units The x3950 X5 database-optimized models have one eXFlash unit standard with space for eight SSDs, and a second eXFlash is optional. The x3690 X5 can have up to 24 SSDs in three eXFlash units. Spinning disks, although an excellent choice for cost per megabyte, are not always the best choice when considered for their cost per I/O operation per second (IOPS). In a production environment where the tier-one capacity requirement can be met by IBM eXFlash, the total cost per IOPS can be lower than any solution requiring attachment to external storage. Host bus adapters (HBAs), switches, controller shelves, disk shelves, cabling, and the actual disks all carry a cost. They might even require an upgrade to the machine room infrastructure, for example, a new rack or racks, additional power lines, or perhaps additional cooling infrastructure. Also, remember that the storage acquisition cost is only a part of the total cost of ownership (TCO). TCO takes the ongoing cost of management, power, and cooling for the additional storage infrastructure detailed previously. SSDs use only a fraction of the power, generate only a fraction of the heat that spinning disks generate, and, because they fit in the chassis, are managed by the server administrator. IBM eXFlash is optimized for a heavy mix of read and write operations, such as transaction processing, media streaming, surveillance, file copy, logging, backup and recovery, and business intelligence. In addition to its superior performance, eXFlash offers superior uptime with three times the reliability of mechanical disk drives. SSDs have no moving parts to fail. They use Enterprise Wear-Leveling to extend their use even longer. All operating systems that are listed in ServerProven for each machine are supported for use with eXFlash. The eXFlash SSD backplane uses two long SAS cables, which are included with the backplane option. If two eXFlash backplanes are installed, four cables are required. You can connect the eXFlash backplane to the dedicated RAID slot if desired. In a system that has two eXFlash backplanes installed, two controllers are required in PCIe slots 1 - 4 to control the drives; however, up to four controllers can be used. In environments where RAID protection is required, use two RAID controllers per backplane to ensure that peak IOPS can be reached. Although use of a single RAID controller results in a functioning solution, peak IOPS can be reduced by a factor of approximately 50%. Remember that each RAID controller controls only its own disks. With four B5015 controllers, each controller controls four disks. The effect of RAID-5 is that four disks (one per array) are used for parity. You can use both RAID and non-RAID controllers. The IBM 6Gb SSD Host Bus Adapter (HBA) is optimized for read-intensive environments, and you can achieve maximum performance with only a single 6Gb SSD HBA. RAID controllers are a better choice for environments with a mix of read and write activity. The eXFlash units can connect to the same types of ServeRAID disk controllers as the SAS and SATA disks. For higher performance, connect them to the IBM 6Gb SAS HBA or the ServeRAID B5015 SSD Controller.
48
In addition to using less power than rotating magnetic media, the SSDs are more reliable, and they can service many more IOPS. These attributes make them well suited to I/O intensive applications, such as complex queries of databases. Figure 2-27 shows an eXFlash unit, with the status lights assembly on the left side.
For more information about system-specific memory eXFlash options, see the following sections: IBM System x3850 X5: 3.9.3, IBM eXFlash and 1.8-inch SSD support on page 93 IBM System x3690 X5: 4.9.2, IBM eXFlash and SSD disk support on page 149.
49
Under similar optimized benchmarking conditions, eight of the 50 GB, 1.8-inch SSDs are able to sustain 48,000 read IOPS and, in a separate benchmark, 16,000 write IOPS. The cost of USD12,000 for the SSDs works out at approximately USD0.25 per IOP and USD60 per gigabyte. Additional spinning disks create additional costs in terms of shelves, rack space, and power and cooling, none of which are applicable for the SSDs, driving their total TCO down even further. The initial cost per GB is higher for the SSDs, but view it in the context of TCO over time. For more information regarding each of the eX5 systems, see the following sections: 3.9, Storage on page 90 5.11, Storage on page 203
50
RHEV-H (KVM) supports 96 CPU threads (an 8-core processor with Hyper-Threading enabled has 16 threads) and can address 1 TB RAM. KVM includes the following features: Advanced memory management support Robust and scalable Linux virtual memory manager Support for large memory systems with greater than 1 TB RAM Support for nonuniform memory access (NUMA) Transparent memory page sharing Memory overcommit KVM also provides the following advanced features: Live migration Snapshots Memory Page sharing SELinux for high security and isolation Thin provisioning Storage overlays
51
be assigned to SSD storage. When you must use memory swap files, assign the swap file space to conventional SAS or SATA hard drives.
the original database lookup. This capability allows the database or web content server to off-load the processing time that is needed to maintain those client sessions. You can define the same physical servers as virtual servers with access to a collection of SSD drives. The number of virtual servers can be dynamically adjusted to fit the demands of the database or web content server.
53
54
Chapter 3.
55
56
The x3850 X5 has the following physical specifications: Width: 440 mm (17.3 in.) Depth: 712 mm (28.0 in.) Height: 173 mm (6.8 in.) or 4 rack units (4U) Minimum configuration: 35.4 kg (78 lb.) Maximum configuration: 49.9 kg (110 lb.) Figure 3-1 shows the x3850 X5.
Figure 3-1 Front view of the x3850 X5 showing eight 2.5-inch SAS drives
In Figure 3-1, two serial-attached SCSI (SAS) backplanes have been installed (at the right of the server). Each backplane supports four 2.5-inch SAS disks (eight disks in total). Notice the orange colored bar on each disk drive. This bar denotes that the disks are hot-swappable. The color coding used throughout the system is orange for hot-swap and blue for non-hot-swap. Changing a hot-swappable component requires no downtime; changing a non-hot-swappable component requires that the server is powered off before removing that component. Figure 3-2 on page 58 shows the major components inside the server and on the front panel of the server.
57
Two 1975W rear-access hot-swap, redundant power supplies Eight memory cards for 64 DIMMs total eight 1066MHz DDR3 DIMMs per card
Dual-port 10Gb Ethernet Adapter (PCIe slot 7) Six available PCIe 2.0 slots
Four Intel Xeon CPUs Two 60mm hot-swap fans Two 120mm hot-swap fans
Additional slot for internal RAID controller Eight SAS 2.5 drives or two eXFlash SSD units
Figure 3-3 shows the connectors at the back of the server. Six available PCIe slots 10 Gigabit Ethernet ports (standard on most models) Power supplies (redundant at 220V power)
Serial port
Video port
58
Figure 3-4 x3850 X5 with the attached MAX5 memory expansion unit
59
The MAX5 has the following specifications: IBM EXA5 chip set. Intel memory controller with eight memory ports (four DIMMs on each port). Intel QuickPath Interconnect (QPI) architecture technology to connect the MAX5 to the x3850 X5. Four QPI links operate at up to 6.4 gigatransfers per second (GT/s). Scalability: Connects to an x3850 X5 server using QPI cables. Memory DIMMs: Minimum: 2 DIMMs, 4 GB. Maximum: 32 DIMM connectors (up to 512 GB of memory using 16 GB DIMMs). Type of DIMMs: PC3-10600, 1067 MHz, ECC, and DDR3 registered SDRAM DIMMs. Supports 2 GB, 4 GB, 8 GB, and 16 GB DIMMs.
All DIMM sockets in the MAX5 are accessible regardless of the number of processors installed on the host system. Five hot-swap 40 mm fans. Power supply: Hot-swap power supplies with built-in fans for redundancy support. 675-watt (110 - 220 V ac auto-sensing). One power supply standard, two maximum (second power supply is for redundancy). Light path diagnostics LEDs: Board LED Configuration LED Fan LEDs Link LED (for QPI and EXA5 links) Locate LED Memory LEDs Power-on LED Power supply LEDs Width: 483 mm (19.0 in.) Depth: 724 mm (28.5 in.) Height: 44 mm (1.73 in.) (1U rack unit) Basic configuration: 12.8 kg (28.2 lb.) Maximum configuration: 15.4 kg (33.9 lb.)
Physical specifications:
With the addition of the MAX5 memory expansion unit, the x3850 X5 gains an additional 32 DIMM sockets for a total of 96 DIMM sockets. Using 16 GB DIMMs means that a total of 1.5 TB of RAM can be installed. All DIMM sockets in the MAX5 are accessible, regardless of the number of processors installed on the host system. Figure 3-5 on page 61 shows the ports at the rear of the MAX5 memory expansion unit. The QPI ports on the MAX5 are used to connect to a single x3850 X5. The EXA ports are reserved for future use.
60
EXA port 1
EXA port 2
EXA port 3
QPI port 3
QPI port 4
Power connectors
EXA port 1 link LED EXA port 2 link LED EXA port 3 link LED
Figure 3-6 shows the internals of the MAX5 including the IBM EXA chip, which acts as the interface to the QPI links from the x3850 X5. IBM EXA chip Intel Scalable Memory buffers 32 DIMM sockets
For an in-depth look at the MAX5 offering, see 3.5, MAX5 on page 68.
61
Table 3-1 Comparison of the x3850 M2 to the x3850 X5 Subsystem CPU card x3850 X5 No Voltage Regulator Modules (VRMs), 4 Voltage Regulator Down (VRDs) Top access to CPUs and CPU card Eight memory cards DDR3 PC3-10600 running at up to 1066 MHz (processor dependent) Eight DIMMs per memory card 64 DIMMs per chassis, maximum With the MAX5, 96 DIMMs per chassis Intel 7500 Boxboro chip set All slots PCIe 2.0 Seven slots total at 5 Gb, 5 GHz, 500 MBps per lane Slot 1 PCIe x16, Slot2 x4 (x8 mechanical), Slots 3 - 7 x8 All slots non-hot-swap Standard ServeRAID BR10i with RAID 0 and 1 (most models) Optional ServeRAID M5015 with RAID 0, 1, and 5 Upgrade to RAID-6 and encryption No external SAS port BCM 5709 dual-port Gigabit Ethernet, PCIe 2.0 x4 Dual-port Emulex 10Gb Ethernet adapter in PCIe slot 7 on all models except ARx Matrox G200 in IMM 16 MB VRAM Maxim VSC452 Integrated BMC (IMM) Remote presence feature is standard Eight 2.5-inch internal drive bays or 16 1.8-inch solid-state drive bays Support for SATA and SSD ICH10 chip set USB: Six external ports, two internal No SuperIO No PS/2 keyboard/mouse connectors No diskette drive controller Optional optical drive 2x 120 mm 2x 60 mm 2x 120 mm in power supplies 1975 W hot-swap, full redundancy high voltage, 875 W low voltage Rear access Two power supplies standard; two maximum (most models)a x3850 M2 No Voltage Regulator Down (VRDs), 4 Voltage Regulator Modules (VRMs) Top access to CPU/VRM and CPU card Four memory cards DDR2 PC2-5300 running at 533 MHz Eight DIMMs per memory card 32 DIMMs per chassis, maximum
Memory
PCIe subsystem
IBM CalIOC2 2.0 chip set All slots PCIe 1.1 Seven slots total at 2.5 GHz, 2.5 Gb, 250 MBps per lane Slot 1 x16, slot 2 x8 (x4), slots 3 - 7 x8 Slots 6 - 7 are hot-swap LSI Logic 1078 with RAID-1 Upgrade key for RAID-5 SAS 4x external port for EXP3000 attach
SAS controller
Ethernet controller
Video controller
ATI RN50 on Remote Supervisor Adapter (RSA2) 16 MB VRAM RSA2 standard Remote presence feature is optional Four 2.5-inch internal drive bays
ICH7 chip set USB: Five external ports, one internal No SuperIO No PS/2 keyboard/mouse connectors No diskette drive controller 4x 120 mm 2x 92 mm 2x 80 mm in power supplies 1440 W hot-swap, full redundancy high voltage, 720 W low voltage Rear access Two power supplies standard; two maximum
Fans
62
63
3.3 Models
This section lists the currently available models. The x3850 X5 and x3950 X5 (both models are machine type 7145) have a three-year warranty. For information about the recent models, consult tools, such as the Configurations and Options Guide (COG) or Standalone Solutions Configuration Tool (SSCT). These tools are available at the Configuration tools website: http://www.ibm.com/systems/x/hardware/configtools.html
Intel Xeon processors (two standard; maximum of four) E7520 4C 1.86 GHz, 18 MB L3, 95W c E7520 4C 1.86 GHz, 18 MB L3, 95W c E7530 6C 1.86 GHz, 12 MB L3, 105W c E7540 6C 2.0 GHz, 18 MB L3, 105W X7550 8C 2.0 GHz, 18 MB L3, 130W X7560 8C 2.26 GHz, 24 MB L3, 130W
Memory speed (MHz) 800 MHz 800 MHz 978 MHz 1066 MHz 1066 MHz 1066 MHz
a. The x character in the seventh position of the machine model denotes the region-specific character. For example, U indicates US, and G indicates EMEA. b. The Emulex 10Gb Ethernet Adapter is installed in PCIe slot 7. c. Any model using the E7520 or E7530 CPU cannot scale beyond single node 4-way.
Model 5Dx
Model 5Dx is designed for database applications and uses SSDs for the best I/O performance. Backplane connections for eight 1.8-inch SSDs are standard and there is space for an additional eight SSDs. You must order the SSDs separately. Because no SAS controllers are standard, you can select from the available cards that are described in 3.9, Storage on page 90.
Model 4Dx
Model 4Dx is designed for virtualization and is fully populated with 4 GB memory DIMMs, including in an attached MAX5 memory expansion unit, for a total of 384 GB of memory.
64
Backplane connections for four 2.5-inch SAS HDDs are standard; however, you must order the SAS HDDs separately. A ServeRAID BR10i SAS controller is standard in this model.
Table 3-3 Models of the x3950 X5: Workload-optimized models Power supplies (std/max) 2/2 2/2 10Gb Ethernet standardb Memory cards (std/max)
Modela
Memory speed
MAX5
Standard memory
Database workload-optimized models 7145-5Dx X7560 8C 2.27 GHz, 24 MB L3, 130W 1066 MHz Opt Server: 8x 4GB 4/8 No Yes None
Virtualization workload optimized models 7145-4Dx 4x X7550 8C 2.0 GHz, 18 MB L3, 130W 1066 MHz Std Server: 64x 4GB MAX5: 32x 4GB 8/8 Yes Yes None
a. The x character in the seventh position of the machine model denotes the region-specific character. For example, U indicates US, and G indicates EMEA. b. The Emulex 10Gb Ethernet Adapter is installed in PCIe slot 7.
Intel Xeon processors (four standard and max) 4x E7530 6C 1.86 GHz, 12 MB L3, 105W c 4x X7550 8C 2.0 GHz, 18 MB L3, 130W 4x X7560 8C 2.27 GHz, 24 MB L3, 130W
Standard memory (MAX5 is standard) Server: 8x 4 GB MAX5: 2x 4 GB Server: 8x 4 GB MAX5: 2x 4 GB Server: 8x 4 GB MAX5: 2x 4 GB
a. The x character in the seventh position of the machine model denotes the region-specific character. For example, U indicates US, and G indicates EMEA. b. The Emulex 10Gb Ethernet Adapter is installed in PCIe slot 7. c. Any model using the E7520 or E7530 CPU cannot scale beyond single node 4-way.
65
Memory card 5
Memory card 7
SMI links MB 1
QPI links
Intel I/O Hub Slot 1: x16 FL Slot 2: x4 FL* Slot 3: x8 FL Slot 4: x8 FL * x8 mechanical Intel Southbridge Slot 5: x8 HL Slot 6: x8 HL Slot 7: x8 HL x8 Intel I/O Hub x4 SAS Dual Gb Ethernet DVD, USB, IMM, LP
Memory card 6
MB 1 MB 2
MB 1 MB 2
Memory card 1
MB 1 MB 2 Intel Xeon CPU 1 QPI MB 2 QPI ports QPI Intel Xeon CPU 2
MB 1 MB 2
Memory card 2
MB 1
MB 1 MB 2
Memory card 4
Memory card 3
QPI links
Memory card 8
In Figure 3-7, the dotted lines indicate where the QPI Wrap Cards are installed in a 4-processor configuration. These wrap cards complete the full QPI mesh to allow all four processors to connect to each other. The QPI Wrap Cards are not needed in 2-processor configurations and are removed when a MAX5 is connected. Figure 3-12 on page 70 is a block diagram of the x3850 X5 connected to a MAX5.
66
Tip: The QPI Wrap Cards are only for single-node configurations with three or four processors installed. They are not necessary for any of the following items: Single-node configurations with two processors Configurations with MAX5 memory expansion units Two-node configurations Figure 3-8 shows the QPI Wrap Card.
For single-node systems with three or four processors installed, but without the MAX5 memory expansion unit connected, install two QPI Wrap Cards. Figure 3-9 shows a diagram of how the QPI Wrap Cards are used to complete the QPI mesh. Although the QPI Wrap Cards are not mandatory, they provide a performance boost by ensuring that all CPUs are only one hop away from each other, as shown in Figure 3-9.
QPI ports QPI Wrap Card connection SMI links Intel Xeon CPU 3 QPI QPI External QPI ports Intel Xeon CPU 4
QPI links
Intel Xeon CPU 2 External QPI ports QPI Wrap Card connection
QPI ports
67
The QPI Wrap Cards are not included with standard server models and must be ordered separately. See Table 3-5.
Table 3-5 Ordering information for the QPI Wrap Card Part number 49Y4379 Feature code Not applicable Description IBM x3850 X5 and x3950 X5 QPI Wrap Card Kit (quantity 2)
Tips: Part number 49Y4379 includes two QPI Wrap Cards. You order only one of these parts per server. QPI Wrap Cards cannot be ordered individually. The QPI Wrap Cards are installed in the QPI bays at the back of the server, as shown in Figure 3-10. QPI Wrap Cards are not needed in a 2-node configuration and not needed in a MAX5 configuration. When the QPI Wrap Cards are installed, no external QPI ports are available. If you later want to attach a MAX5 expansion unit or connect a second node, you must first remove the QPI Wrap Cards.
3.5 MAX5
As introduced in 3.1.3, IBM MAX5 memory expansion unit on page 59, the MAX5 memory expansion drawer is available for both the x3850 X5 and the x3950 X5. Models of the x3850 X5 and x3950 X5 are available that include the MAX5, as described in 3.3, Models on
68
page 64, Also, you can order the MAX5 separately, as listed in Table 3-6. When ordering a MAX5, remember to order the cable kit as well. For power supply fault redundancy, order the optional power supply.
Table 3-6 Ordering information for the IBM MAX5 for System x Part number 59Y6265 60Y0332 59Y6267 Feature code 4199 4782 4192 Description IBM MAX5 for System x IBM 675W HE Redundant Power Supply IBM MAX5 to x3850 X5 Cable Kit
The eX5 chip set in the MAX5 is an IBM unique design that attaches to the QPI links as a node controller, giving it direct access to all CPU bus transactions. It increases the number of DIMMs supported in a system by a total of 32, and it also adds another 16 channels of memory bandwidth, boosting overall throughput. Therefore, the MAX5 adds additional memory and performance. The eX5 chip connects directly through QPI links to all of the CPUs in the x3850 X5, and it maintains a directory of each CPUs last-level cache. Therefore, when a CPU requests content stored in the cache of another CPU, the MAX5 not only has that same data stored in its own cache, it is able to return the acknowledgement of the snoop and the data to the requesting CPU in the same transaction. For more information about QPI links and snooping, see 2.2.4, QuickPath Interconnect (QPI) on page 18. The MAX5 also has EXA scalability ports used in an EXA-scaled configuration (that is, a 2-node and MAX5 configuration). These ports are reserved for future use. In summary, the MAX5 offers the following major features: Adds 32 DIMM slots to either the x3850 X5 or the x3690 X5 Adds 16 channels of memory bandwidth Improves snoop latencies Figure 3-11 shows a diagram of the MAX5.
External connectors QPI QPI QPI QPI EXA EXA EXA
Memory buffer Memory buffer Memory buffer Memory buffer SMI links IBM EXA chip
SMI links
69
The MAX5 is connected to the x3850 X5 using four cables, connecting the QPI ports on the server to the four QPI ports on the MAX5. Figure 3-12 shows architecturally how a single-node x3850 X5 connects to a MAX5.
External QPI cables
System x3850 X5
MAX5
QPI
QPI links
EXA
EXA
EXA
SMI links
QPI
QPI
EXA
Figure 3-12 The x3850 X5: Connectivity of the system unit with the MAX5
Tip: As shown in Figure 3-12 on page 70, you maximize performance when you have four processors installed, because you then have four active QPI links to the MAX5. However, configurations of two and three processors are still supported. If only two processors are required, consider the use of the x3690 X5. We describe the connectivity of the MAX5 to the x3850 X5 in 3.6, Scalability on page 70. For memory configuration information, see 3.8.4, Memory mirroring on page 87. For information about power and fans, see 3.14, Power supplies and fans of the x3850 X5 and MAX5 on page 112.
3.6 Scalability
In this section, we describe how to expand the x3850 X5 to increase the number of processors and the number of memory DIMMs. The x3850 X5 currently supports the following scalable configurations: A single x3850 X5 server with four processor sockets. This configuration is sometimes referred to as a single-node server. A single x3850 X5 server with a single MAX5 memory expansion unit attached. This configuration is sometimes referred to as a memory-expanded server. 70
IBM eX5 Implementation Guide
QPI
QPI
QPI
Two x3850 X5 servers connected to form a single image 8-socket server. This configuration is sometimes referred to as a 2-node server. MAX5: The configuration of two nodes with MAX5 is not supported.
Rack rear
Connecting the MAX5 to a single-node x3850 X5 requires one IBM MAX5 to x3850 X5 Cable Kit, which consists of four QPI cables. See Table 3-7.
Table 3-7 Ordering information for the IBM MAX5 to x3850 X5 Cable Kit Part number 59Y6267 Feature code 4192 Description IBM MAX5 to x3850 X5 Cable Kit (quantity 4 cables)
71
Rack rear
Connecting the two x3850 X5 servers to form a 2-node system requires one IBM x3850 X5 and x3950 X5 QPI Scalability Kit, which consists of four QPI cables. See Table 3-8.
Table 3-8 Ordering information for the IBM x3850 X5 and x3950 X5 QPI Scalability Kit Part number 46M0072 Feature code 5103 Description IBM x3850 X5 and x3950 X5 QPI Scalability Kit (quantity 4 cables)
No QPI ports are visible on the rear of the server. The QPI scalability cables have long rigid connectors, allowing them to be inserted into the QPI bay until they connect to the QPI ports, which are located a few inches inside on the planar. Completing the QPI scaling of two x3850 X5 servers into a 2-node complex does not require any other option. Intel E7520 and E7530: The Intel E7520 and E7530 processors cannot be used to scale to an 8-way 2-node complex. They support a maximum of four processors. At the time of this writing, the following models use those processors: 7145-ARx 7145-1Rx 7145-2Rx 7145-2Sx Figure 3-15 on page 73 shows the QPI links that are used to connect two x3850 X5 servers to each other. Both nodes must have four processors each, and all processors must be identical.
72
1 4
3 2
3 2
1 4
QPI Links
QPI-based scaling is managed primarily through the Unified Extensible Firmware Interface (UEFI) firmware of the x3850 X5. For the 2-node x3850 X5 scaled through the QPI ports, when those cables are connected, the two nodes act as one system until the cables are physically disconnected. Firmware levels: It is important to ensure that both of the x3850 X5 servers have the identical UEFI, integrated management module (IMM), and Field-Programmable Gate Array (FPGA) levels before scaling. If they are not at the same levels, unexpected issues occur and the server might not boot. See 9.10, Firmware update tools and methods on page 509 for ways to check and update the firmware.
73
a. GT/s is gigatransfers per second. For an explanation, see 2.3.1, Memory speed on page 22. b. Intel Hyper-Threading Technology. For an explanation, see 2.2.2, Hyper-Threading Technology on page 17. c. Intel Turbo Boost Technology. For an explanation, see 2.2.3, Turbo Boost Technology on page 18. d. Scalable to a 4-socket maximum, and therefore, it cannot be used in a 2-node x3850 X5 complex that is scaled with native QPI cables.
With the exception of the E7520, all processors listed in Table 3-9 support Intel Turbo Boost Technology. When a processor operates below its thermal and electrical limits, Turbo Boost dynamically increases the clock frequency of the processor by 133 MHz on short and regular intervals until an upper limit is reached. See 2.2.3, Turbo Boost Technology on page 18 for more information. With the exception of the X7542, all of the processors that are shown in Table 3-9 support Intel Hyper-Threading Technology, which is an Intel technology that can improve the parallelization of workloads. When Hyper-Threading is engaged in the BIOS, for each processor core that is physically present, the operating system addresses two. For more information, see 2.2.2, Hyper-Threading Technology on page 17. All processor options include the heat-sink and CPU installation tool. This tool is extremely important due to the high possibility of bending pins on the processor socket when using the incorrect procedure. The x3850 X5 includes at least two CPUs as standard. Two CPUs are required to access all seven of the PCIe slots (shown in Figure 3-7 on page 66): Either CPU 1 or CPU 2 is required for the operation of PCIe slots 5-7. Either CPU 3 or CPU 4 is required for the operation of PCIe Slots 1-4. All CPUs are also required to access all memory cards on the x3850 X5 but they are not required to access memory on the MAX5, as explained in 3.8, Memory on page 76.
74
Use these population guidelines: Each CPU requires a minimum of two DIMMs to operate. All processors must be identical. Only configurations of two, three, or four processors are supported. The number of installed processors dictates what memory cards can be used: Two installed processors enable four memory cards. Three installed processors enable six memory cards. Four installed processors enable all eight memory cards. A processor must be installed in socket 1 or 2 for the system to successfully boot. A processor is required in socket 3 or 4 to use PCIe slots 1 - 4. See Figure 3-7 on page 66. When installing three or four processors, use a QPI Wrap Card Kit (part number 49Y4379) to improve performance. The kit contains two wrap cards. See 3.4.2, QPI Wrap Card on page 66. When using a MAX5 memory expansion unit, as shown in Figure 3-12 on page 70, you maximize performance when you have four installed processors because there are four active QPI links to the MAX5. However, configurations of two and three processors are still supported. Consider the X7542 processor for CPU frequency-dependent workloads because it has the highest core frequency of the available processor models. If high processing capacity is not required for your application but high memory bandwidth is required, consider using four processors with fewer cores or a lower core frequency rather than two processors with more cores or a higher core frequency. Having four processors enables all memory channels and maximizes memory bandwidth. We describe this situation in 3.8, Memory on page 76.
75
3.8 Memory
Memory is installed in the x3850 X5 in memory cards. Up to eight memory cards can be installed in the server, and each card holds eight DIMMs. Therefore, the x3850 X5 supports up to 64 DIMMs. This section includes the following topics: 3.8.1, Memory cards and DIMMs on page 76 3.8.2, DIMM population sequence on page 79 3.8.3, Maximizing memory performance on page 84 3.8.4, Memory mirroring on page 87 3.8.5, Memory sparing on page 89 3.8.6, Effect on performance by using mirroring or sparing on page 89
DIMM 1
DIMM 8
Figure 3-16 x3850 x5 memory card
76
Standard models contain two or more memory cards. You can configure additional cards, as listed in Table 3-10.
Table 3-10 IBM System x3850 X5 and x3950 X5 memory card Part number 46M0071 Feature code 5102 Description IBM x3850 X5 and x3950 X5 Memory Expansion Card
The memory cards are installed in the server, as shown in Figure 3-17. Each processor is electrically connected to two memory cards as shown (for example, processor 1 is connected to memory cards 1 and 2). Memory cards 1 - 8 Processors 1 - 4
77
2 GB (1x 2GB) 1Rx8, 2 Gb PC3-10600R DDR3-1333 4 GB (1x 4GB), 2Rx8, 2 Gb PC3-10600R DDR3-1333 4 GB (1x 4GB), 4Rx8, 1 Gb PC3-8500 DDR3-1066 8 GB (1x 8GB), 4Rx8, 2 Gb PC3-8500 DDR3-1066 16 GB (1x 16GB), 4Rx4, 2 Gb PC3-8500 DDR3-1066
a. Memory speed is also controlled by the memory bus speed as specified by the processor model selected. The actual memory bus speed is the lower of both the processor memory bus speed and the DIMM memory bus speed. b. Although 1333 MHz memory DIMMs are supported in the x3850 X5, the memory DIMMs run at a maximum speed of 1066 MHz. c. The 16 GB memory option is supported in the MAX5 only when it is the only type of memory that is used in the MAX5. No other memory options can be used in the MAX5 if this option is installed in the MAX5. This DIMM also supports redundant bit steering (RBS) when used in the MAX5, as described in Redundant bit steering on page 29.
Guidelines: Memory options must be installed in matched pairs. Single options cannot be installed, so the options that are shown in Table 3-11 need to be ordered in quantities of two. You can achieve additional performance by enabling Hemisphere Mode, which is described in Hemisphere Mode on page 26. This mode requires that the memory options are installed in matched quads. The maximum memory speed that is supported by Xeon 7500 and 6500 (Nehalem-EX) processors is 1066 MHz (1,333 MHz speed is not supported). Although the 1333 MHz DIMMs are still supported in the x3850 X5, they can operate at a speed of at most 1066 MHz. As with Intel Xeon 5500 processor (Nehalem-EP), the speed at which the memory that is connected to the Xeon 7500 and 6500 processors (Nehalem-EX) runs depends on the capabilities of the specific processor. With Nehalem-EX, the scalable memory interconnect (SMI) link runs from the memory controller that is integrated in the processor to the memory buffers on the memory cards. The SMI link speed is derived from the processor QPI link speed: 6.4 GT/s QPI link speed capable of running memory speeds up to 1066 MHz 5.86 GT/s QPI link speed capable of running memory speeds up to 978 MHz 4.8 GT/s QPI link speed capable of running memory speeds up to 800 MHz To see more information about how memory speed is calculated with QPI, see 2.3.1, Memory speed on page 22.
78
MAX5 memory
The MAX5 memory expansion unit has 32 DIMM sockets and is designed to augment the installed memory in the attached x3850 X5 server. Table 3-12 shows the available memory options that are supported in the MAX5 memory expansion unit. These options are a subset of the options that are supported in the x3850 X5 because the MAX5 requires that all DIMMs use identical DRAM technology: either 2 Gb x8 or 2 Gb x4 (but not both at the same time). x3850 X5 memory options: The memory options listed here are also supported in the x3850 X5, but under other feature codes for configure-to-order (CTO) clients. Additional memory options are also supported in the x3850 X5 server but not in the MAX5; these options are listed in Table 3-11 on page 78.
Table 3-12 DIMMs supported in the MAX5 Part number 44T1592 44T1599 46C7482 46C7483 MAX5 feature code 2429 2431 2432 2433 Memory Supporte d in MAX5 Yes Yes Yes Yesc d Memory speeda 1333 MHzb 1333 MHz 1066 MHz 1066 MHz Ranks
2 GB (1x 2GB) 1Rx8, 2Gbit PC3-10600R DDR3-1333 4 GB (1x 4GB), 2Rx8, 2Gbit PC3-10600R DDR3-1333 8 GB (1x 8GB), 4Rx8, 2Gbit PC3-8500 DDR3-1066 16 GB (1x 16GB), 4Rx4, 2Gbit PC3-8500 DDR3-1066
a. Memory speed is also controlled by the memory bus speed, as specified by the selected processor model. The actual memory bus speed is the lower of both the processor memory bus speed and the DIMM memory bus speed. b. Although 1333 MHz memory DIMMs are supported in the x3690 X5, the memory DIMMs run at a maximum speed of 1066 MHz. c. The 16 GB memory option is supported in the MAX5 only when it is the only type of memory used in the MAX5. No other memory options can be used in the MAX5 if this option is installed in the MAX5. d. This DIMM supports redundant bit steering (RBS), as described in Redundant bit steering on page 29.
Use of the 16 GB memory option: The 16 GB memory option, 46C7483, is supported in the MAX5 only when it is the only type of memory that is used in the MAX5. No other memory options can be used in the MAX5 if this option is installed in the MAX5.
Redundant bit steering: Redundant bit steering (RBS) is not supported on the x3850 X5 itself, because the integrated memory controller of the Intel Xeon 7500 processors does not support the feature. See Redundant bit steering on page 29 for details. The MAX5 memory expansion unit support RBS, but only with x4 memory and not x8 memory. As shown in Table 3-12, the 16 GB DIMM, part 46C7483, uses x4 DRAM technology. RBS is automatically enabled in the MAX5 memory port, if all DIMMs installed to that memory port are x4 DIMMs.
79
Installing DIMMs in the x3850 X5 and MAX5 in the correct order is essential for system performance. See Mixed DIMMs and the effect on performance on page 86 for performance effects when this guideline is not followed. Tip: The tables in this section list only memory configurations that are considered the best practices in obtaining the optimal memory and processor performance. For a full list of supported memory configurations, see the IBM System x3850 X5 Installation and Users Guide or the IBM System x3850 X5 Problem Determination and Service Guide. We list the download links to these documents in Related publications on page 541.
4 8 12 16 20 24 28 32
N Y N Y N Y N Y
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26.
Table 3-14 on page 81 shows the NUMA-compliant memory installation sequence for three processors.
80
Table 3-14 NUMA-compliant DIMM installation (three processors): x3850 X5 Processor 1 Hemisphere Modea Number of DIMMs Card 1 DIMM 1 and 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 2 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Processor 4 Card 7 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Processor 2 or 3 Card 3 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 4 DIMM 3 and 6 x x x x x x x x x DIMM 2 and 7 DIMM 4 and 5 DIMM 4 and 5 Card 6 DIMM 1 and 8 DIMM 3 and 6 x x x x x x DIMM 2 and 7 DIMM 4 and 5 x x x x x x x x x x x x x x x
6 12 18 24 30 36 42 48
N Y N Y N Y N Y
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26.
Three-processor system: For a 3-processor system, you can use either processor slot 2 or processor 3. Processor 3 uses cards 5 and 6 instead of cards 3 and 4, which are used for processor 2. Table 3-15 shows the NUMA-compliant memory installation sequence for four processors.
Table 3-15 NUMA-compliant DIMM installation (four processors): x3850 X5 Processor 1 Hemisphere Modea Number of DIMMs Card 1 DIMM 1 and 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 2 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Processor 4 Card 7 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Processor 2 Card 3 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 4 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 x x x x x x x x x x x x x x x x x x x x x x x x x x x x Processor 3 Card 5 DIMM 3 and 6 DIMM 2 and 7
8 16 24 32 40 48 56
N Y N Y N Y N
x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x
81
Processor 1 Hemisphere Modea Number of DIMMs Card 1 DIMM 1 and 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 2 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8
Processor 4 Card 7 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8
Processor 2 Card 3 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 4 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8
Processor 3 Card 5 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 6 DIMM 3 and 6 x DIMM 2 and 7 x DIMM 4 and 5 x
64
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26.
MAX5 configurations
The memory installed in the MAX5 operates at the same speed as the memory that is installed in the x3850 X5 server. As explained in 2.3.1, Memory speed on page 22, the memory speed is derived from the QPI link speed of the installed processors, which in turn dictates the maximum SMI link speed, which in turn dictates the memory speed. Table 3-9 on page 74 summarizes the memory speeds of all the models of Intel Xeon 7500 series CPUs. One important consideration when installing memory in MAX5 configurations is that the server must be fully populated before adding DIMMs to the MAX5. As we described in 2.3.2, Memory DIMM placement on page 23, you get the best performance by using all memory buffers and all DIMM sockets on the server first, and then add DIMMs to the MAX5. Figure 3-18 on page 83 shows the numbering scheme for the DIMM slots on the MAX5, and the pairing of DIMMs in the MAX5. As DIMMs are added in pairs, they must be matched on a memory port (as shown by using the colors). For example, DIMM1 is matched to DIMM 8, DIMM 2 to DIMM 7, DIMM 20 to DIMM 21, and so on.
82
16 15 14 13
12 11 10 9
DIMM 16
DIMM 14
DIMM 12
DIMM 11
DIMM 15
DIMM 13
DIMM 10
DIMM 7
DIMM 9
DIMM 8
DIMM 6
DIMM 5
DIMM 4
DIMM 3
DIMM 2
Memory buffer 4
Memory buffer 3
Memory buffer 5
Memory buffer 6
Quad D Quad C 32 31 30 29 28 27 26 25
Quad B Quad A 24 23 22 21 20 19 18 17
DIMM 26
DIMM 31
DIMM 29
DIMM 28
DIMM 27
DIMM 25
DIMM 24
DIMM 23
DIMM 22
DIMM 21
DIMM 20
DIMM 32
DIMM 30
DIMM 19
DIMM 18
Memory buffer 1
Memory buffer 2
Memory buffer 8
DIMM 17
Memory buffer 7
Quad H Quad G
Quad F Quad E
Table 3-16 shows the population order of the MAX5 DIMM slots, ensuring that memory is balanced among the memory buffers. The colors in the table match the colors in Figure 3-18.
Table 3-16 DIMM pair 1 2 3 4 5 6 7 8 DIMM installation sequence in the MAX5 DIMM slots 28 and 29 9 and 16 1 and 8 20 and 21 26 and 31 11 and 14 3 and 6 18 and 23
DIMM 1
83
DIMM pair 9 10 11 12 13 14 15 16
DIMM slots 27 and 30 10 and 15 2 and 7 19 and 22 25 and 32 12 and 13 4 and 5 17 and 24
84
Figure 3-19 shows eight possible memory configurations for the two memory cards and 16 DIMMs connected to one processor socket. Each configuration has a relative performance score. Note the key information from this chart: The best performance is achieved by populating all memory DIMMs in two memory cards for each processor installed (configuration 1). Populating only one memory card per socket can result in approximately a 50% performance degradation (compare configuration 1 with 5). Memory performance is better if you install DIMMs on all memory channels than if you leave any memory channels empty (compare configuration 2 with 3). Two DIMMs per channel result in better performance that one DIMM per channel (compare configuration 1 with 2, and compare configuration 5 with 6).
Relative performance
Each processor: 2 memory controllers 2 DIMMs per channel 8 DIMMs per MC Each processor: 2 memory controllers 1 DIMM per channel 4 DIMMs per MC Each processor: 2 memory controllers 2 DIMMs per channel 4 DIMMs per MC Each processor: 2 memory controllers 1 DIMM per channel 2 DIMMs per MC Each processor: 1 memory controller 2 DIMMs per channel 8 DIMMs per MC Each processor: 1 memory controller 1 DIMM per channel 4 DIMMs per MC Each processor: 1 memory controller 2 DIMMs per channel 4 DIMMs per MC Each processor: 1 memory controller 1 DIMM per channel 2 DIMMs per MC
Memory configurations
1 2 3 4 5 6 7 8
1.0
Memory card DIMMs Channel Memory buffer SMI link Memory controller
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
0.94
Mem Ctrl 1
0.61
1
1 0.94
Mem Ctrl 1
Mem Ctrl 2
0.58
0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 1 2 3 4 5 6 7 8 Configuration
0.61 0.58 0.51 0.47 0.31
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
0.29
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
Mem Ctrl 1
Mem Ctrl 2
0.29
Figure 3-19 Relative memory performance based on DIMM placement (one processor and two memory cards shown)
Use the following general memory population rules: DIMMs must be installed in matching pairs. Each memory card requires at least two DIMMs. Each processor and memory card must have identical amounts of RAM. Install and populate two memory cards per processor or you can lose memory bandwidth.
85
Populate one DIMM per channel on every channel on memory before populating a second DIMM in any channel. Populate DIMMs at the end of a memory channel first before populating the DIMM closer to the memory buffer. That is, install to sockets 1, 3, 6, and 8 first. If you have a mix of DIMM capacities (such as 4 GB and 8 GB DIMMs), insert the largest DIMMs first (spreading out the DIMMs across every memory channel), then move to next largest DIMMs, and finish with the smallest capacity DIMMs that you have. Therefore, where memory performance is key to a successful deployment, the best configuration is to install 32 or 64 identical DIMMs across eight memory cards and four processors. A system with fewer than four installed processors or fewer than eight installed memory cards has fewer memory channels, therefore less bandwidth and lower performance.
86
Relative performance: 97
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
Relative performance: 92
Relative performance: 82
Buffer
Buffer
DIMM
DIMM DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
As you can see, mixing DIMM sizes can cause performance loss up to 18%, even if all channels are occupied and Hemisphere Mode is maintained.
87
Table 3-17 x3850 X5 memory mirroring 4-processor 2-node Processor 1 Number of DIMMs Card 1 DIMM 1 and 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 2 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Processor 4 Card 7 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 8 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Processor 2 Card 3 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 4 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Processor 3 Card 5 DIMM 3 and 6 DIMM 2 and 7 DIMM 4 and 5 DIMM 1 and 8 Card 6 DIMM 3 and 6 x x x x x x x x x x x DIMM 2 and 7 x DIMM 4 and 5 x
4 8 12 16 20 24 28 32 36 40 44 48 52 56 60 64
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
88
DIMM 10
DIMM 11
DIMM 12
DIMM 13
DIMM 14
DIMM 15
DIMM 16
DIMM 17
DIMM 18
DIMM 19
DIMM 20
DIMM 21
DIMM 22
DIMM 23
DIMM 24
DIMM 25
DIMM 26
DIMM 27
DIMM 28
DIMM 29
DIMM 30 x x x x
DIMM 31 x x x x x x
4 8 12 16 20 24 28 32 x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x
x x x x x x x x
89
DIMM 32 x x
DIMM 1
DIMM 2
DIMM 3
DIMM 4
DIMM 5
DIMM 6
DIMM 7
DIMM 8
DIMM 9
Sparing
62
Mirroring
50
Normal
100
20
40
60
80
100
120
3.9 Storage
In this section, we look at the internal storage and RAID options for the x3850 X5, with suggestions where you can obtain the details about supported external storage arrays. This section includes the following topics: 3.9.1, Internal disks on page 90 3.9.2, SAS and SSD 2.5-inch disk support on page 91 3.9.3, IBM eXFlash and 1.8-inch SSD support on page 93 3.9.4, SAS and SSD controllers on page 96 3.9.6, External storage connectivity on page 101
90
Figure 3-22 Front view of the x3850 X5 showing eight 2.5-inch SAS drives
The SAS backplane uses a short SAS cable (included with the part number 59Y6135), and it is always controlled by the RAID adapter in the dedicated slot behind the disk cage, never from an adapter in the PCIe slots. The required power/signal Y cable is also included with the x3850 X5. Up to two 2.5-inch backplanes (each holding up to four disks) can connect to a RAID controller installed in the dedicated RAID slot. Table 3-21 lists the supported RAID controllers. For more information about each RAID controller, see 3.9.4, SAS and SSD controllers on page 96.
Table 3-21 RAID controllers that are compatible with SAS backplane and SAS disk drives Part number 44E8689 46M0831 46M0829 Feature code 3577 0095 0093 Description ServeRAID BR10i (standard on most models; see 3.3, Models on page 64) ServeRAID M1015 SAS/SATA Controller ServeRAID M5015 SAS/SATA Controllera
91
Description ServeRAID M5014 SAS/SATA Controller ServeRAID B5015 SSD IBM ServeRAID M5000 Advance Feature Key: Adds RAID-6, RAID-60, and SED Data Encryption Key Management to the ServeRAID M5014, M5015, and M5025 controllers IBM ServeRAID M5000 Performance Accelerator Key: Adds Cut Through I/O (CTIO) for SSD FastPath optimization on ServeRAID M5014, M5015, and M5025 controllers
81Y4426
A10C
Table 3-22 lists the 2.5-inch SAS 10K and 15K RPM disk drives and the 2.5-inch SSD that are supported in the x3850 X5. These drives are supported with the SAS hard disk backplane, 59Y6135.
Table 3-22 Supported 2.5-inch SAS drives and 2.5-inch SSDs Part number 42D0632 42D0637 42D0672 42D0677 43W7714 Feature code 5537 5599 5522 5536 3745 Description IBM 146 GB 10K 6 Gbps SAS 2.5-inch SFF Slim-HS HDD IBM 300 GB 10K 6 Gbps SAS 2.5-inch SFF Slim-HS HDD IBM 73 GB 15K 6 Gbps SAS 2.5-inch SFF Slim-HS HDD IBM 146 GB 15K 6 Gbps SAS 2.5-inch SFF Slim-HS HDD IBM 50 GB SATA 2.5-inch SFF Slim-HS High IOPS SSD
Table 3-23 lists the 2.5-inch Nearline SATA 7.2K drives that are supported in the x3850 X5. These drives are supported with the SAS hard disk backplane, part number 59Y6135.
Table 3-23 Supported 2.5-inch Nearline SATA drives Part number 42D0747 42D0752 Feature code 5405 5407 Description IBM 160GB 7200 NL SATA 2.5'' SFF Slim-HS HDD IBM 500GB 7200 NL SATA 2.5'' SFF Slim-HS HDD
The 2.5-inch drives require less space than 3.5-inch drives, consume half the power, produce less noise, can seek faster, and offer increased reliability. Compatibility: As listed in Table 3-22, the 2.5-inch 50 GB SSD is also supported with the standard SAS backplane and the optional SAS backplane, part number 59Y6135. It is incompatible with the 1.8-inch SSD eXFlash backplane, part number 59Y6213. A typical configuration can be two 2.5-inch SAS disks for the operating system and two High IOPS disks for data. Only the 2.5-inch High IOPS SSD disk can be used on the SAS backplane. The 1.8-inch disks for the eXFlash cannot be used on the SAS backplane.
92
The IBM eXFlash 8x 1.8-inch HS SAS SSD Backplane, part number 59Y6213, supports eight 1.8-inch SSDs. The eight drive bays require the same physical space as four SAS hard disk bays. A single eXFlash backplane requires two SAS x4 input cables and a power/configuration cable, which are both shipped standard. Up to two SSD backplanes and 16 SSDs are supported in the x3850 X5 chassis. For more information regarding eXFlash and SSD information, including a brief overview of the benefits of using eXFlash, see 2.8, IBM eXFlash on page 47. Two-node configurations: Spanning an array on any disk type between two chassis is not possible because the RAID controllers operate separately. Figure 3-23 shows an x3850 X5 with one of two eXFlash units installed.
93
Table 3-25 Controllers supported with the eXFlash SSD backplane option Part number 46M0914 46M0831 46M0829 46M0916 46M0969 46M0930 Feature code 3876 0095 0093 3877 3889 5106 Description IBM 6Gb SSD Host Bus Adapter (No RAID support) ServeRAID M1015 SAS/SATA Controller ServeRAID M5015 SAS/SATA Controllera ServeRAID M5014 SAS/SATA Controllera ServeRAID B5015 SSDa IBM ServeRAID M5000 Advance Feature Key: Adds RAID-6, RAID-60, and self-encrypting drives (SED) Data Encryption Key Management to the ServeRAID M5014, M5015, and M5025 controllers IBM ServeRAID M5000 Performance Accelerator Key: Adds Cut Through I/O (CTIO) for SSD FastPath optimization on ServeRAID M5014, M5015, and M5025 controllers
81Y4426
A10C
a. When using SSD drives, you must disable the write back cache to prevent latency and bottlenecks by using the controller settings or by adding the ServeRAID M5000 Series Performance Accelerator Key. See ServeRAID M5000 Series Performance Accelerator Key on page 95 for more information.
When ordering M5000 series controllers (M5014, M5015, or M5025) for use only with SSD drives, the cache battery must not be used for performance reasons. If using M5000 series controllers in a mixed SSD and SAS environment cache, order the battery along with the Performance Accelerator Key. If the ServeRAID controller being used is already set up and you want to leave the battery attached, you can still disable the write back cache by using the MegaRAID web BIOS, as shown in Figure 3-24.
94
95
The failure rate of SSDs is low because, in part, the drives have no moving parts. The 50 GB High IOPS SSD is a Single Level Cell (SLC) device with Enterprise Wear Leveling. As a consequence of both of these technologies, the additional layer of protection that is provided by a RAID controller might not always be necessary in every client environment and, in certain cases, RAID-0 might even be an acceptable option.
Name ServeRAID BR10ib ServeRAID M1015 ServeRAID M5014 ServeRAID M5015 ServeRAID M5025 IBM 6Gb SSD HBA ServeRAID B5015 SSD
a. See 3.9.5, Dedicated controller slot on page 100. b. The BR10i is standard on most models. See 3.3, Models on page 64. c. M1015 support for RAID-5 and RAID-50 requires the M1000 Advanced Feature Key (46M0832, 9749). d. M5014, M5015, and M5025 support for RAID-6 and RAID-60 requires the M5000 Advanced Feature Key (46M0930, fc 5106). e. ServeRAID M5015 option part number 46M0829 includes the M5000 battery; however, the feature code 0093 does not contain the battery. Order feature code 5744 if you want to include the battery in the server configuration. f. The IBM 6Gb SSD Host Bus Adapter is currently not available as a separately orderable option. Use the feature code to add the adapter to a customized order, using the configure-to-order (CTO) process. Part number 46M0914 is the L1 manufacturing part number. Part number 46M0983 is the pseudo option number, which is also used in manufacturing.
RAID levels 0 and 1 are standard on all models (except 7145-ARx) with the integrated BR10i ServeRAID controller. Model 7145-ARx has no RAID capability standard. All servers, even those servers that are not standard with the BR10i (model 7145-ARx), include the blue mounting bracket (see Figure 3-25 on page 101), which allows for the easy installation of a supported RAID controller in the dedicated x8 PCIe slot behind the disk cage. Only RAID controllers that are supported with the 2.5-inch SAS backplane can be used in this slot. See Table 3-27 for a summary of these supported options.
97
Performance Key Accelerator: Performance Accelerator Key uses the same features as the Advanced Feature Key, but it also includes performance enhancements to enable SSD support in a mixed HDD environment. For more information, see ServeRAID M5015 and M5014 SAS/SATA Controllers for IBM System x, TIPS0738, at the following website: http://www.redbooks.ibm.com/abstracts/tips0738.html?Open
98
Performance: Performance Accelerator Key uses the same features as the Advanced Feature Key. However, it also includes performance enhancements to enable SSD support in a mixed HDD environment. For more information, see ServeRAID M5025 SAS/SATA Controller for IBM System x, TIPS0739, at the following website: http://www.redbooks.ibm.com/abstracts/tips0739.html?Open
99
For more information, see IBM 6Gb SSD Host Bus Adapter for IBM System x, TIPS0744, which is available at the following website: http://www.redbooks.ibm.com/abstracts/tips0744.html?Open Important: Two variants of the 6Gb Host Bus Adapter exist. The SSD variant has no external port and is part number 46M0914. Do not confuse it with the IBM 6Gb SAS HBA, part number 46M0907, which is not supported for use with eXFlash.
100
Front of server The M5015 installs behind the disk cage and is shown here
Latch
This card installs in a special PCIe slot, and it is not installed in one of the seven PCIe slots for other expansion cards.
Edge clips
Optional battery
RAID-6, RAID-60, and encryption are a further optional upgrade for the M5015 through the ServeRAID M5000 Series Advance Feature Key or the Performance Accelerator Key.
The M5025 has two external SAS 2.0 x4 connectors and supports the following features: Eight external 6 Gbps SAS 2.0 ports implemented through two four-lane (x4) connectors. Two mini-SAS external connectors (SFF-8088). Six Gbps throughput per SAS port. 800 MHz PowerPC processor with LSI SAS2108 6 Gbps RAID on Chip (ROC) controller. PCI Express 2.0 x8 host interface. 512 MB onboard data cache (DDR2 running at 800 MHz).
101
Intelligent lithium polymer battery backup unit is standard with up to 48 hours of data retention. Support for RAID levels 0, 1, 5, 10, and 50 (RAID-6 and 60 support with either the optional M5000 Advanced Feature Key or the optional M5000 Performance Key). Connections: Up to 240 SAS or SATA drives Up to 9 daisy-chained enclosures per port SAS and SATA drives are supported, but mixing SAS and SATA in the same RAID array is not supported. Support for up to 64 logical volumes. Support for LUN sizes up to 64 TB. Configurable stripe size up to 1024 KB. Compliant with Disk Data Format (DDF) configuration on disk (COD). S.M.A.R.T. support. Support for the optional M5000 Advanced Feature Key, which enables the following features: RAID levels 6 and 60 LSI SafeStore: Support for SED services, such as instant secure erase and local key management (which requires the use of self-encrypting drives). Support for SSD drives in a mixed environment with SAS and SSD with the optional M5000 Series Performance Accelerator Key, which enables the following features: RAID levels 6 and 60 Performance optimizations for SSDs LSI SafeStore: Support for SED services, such as instant secure erase and local key management (which requires the use of self-encrypting drives). For more information, see the IBM Redbooks at-a-glance guide ServeRAID M5025 SAS/SATA Controller for IBM System x, TIPS0739, which is available at this website: http://www.redbooks.ibm.com/abstracts/tips0739.html?Open The x3850 X5 is qualified with a wide range of external storage options. To view the available solutions, see the Configuration and Options Guide, which is available at this website: http://ibm.com/systems/xbc/cog/x3850x5_7145/x3850x5_7145io.html The System Storage Interoperation Center (SSIC) is a search engine that provides details about supported configurations: http://www.ibm.com/systems/support/storage/config/ssic
102
All slots are PCI Express 2.0, full height, and not hot-swap. PCI Express 2.0 has several improvements over PCI Express 1.1 (as implemented in the x3850 M2). The chief benefit is the enhanced throughput. PCI Express 2.0 is rated for 500 MBps per lane (5 Gbps per lane); PCI Express 1.1 is rated for 250 MBps per lane (2.5 Gbps per lane). Note the following information about the slots: Slot 1 can accommodate a double-wide x16 card, but access to slot 2 is then blocked. Slot 2 is described as x4 (x8 mechanical). This host interface is sometimes shown as x4 (x8) and means that the slot is only capable of x4 speed, but is physically large enough to accommodate an x8 card. Any x8-rated card physically fits in the slot, but it runs at only x4 speed. Do not add RAID cards to this slot, because RAID cards in this slot cause bottlenecks and possible crashes. Slot 7 has been extended in length to 106 pins, making it a nonstandard connector. It still accepts PCI Express x8, x4, and x1 standard adapters. It is the only slot that is compatible with the extended edge connector on the Emulex 10Gb Ethernet Adapter, which is standard with most models.
103
Slots 5 - 7, the onboard Broadcom-based Ethernet dual-port chip and the custom slot for the RAID controller are on the first PCIe bridge and require that either CPU 1 or 2 is installed and operational. Slots 1 - 4 are on the second PCIe bridge and require that either CPU 3 or 4 is installed and operational. Table 3-31 shows the order in which to add cards to balance bandwidth between the two PCIe controllers. However, this installation order assumes that the cards are installed in matched pairs, or that they have similar throughput capabilities.
Table 3-31 Order for adding cards Installation order 1 2 3 4 5 6 7 PCIe slot 1 5 3 6 4 7 2 Slot width x16 PCIe slot x8 PCIe slot x8 PCIe slot x8 PCIe slot x8 PCIe slot x8 PCIe slot x4 PCIe slot Slot bandwidtha 8 GBps (80 Gbps) 4 GBps (40 Gbps) 4 GBps (40 Gbps) 4 GBps (40 Gbps) 4 GBps (40 Gbps) 4 GBps (40 Gbps) 2 GBps (20 Gbps)
a. This column correctly shows bandwidth expressed as GB for gigabyte or Gb for gigabit. 10 bits of traffic correspond to 1 byte of data due to the 8:10 encoding scheme. A single PCIe 2.0 lane provides a unidirectional bandwidth of 500 MBps or 5 Gbps.
Two additional power connectors, one 2x4 and one 2x3, are provided on the planar for high-power adapters, such as graphics cards. If there is a requirement to use an x16 PCIe card that is not shown as supported in ServerProven, initiate the SPORE process. To determine whether a vendor has qualified any x16 cards with the x3850 X5, see IBM ServerProven at the following website: http://www.ibm.com/servers/eserver/serverproven/compat/us/serverproven If the preferred vendors logo is displayed, click it to assess options that the vendor has qualified on the x3850 X5. You can obtain the support caveats for third-party options in 3.12.2, Optional adapters on page 107. In a 2-node configuration, all PCIe slots are available to the operating system that is running on the complex. They appear as devices on separate PCIe buses.
104
Tip: The Emulex 10Gb Ethernet Adapter that is standard with most models is a custom version of the Emulex 10Gb Virtual Fabric Adapter for IBM System x, part number 49Y4250. However, the features and functions of the two adapters are identical.
Figure 3-26 Top view of slot 6 and 7 showing that slot 7 is slightly longer than slot 6
The Emulex 10Gb Ethernet Adapter in the x3850 X5 has been customized with a special type of connector called an extended edge connector. The card itself is colored blue instead of green to indicate that it is nonstandard and cannot be installed in a standard x8 PCIe slot. At the time of writing, only the x3850 X5 and the x3690 X5 have slots that are compatible with the custom-built Emulex 10Gb Ethernet Adapter that is shown in Figure 3-27.
Figure 3-27 The Emulex 10Gb Ethernet Adapter has a blue circuit board and a longer connector
The Emulex 10Gb Ethernet Adapter is a customer-replaceable unit (CRU). To replace the adapter (for example, under warranty), order the CRU number, as shown in Table 3-32 on page 106. The table also shows the regular Emulex 10Gb Virtual Fabric Adapter (VFA) for IBM System x option, which differs only in the connector type (standard x8) and color of the circuit board (green).
105
Emulex VFA: The standard version of the Emulex VFA and the eX5 extended edge custom version can be used together as a redundant pair. This pair is a supported combination.
Table 3-32 Emulex adapter part numbers Option description Emulex 10Gb Ethernet Adapter for x3850 X5 Emulex 10Gb Virtual Fabric Adapter for IBM System x Part number None 49Y4250 Feature code 1648 5749 CRU number 49Y4202 Not applicable
General details about this card are in Emulex 10Gb Virtual Fabric Adapter for IBM System x, TIPS0762, which is available at the following website: http://www.redbooks.ibm.com/abstracts/tips0762.html Important: Although these cards are functionally identical, the availability of iSCSI and Fibre Channel over Ethernet (FCoE) upgrades for one card does not automatically mean availability for both cards. At the time of writing, the target availability of these features is the second quarter of 2011. Check the availability of iSCSI and FCoE feature upgrades with your local IBM representative. The Emulex 10Gb Ethernet Adapter for x3850 X5 offers the following features: Dual-channel, 10 Gbps Ethernet controller Near line rate 10 Gbps performance Two SFP+ empty cages to support either of the following items: SFP+ SR link is with SFP+ SR Module with LC connectors SFP+ twinaxial copper link is with SFP+ direct-attached copper module/cable Note: Servers that include the Emulex 10Gb Ethernet Adapter do not include transceivers. You must order transceivers separately if needed, as listed in Table 3-33. TCP/IP stateless off-loads TCP chimney off-load Based on Emulex OneConnect technology FCoE support as a future feature entitlement upgrade Hardware parity, cyclic redundancy check (CRC), error checking and correcting (ECC), and other advanced error checking PCI Express 2.0 x8 host interface Low-profile form-factor design IPv4/IPv6 TCP, user datagram protocol (UDP) checksum off-load Virtual LAN (VLAN) insertion and extraction Support for jumbo frames up to 9000 bytes Preboot eXecutive Environment (PXE) 2.0 network boot support Interrupt coalescing Load balancing and failover support
106
Deployment of this adapter and other Emulex OneConnect-based adapters with OneCommand Manager Interoperable with BNT 10Gb Top of Rack (ToR) switch for FCoE functions Interoperable with Cisco Nexus 5000 and Brocade 10Gb Ethernet switches for NIC/FCoE SFP+ transceivers are not included with the server. You must order them separately. Table 3-33 lists the compatible transceivers.
Table 3-33 Transceiver ordering information Option number 49Y4218 49Y4216 46C3447 Feature code 0064 0069 5053 Description QLogic 10Gb SFP+ SR Optical Transceiver Brocade 10Gb SFP+ SR Optical Transceiver BNT SFP+ Transceiver
107
Option 42D0511 46M6051 46M6052 59Y1988 42C2182 43W7491 43W7492 Graphics 49Y6804
Description QLogic 8Gb FC Dual-port HBA Brocade 8Gb FC Single-port HBA Brocade 8Gb FC Dual-port HBA Brocade 4Gb FC Single-port HBA QLogic 4Gb FC Dual-Port PCIe HBA Emulex 4GB FC Single-Port PCI-E HBA Emulex 4GB FC Dual-Port PCI-E HBA
1826
This list is constantly updated and changed. To see the latest updates, see the following website: http://ibm.com/systems/xbc/cog/x3850x5_7145/x3850x5_7145io.html Tools, such as the Configurations and Options Guide (COG) or SSCT, contain information about supported part numbers. Many System x tools, including those tools that we have mentioned, are located on the following configuration tools website: http://www.ibm.com/systems/x/hardware/configtools.html See the ServerProven website for a complete list of the available options: http://www.ibm.com/systems/info/x86servers/serverproven/compat/us/ In any circumstance where this list of options differs from the options that are shown in ServerProven, use ServerProven as the definitive resource. The main function of ServerProven is to show the options that have been successfully tested by IBM with a System x server. Another useful page in the ServerProven site is the list of vendors. On the home page for ServerProven, click the industry leaders link, as shown in Figure 3-28.
You can use the following ServerProven web address: http://www.ibm.com/systems/info/x86servers/serverproven/compat/us/serverproven 108
IBM eX5 Implementation Guide
This page lists the third-party vendors that have performed their own testing of their options with our servers. This support information means that those vendors agree to support the combinations that are shown in those particular pages. Tip: To see the tested hardware, click the logo of the vendor. Clicking the About link under the logo takes you to a separate About page. Although IBM supports the rest of the System x server, technical issues traced to the vendor card are, in most circumstances, directed to the vendor for resolution.
Electrical input: 100 - 127 V, 200 - 240 V, 50 - 60 Hz Approximate input kilovolt-amperes (kVA): Minimum: 0.25 kVA Typical: 0.85 kVA Maximum: 1.95 kVA (110 V ac) Maximum: 2.17 kVA (220 V ac)
Chapter 3. IBM System x3850 X5 and x3950 X5
109
3.13.4 UEFI
The x3850 X5 uses an integrated Unified Extensible Firmware Interface (UEFI) next-generation BIOS. UEFI includes the following capabilities: Human-readable event logs; no more beep codes Complete setup solution by allowing adapter configuration function to be moved to UEFI Complete out-of-band coverage by Advanced Settings Utility to simplify remote setup
110
Using all of the features of UEFI requires an UEFI-aware operating system and adapters. UEFI is fully backward-compatible with BIOS. For more information about UEFI, see the IBM white paper, Introducing UEFI-Compliant Firmware on IBM System x and BladeCenter Servers, which is available at the following website: http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5083207 For UEFI menu setup, see 6.9, UEFI settings on page 259.
111
Figure 3-29 shows the light path diagnostics panel on the x3850 X5.
Full details about the functionality and operation of the light path diagnostics in this system are in the IBM System x3850 X5 and x3950 X5 Problem Determination and Service Guide, which is available at the following website: http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5083418
112
Six strategically located hot-swap/redundant fans, combined with efficient airflow paths, provide highly effective system cooling for the eX5 systems. This technology is known as IBM Calibrated Vectored Cooling technology. The fans are arranged to cool three separate zones, with one pair of redundant fans per zone. The fans automatically adjust speeds in response to changing thermal requirements, depending on the zone, redundancy, and internal temperatures. When the temperature inside the server increases, the fans speed up to maintain the proper ambient temperature. When the temperature returns to a normal operating level, the fans return to their default speed. All x3850 X5 system fans are hot-swappable, except for Fan 3 in the bottom x3850 X5 of a 2-node complex, when QPI cables directly link the two servers.
A fan that is located inside each power supply cools the power modules. MAX5 has five redundant hot-swap fans, which are all in one cooling zone. The IMM of the attached host controls the MAX5 fan speed, based on altitude and ambient temperature. Fans also respond to certain conditions and come up to speed accordingly: If a fan fails, the remaining fans ramp up to full speed. As the internal temperature rises, all fans ramp to full speed.
Chapter 3. IBM System x3850 X5 and x3950 X5
113
114
Because a short delay in qualification for several of these operating systems might exist, check the ServerProven Operating System support page for a current statement at the following website: http://www.ibm.com/servers/eserver/serverproven/compat/us/nos/matrix.shtml Table 3-36 summarizes the hardware maximums of the possible x3850 X5 configurations.
Table 3-36 Thread and memory maximums x3850 X5 x3850 X5 with MAX5 64 1.5 TB Two-node x3850 X5 128 2 TB Two-node x3850 X5 with MAX5 128 3 TB
64 1 TB
115
116
Chapter 4.
117
118
Figure 4-1 shows the x3690 X5 server with 16 hot-swap 2.5-inch SAS disk drives installed. The x3690 server has these physical specifications: Height: 86 mm (3.5 inches, 2U) Depth: 698 mm (27.4 inches) Width: 429 mm (16.8 inches) Maximum weight: 31.3 kg (69 lb) when fully configured Each disk drive has an orange-colored bar. This color denotes that these disks are hot-swappable. The color coding that is used throughout the system is orange for hot-swappable and blue for non-hot-swappable. The only hot-swappable parts in this server are the HDDs, fans, and power supplies. All other parts require that the server is powered off before removing that component.
Drive bays
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
Rack release latch Drive activity LED (green) Drive status LED (amber)
Rack release latch Optical drive Optical drive activity LED eject button
119
Ethernet 1
Serial connector
QPI port 1
QPI port 2
Figure 4-4 shows the system with the top cover removed. Five PCIe 2.0 slots Two CPU sockets (partially covered by the memory mezzanine) Memory mezzanine with 16 DIMM sockets. A further 16 DIMMs are on the system planar underneath (not visible).
Drive bays
120
important: The x3950 X5 top cover cannot be removed and the server remain powered on. If the top cover is removed, the server powers off immediately.
Figure 4-5 x3690 X5 with the attached MAX5 memory expansion unit
The MAX5 has the following specifications: IBM EXA5 chip set. Intel memory controller with eight memory ports (four DIMMs on each port). Intel QPI architecture technology to connect the MAX5 to the x3690 X5. There are two QPI links and each QPI link operates at up to 6.4 GT/s depending on the processors installed. Memory DIMMs: Minimum: 2 DIMMs, 4 GB. Maximum: 32 DIMM connectors (up to 512 GB of memory using 16 GB DIMMs). Type of DIMMs: PC3-10600, 1067 MHz, error checking and correction (ECC), and DDR3 registered SDRAM DIMMs. Support for 2 GB, 4 GB, 8 GB, and 16 GB DIMMs.
121
Five hot-swap 40 mm fans. Power supply: Hot-swap power supplies with built-in fans for redundancy support. 675-watt (110 - 220 V ac auto-sensing). One power supply standard; a second redundant power supply is optional. Light path diagnostics LEDs: Board LED Configuration LED Fan LEDs Link LED (for QPI and EXA5 links) Locate LED Memory LEDs Power-on LED Power supply LEDs Width: 483 mm (19.0 inches) Depth: 724 mm (28.5 inches) Height: 44 mm (1.73 inches) (1U rack unit) Basic configuration: 12.8 kg (28.2 lb) Maximum configuration: 15.4 kg (33.9 lb)
Physical specifications:
Tip: The MAX5 that is used with the x3690 X5 is the same as the MAX5 offered with the x3850 X5. With the addition of the MAX5 memory expansion unit, the x3690 X5 gains an additional 32 DIMM sockets for a total of 64 DIMM sockets. Using 16 GB DIMMs, you can install a total of 1 TB of RAM. All DIMM sockets in the MAX5 are accessible, regardless of whether one or two processors are installed in the x3690 X5. Figure 4-6 shows the ports at the rear of the MAX5 memory expansion unit. When connecting the MAX5 to an x3690 X5, the QPI ports are used. The EXA ports are unused.
EXA port 1 EXA port 2 EXA port 3 QPI QPI port 1 port 2 QPI port 3 QPI port 4 Power connectors
EXA port 1 link LED EXA port 2 link LED EXA port 3 link LED
Figure 4-7 on page 123 shows the internals of the MAX5, including the IBM EXA chip that acts as the interface to the QPI links from the x3690 X5.
122
32 DIMM sockets
For an in-depth look at the MAX5 offering, see 4.5, MAX5 on page 127.
123
4.3 Models
In addition to the details in the tables in this chapter, each standard model has the following specifications: The servers have 16 DIMM sockets on the system planar. The additional 16-DIMM socket memory mezzanine (memory tray) is optional on most models and must be ordered separately. See 4.8, Memory on page 131 for details. The MAX5 is optional on certain models and standard on others. The optical drive is not standard and must be ordered separately if an optical drive is required. See 4.9.8, Optical drives on page 163 for details. As noted in the tables, most models have drive bays standard (std). However, disk drives are not standard and must be ordered separately. In the tables, max indicates maximum.
Intel Xeon processors (two maximum) 1x E7520 4C, 1.86 GHz, 95W 1x E7520 4C, 1.86 GHz, 95W 1x E6540 6C, 2.00 GHz, 105W 1x X6550 8C, 2.00 GHz, 130W 1x X6550 8C, 2.00 GHz, 130W 1x X7560 8C, 2.26 GHz, 130W 1x X7550 8C, 2.00GHz, 130W 1x X7560 8C, 2.26GHz, 130W
Memory tray
Memory speed 800 MHz 800 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz
Standard memorya Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB Server: 2x 4GB MAX5: 2x 4GB Server: 2x 4GB MAX5: 2x 4GB
Power supplies std/max 1/4 1/4 1/4 1/4 1/4 1/4 Server: 2/4 MAX5: 1/2 Server: 2/4 MAX5: 1/2
Drive bays std/max None 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16 4x 2.5/16
a. Up to 64 DIMM sockets: Each server has 16 DIMM sockets standard or 32 sockets with the addition of the internal memory tray (mezzanine). With the addition of the MAX5 memory expansion unit, 64 DIMM sockets total are available. b. Emulex 10Gb Ethernet Adapter
124
Model
Memory speed
MAX5
Standard memorya
Database workload-optimized models 7148-3Dx 2x X6550 8C, 2.00 GHz, 130W 1066 MHz Opt Server: 4x 4 GB Std Opt Opt Server: 4/4 16x 1.8/24
Virtualization workload-optimized models 7148-2Dx 2x E6540 6C, 2.00 GHz, 105W 1066 MHz Std Server: 32x 4GB MAX5: 32x 4GB Std Opt Std Server: 4/4 MAX5: 2/2 4x 2.5/16
a. Up to 64 DIMM sockets: Each server has 16 DIMM sockets standard or 32 sockets with the addition of the internal memory tray (mezzanine). With the addition of the MAX5 memory expansion unit, 64 DIMM sockets total are available. b. Emulex 10Gb Ethernet Adapter.
125
PCIe x16
OR
QPI
ESI x4
Memory tray (16 DIMMs) SMI links Memory buffer Memory buffer Memory buffer Memory buffer FPGA Video ports Serial port 10/100 Management port IBM IMM Controller Intel Xeon Processor 1
QPI
Intel Southbridge
SAS Adapters
PCIe x1
PCIe x4
SAS Backplanes
Figure 4-9 shows the block diagram of the MAX5. The MAX5 is connected to the x3690 X5 using two cables connecting the two QPI ports on the server to two of the QPI ports on the MAX5. The EXA ports and other two QPI are unused in this configuration.
External connectors QPI QPI QPI QPI EXA EXA EXA
Memory buffer Memory buffer Memory buffer Memory buffer SMI links IBM EXA chip
SMI links
126
4.5 MAX5
As introduced in 4.1.2, IBM MAX5 memory expansion unit on page 121, the MAX5 memory expansion drawer is available for the x3690 X5. Certain standard models include the MAX5, as described in 4.3, Models on page 124, and the MAX5 can also be ordered separately, as listed in Table 4-3.
Table 4-3 Ordering information for the IBM MAX5 for System x Part number 59Y6265 60Y0332 59Y6269 Feature code 4199 4782 7481 Description IBM MAX5 for System x IBM 675W HE Redundant Power Supply IBM MAX5 to x3690 X5 Cable Kit (two cables)
The eX5 chip set in the MAX5 is an IBM unique design that attaches to the QPI links as a node controller, giving it direct access to all CPU bus transactions. It increases the number of DIMMs supported in a system by a total of 32, and also adds another 16 channels of memory bandwidth, boosting overall throughput. The MAX5 adds additional memory performance. The eX5 chip connects directly through QPI links to both CPUs in the x3690 X5, and it maintains a directory of each CPUs last-level cache. This directory allows the eX5 chip to respond to memory requests prior to the end of a broadcast snoop cycle, thereby improving performance. For more information about eX5 technology, see 2.1, eX5 chip set on page 16. Figure 4-10 shows a diagram of the MAX5.
External connectors QPI QPI QPI QPI EXA EXA EXA
Memory buffer Memory buffer Memory buffer Memory buffer SMI links IBM EXA chip
SMI links
127
The MAX5 is connected to the x3690 X5 using two cables, connecting the QPI ports on the server to two of the four QPI ports on the MAX5. The other two QPI ports of the MAX5 are unused. The EXA ports are for future scaling capabilities. Figure 4-11 shows architecturally how a single-node x3690 X5 is connected to a MAX5.
External QPI cables
x3690 X5
16 DIMMs on system planar CPU 0 QPI SMI links Memory tray (16 DIMMs) QPI links CPU 1
MAX5
EXA 32 DIMMs on MAX5
QPI
QPI
EXA
EXA
EXA
Figure 4-11 Connectivity of the x3690 X5 with a MAX5 memory expansion unit
As shown in Figure 4-11, the x3690 X5 attaches to the MAX5 using QPI links; you can see that the eX5 chip set in the MAX5 simultaneously connects to both CPUs in the server. One benefit of this connectivity is that the MAX5 is able to store a copy of the contents of the last-level cache of all the CPUs in the server. Therefore, when a CPU requests content stored in the cache of another CPU, the MAX5 not only has that same data stored in its own cache, it is able to return the acknowledgement of the snoop and the data to the requesting CPU in the same transaction. For more information about QPI links and snooping, see 2.2.4, QuickPath Interconnect (QPI) on page 18. Tip: The Xeon E6510 processor does not support the use of the MAX5. Connectivity of the MAX5 to the x3690 X5 is described in 4.6, Scalability on page 128. For memory configuration information, see 4.8.3, MAX5 memory on page 136. For a description of the power and fans, see 4.12, Power supplies on page 173.
4.6 Scalability
In this section, we describe how the x3690 X5 can be expanded to increase the number of memory DIMMs.
128
QPI
QPI
QPI
The x3690 X5 supports the following configurations: A single x3690 X5 server with two processor sockets. This configuration is sometimes referred to as a single-node server. A single x3690 X5 server with a single MAX5 memory expansion unit attached. This configuration is sometimes referred to as a memory-expanded server. Two-node configurations: The 2-node configurations (with and without MAX5) are not supported. The MAX5 memory expansion unit permits the x3690 X5 to scale to an additional 32 DDR3 DIMM sockets. Connecting the single-node x3690 X5 with the MAX5 memory expansion unit uses two QPI cables, part number 59Y6269, as listed in Table 4-4. Figure 4-12 shows the connectivity. Number of processors: The MAX5 is supported with either one or two processors installed in the x3690 X5. However, the recommendation is to have two processors installed and memory installed in every DIMM socket in the server to maximize performance.
Rack rear
Connecting the MAX5 to a single-node x3690 X5 requires one IBM MAX5 to x3690 X5 Cable Kit, which consists of two QPI cables. See Table 4-4.
Table 4-4 Ordering information for the IBM MAX5 to x3690 X5 Cable Kit Part number 59Y6269 Feature code 7481 Description IBM MAX5 to x3690 X5 Cable Kit (two cables)
Note: There is no setup on the Integrated Management Module (IMM) for scalability. Ensure that you have updated all firmware before attaching the MAX5, especially the Field Programmable Gate Array (FPGA). For more information about updating firmware, see 9.10, Firmware update tools and methods on page 509.
129
Advanced processors (X) 60Y0311 60Y0313 60Y0321 60Y0319 4469 4471 4479 4477 Xeon X7560 8C Xeon X7550 8C Xeon X7542 6C Xeon X6550 8C 2.26 GHz 2.00 GHz 2.66 GHz 2.00 GHz 24 MB 18 MB 18 MB 18 MB 6.4 GT/s 6.4 GT/s 5.86 GT/s 6.4 GT/s 1066 MHz 1066 MHz 977 MHz 1066 MHz 130W 130W 130W 130W Yes Yes No Yes Yes Yes Yes Yes Yes Yes Yes Yes
Standard processors (E) 60Y0315 60Y0316 60Y0317 60Y0320 60Y0318 4473 4474 4475 4478 4476 Xeon E7540 6C Xeon E7530 6C Xeon E7520 4C Xeon E6540 6C Xeon E6510 4C 2.00 GHz 1.86 GHz 1.86 GHz 2.00 GHz 1.73 GHz 18 MB 12 MB 18 MB 18 MB 12 MB 6.4 GT/s 5.86 GT/s 4.8 GT/s 6.4 GT/s 4.8 GT/s 1066 MHz 977 MHz 800 MHz 1066 MHz 800 MHz 105W 105W 95W 105W 105W Yes Yes Yes Yes Yes Yes Yes No Yes No Yes Yes Yes Yes No
Low-power processors (L) 60Y0312 60Y0314 4470 4472 Xeon L7555 8C Xeon L7545 6C 1.86 GHz 1.86 GHz 24 MB 18 MB 5.86 GT/s 5.86 GT/s 977 MHz 977 MHz 95W 95W Yes Yes Yes Yes Yes Yes
a. Thermal design power b. Intel Hyper-Threading Technology c. Intel Turbo Boost Technology
Clarification: The Xeon E6510 does not support the use of the MAX5 memory expansion unit. The x3690 X5 announcement letter incorrectly reports that the L3 cache of the X7542 is 12 MB. Also, the announcement letter incorrectly states that the E6540 does not have Turbo Boost mode. See 2.2, Intel Xeon 6500 and 7500 family processors on page 16 for an in-depth description of the Intel Xeon 6500/7500 processor family and features. Most processors support Intel Turbo Boost Technology with a couple of exceptions, as listed in Table 4-5. When a CPU operates beneath its thermal and electrical limits, Turbo Boost dynamically increases the processors clock frequency by 133 MHz on short and regular intervals until an upper limit is reached. See 2.2.3, Turbo Boost Technology on page 18 for more information. With the exception of the X7542, all CPUs that are listed support Intel Hyper-Threading Technology. Hyper-Threading Technology (HT) is an Intel technology that is used to improve 130
IBM eX5 Implementation Guide
the parallelization of workloads. When Hyper-Threading is engaged in the BIOS, for each processor core that is physically present, the operating system addresses two. For more information, see 2.2.2, Hyper-Threading Technology on page 17. All CPU options include a heat-sink. The x3690 X5 models include one CPU standard. All five PCIe slots are usable, even with only one processor installed, as shown in Figure 4-8 on page 126. The second CPU is required to access the memory in the memory mezzanine (if the memory mezzanine is installed). The second CPU can be installed without the memory mezzanine, but its only access to memory is through the primary CPU. For optimal performance, if two CPUs are installed, install a memory mezzanine also. Follow these population guidelines: Each CPU requires a minimum of two DIMMs to operate. If the memory mezzanine is installed, it needs a minimum of two DIMMs installed. Both processors must be identical. Consider the X7542 processor for CPU frequency-dependent workloads, because it has the highest core frequency of the available processor models. The MAX5 is supported with either one or two processors installed in the x3690 X5, although the recommendation is to have two processors installed and memory installed in every DIMM socket in the server to maximize performance. If high processing capacity is not required for your application but high memory bandwidth is required, consider using two processors with fewer cores or a lower core frequency rather than two processors with more cores or a higher core frequency. Having four processors can enable all memory channels and can maximize memory bandwidth. We describe this technique in 4.8, Memory on page 131.
4.8 Memory
The x3690 X5 offers up to 32 DIMM sockets that are internal to the server chassis, plus an additional 32 DIMM sockets in the MAX5 memory expansion unit. This section covers the following topics: 4.8.1, Memory DIMM options on page 133 4.8.2, x3690 X5 memory population order on page 133 4.8.3, MAX5 memory on page 136 4.8.4, Memory balance on page 139 4.8.5, Mixing DIMMs and the performance effect on page 140 4.8.6, Memory mirroring on page 141 4.8.7, Memory sparing on page 143 4.8.8, Effect on performance of using mirroring or sparing on page 144 Implement the memory DIMMs that are internal to the x3690 X5 chassis: 16 DIMM sockets on the system planar 16 DIMM sockets in an optional memory mezzanine Tip: The memory mezzanine is referred to in the announcement letter as the memory expansion card. It is referred to as the memory tray, in the Installation and Users Guide - IBM System x3690 X5.
131
The memory mezzanine is an optional component and orderable as listed in Table 4-6.
Table 4-6 x3690 X5 memory mezzanine option part number Option 60Y0323 Feature code 9278 Description IBM x3690 X5 16-DIMM Internal Memory Expansion
Memory mezzanine
Memory mezzanine
With the Intel Xeon 6500 and 7500 processors, the memory controller is integrated into the processor, as shown in the architecture block diagram in Figure 4-8 on page 126: Processor 0 connects directly to the memory buffers and memory DIMM sockets on the system planar. Processor 1 connects directly to the memory buffers and memory in the memory mezzanine. If you plan to install the memory mezzanine, you are required to also install the second processor. The x3690 X5 uses the Intel scalable memory buffer to provide DDR3 SDRAM memory functions. The memory buffers connect to the memory controller in each processor through Intel Scalable Memory Interconnect links. Each memory buffer has two memory channels, and the DIMM sockets are connected to the memory buffer with two DIMMs per memory channel (2 DPC).
132
The memory uses DDR3 technology and operates at memory speeds of 800, 978, and 1066 MHz. The memory speed is dictated by the memory speed of the processor (see Table 4-5 on page 130). For more information about how this is calculated, see 2.3.1, Memory speed on page 22.
2 GB (1x 2GB) 1Rx8, 2Gbit PC3-10600R DDR3-1333 4 GB (1x 4GB), 2Rx8, 2Gbit PC3-10600R DDR3-1333 4 GB (1x 4GB), 4Rx8, 1Gbit PC3-8500 DDR3-1066 8 GB (1x 8GB), 4Rx8, 2Gbit PC3-8500 DDR3-1066 16 GB (1x 16GB), 4Rx4, 2Gbit PC3-8500 DDR3-1066
a. Memory speed is also controlled by the memory bus speed, as specified by the selected processor model. The actual memory bus speed is the lower of both the processor memory bus speed and the DIMM memory bus speed. b. Although 1333 MHz memory DIMMs are supported in the x3690 X5, the memory DIMMs run at a maximum speed of 1066 MHz. c. The 16 GB memory option is only supported in the MAX5 when it is the only type of memory used in the MAX5. You cannot use any other memory options in the MAX5 if this option is installed in the MAX5. This DIMM also supports redundant bit steering (RBS) when used in the MAX5, as described in Redundant bit steering on page 29.
Details: Only certain memory options supported in the x3690 X5 are also supported in the MAX5, as indicated in Table 4-7. When ordering DIMMs for the MAX5 with feature codes (configure-to-order (CTO) clients only), use the special MAX5 feature codes 24xx as listed. See MAX5 memory options on page 137 for more information. Memory options must be installed in matched pairs. Single options cannot be installed; therefore, the options listed must be ordered in a quantity of two. The maximum memory speed that is supported by Xeon 7500 and 6500 (Nehalem-EX) processors is 1066 MHz. DIMMs will not run at 1333 MHz. Mixing DIMMs is supported by all DIMMs, except 16 GB DIMMs, and operates at the speed of the slowest installed DIMM. We do not recommend mixing DIMMs for performance reasons.
133
134
Table 4-8 One or two processor DIMM installation when the memory mezzanine is not installed Hemisphere Modea Memory buffer Memory buffer Memory buffer Memory buffer
Number of processors
Number of DIMMs
DIMM 10
DIMM 11
DIMM 12
DIMM 13
DIMM 14
DIMM 15 x x x
1 or 2 1 or 2 1 or 2 1 or 2 1 or 2 1 or 2 1 or 2 1 or 2
2 4 6 8 10 12 14 16
N Y N Y N Y N Y
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26.
Important: If using two processors with no mezzanine board, memory is not in a nonuniform memory access (NUMA)-compliant state, which causes great performance degradation. See 2.3.4, Nonuniform memory architecture (NUMA) on page 26 for details.
Important VMware ESX considerations: When installing and running VMware ESX on this server, the operating system might fail to install or boot with the following error message when the server memory configuration is not NUMA-compliant: NUMA node 1 has no memory There are only three possible configurations to support VMware: One processor is installed and no mezzanine board is installed. Two processors are installed and matching memory is installed on both the system board and the mezzanine board. Two processors are installed, no internal memory is installed, and the memory installed in an attached MAX5 memory expansion is configured as non-pooled memory.
DIMM 16
DIMM 1
DIMM 2
DIMM 3
DIMM 4
DIMM 5
DIMM 6
DIMM 7
DIMM 8
DIMM 9
135
Figure 4-15 Memory mezzanine tray Table 4-9 NUMA-compliant DIMM installation: Two processors and the memory mezzanine installed Hemisphere Modea Number of DIMMs Processor 1 (planar DIMMs) Buffer Buffer Buffer DIMM 10 DIMM 11 DIMM 12 DIMM 13 Buffer DIMM 14 DIMM 15 DIMM 16 DIMM 17 Buffer DIMM 18 DIMM 19 DIMM 20 DIMM 21 Processor 2 (mezzanine DIMMs) Buffer DIMM 22 DIMM 23 DIMM 24 DIMM 25 Buffer DIMM 26 DIMM 27 DIMM 28 DIMM 29 Buffer DIMM 30 x x x x x x x x x x DIMM 31 DIMM 32 x x x x x x x x x x x x x x x
DIMM 1
DIMM 2
DIMM 3
DIMM 4
DIMM 5
DIMM 6
DIMM 7
DIMM 8 x x
4 8 12 16 20 24 28 32
N Y N Y N Y N Y
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
DIMM 9
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x
x x x x x x
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26.
Tip: Table 4-9 lists only memory configurations that are considered best practice in obtaining optimal memory and processor performance. For a full list of supported memory configurations, see the IBM System x3690 X5 Installation and User Guide or the IBM System x3690 X5 Problem Determination and Service Guide. You can obtain both of these documents at the following website: http://www.ibm.com/support
136
2 GB (1x 2GB) 1Rx8, 2Gbit PC3-10600R DDR3-1333 4 GB (1x 4GB), 2Rx8, 2Gbit PC3-10600R DDR3-1333 8 GB (1x 8GB), 4Rx8, 2Gbit PC3-8500 DDR3-1066 16 GB (1x 16GB), 4Rx4, 2Gbit PC3-8500 DDR3-1066
a. Memory speed is also controlled by the memory bus speed, as specified by the processor model selected. The actual memory bus speed is the lower of both the processor memory bus speed and the DIMM memory bus speed. b. Although 1333 MHz memory DIMMs are supported in the x3690 X5, the memory DIMMs run at a maximum speed of 1066 MHz. c. The 16 GB memory option is only supported in the MAX5 when it is the only type of memory used in the MAX5. No other memory options can be used in the MAX5 if this option is installed in the MAX5. d. This DIMM supports redundant bit steering (RBS), as described in Redundant bit steering on page 29.
Use of the 16 GB memory option: The 16 GB memory option, part number 46C7483, is supported in the MAX5 only when it is the only type of memory used in the MAX5. No other memory options can be used in the MAX5 if this option is installed in the MAX5.
Redundant bit steering: Redundant bit steering (RBS) is not supported on the x3690 X5 itself, because the integrated memory controller of the Intel Xeon 7500 processor does not support the feature. See Redundant bit steering on page 29 for details. The MAX5 memory expansion unit supports RBS, but only with x4 memory and not x8 memory. As shown in Table 4-10, the 16 GB DIMM, part number 46C7483, uses x4 DRAM technology. RBS is automatically enabled in the MAX5 memory port, if all DIMMs that are installed to that memory port are x4 DIMMs.
137
One important consideration when installing memory in MAX5 configurations is that the server must be fully populated before adding DIMMs to the MAX5. As we describe in 2.3.2, Memory DIMM placement on page 23, you get the best performance by using all memory buffers and all DIMM sockets on the server first, and then adding DIMMs to the MAX5. Figure 4-16 shows the numbering scheme for the DIMM slots on the MAX5, and the pairing of DIMMs in the MAX5. Because DIMMs are added in pairs, they must be matched on a memory port (as shown using the colors). For example, DIMM1 is matched to DIMM 8, DIMM 2 to DIMM 7, DIMM 20 to DIMM 21, and so on.
16 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1
DIMM 15
DIMM 12
DIMM 11
DIMM 16
DIMM 14
DIMM 13
DIMM 10
DIMM 9
DIMM 8
DIMM 7
DIMM 6
DIMM 5
DIMM 4
DIMM 3
DIMM 2
Memory buffer 4
Memory buffer 3
Memory buffer 5
Memory buffer 6
Quad D Quad C 32 31 30 29 28 27 26 25
Quad B Quad A 24 23 22 21 20 19 18 17
DIMM 32
DIMM 29
DIMM 28
DIMM 26
DIMM 25
DIMM 24
DIMM 23
DIMM 22
DIMM 21
DIMM 20
DIMM 19
DIMM 31
DIMM 30
DIMM 27
Memory buffer 1
Memory buffer 2
Memory buffer 8
DIMM 18
DIMM 17
Memory buffer 7
Quad H Quad G
Quad F Quad E
Table 4-11 shows the population order of the MAX5 DIMM slots, which ensures that memory is balanced among the memory buffers. The colors in Table 4-11 match the colors in Figure 4-16.
Table 4-11 DIMM pair 1 2 3 DIMM installation sequence in the MAX5 DIMM slot 28 and 29 9 and 16 1 and 8
138
DIMM 1
DIMM pair 4 5 6 7 8 9 10 11 12 13 14 15 16
DIMM slot 20 and 21 26 and 31 11 and 14 3 and 6 18 and 23 27 and 30 10 and 15 2 and 7 19 and 22 25 and 32 12 and 13 4 and 5 17 and 24
139
access only to remote or far memory. So, threads executing on Processor 1 have a longer latency to access memory as compared to threads on Processor 0. This situation is due to the latency penalty incurred to traverse the QPI links to access the data on the other processors memory controller. The bandwidth to remote memory is also limited by the capability of the QPI links. The latency to access remote memory is more than 50% higher than local memory access. For these reasons, we advise that you populate all processors with memory, remembering the requirements that are necessary to ensure optimal interleaving and Hemisphere Mode.
LOCAL
REMOTE
QPI links
Memory controller
Memory controller
Buffer DIMM
Buffer
Buffer
DIMM DIMM
DIMM DIMM
Figure 4-17 Memory latency when not spreading DIMMs across both processors
memory channels, with configurations balanced across the memory buffers, so that Hemisphere Mode is maintained.
Relative performance: 100 Relative performance: 97
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
DIMM DIMM
Relative performance: 92
Relative performance: 82
Buffer
Buffer
DIMM
DIMM DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
As you can see, mixing DIMM sizes can cause performance loss up to 18%, even if all channels are occupied and Hemisphere Mode is maintained.
141
Table 4-13 shows the DIMM population order for memory-mirroring mode without the mezzanine installed.
Table 4-12 Mirror DIMM installation: Two processors and no memory mezzanine installed Number of DIMMs Processor 1 (planar DIMMs) Buffer DIMM 1 DIMM 2 DIMM 3 DIMM 4 DIMM 5 Buffer DIMM 6 DIMM 7 DIMM 8 DIMM 9 Buffer DIMM 10 DIMM 11 DIMM 12 DIMM 13 Buffer DIMM 14 DIMM 15 DIMM 16 DIMM 17 Processor 2 (No mezzanine DIMMs) Buffer DIMM 18 DIMM 19 DIMM 20 DIMM 21 Buffer DIMM 22 DIMM 23 DIMM 24 DIMM 25 Buffer DIMM 26 DIMM 27 DIMM 28 DIMM 29 Buffer DIMM 30 DIMM 30 x x x x x x x DIMM 31 DIMM 31 DIMM 32 DIMM 32 x x x x x x x x x
4 8 12 16
x x x x x x x x x x x x x x x x
x x x x
x x x x x x x x x x x x x x x x
x x x x
Table 4-13 DIMM population order: Memory-mirroring mode without the mezzanine installed Sets of DIMMs Set 1 Set 2 Set 3 Set 4 Number of installed processors 1 or 2 1 or 2 1 or 2 1 or 2 DIMM connector population sequence with no memory tray 1, 8, 9, 16 3, 6, 11, 14 2, 7, 10, 15 4, 5, 12, 13
Table 4-14 lists the DIMM installation sequence for memory-mirroring mode when two processors and a memory tray are installed in the server.
Table 4-14 Mirror DIMM installation: Two processors and memory mezzanine installed Number of DIMMs Processor 1 (planar DIMMs) Buffer DIMM 1 DIMM 2 DIMM 3 DIMM 4 DIMM 5 Buffer DIMM 6 DIMM 7 DIMM 8 DIMM 9 Buffer DIMM 10 DIMM 11 DIMM 12 DIMM 13 Buffer DIMM 14 DIMM 15 DIMM 16 DIMM 17 Processor 2 (with mezzanine DIMMs) Buffer DIMM 18 DIMM 19 DIMM 20 DIMM 21 Buffer DIMM 22 DIMM 23 DIMM 24 DIMM 25 Buffer DIMM 26 DIMM 27 DIMM 28 DIMM 29 Buffer
8 16 24 32
x x x x x x x x x x x x x x x x
x x x x
x x x x x x x x x x x x x x x x
x x x x
x x x x x x x x x x x x x x x x
x x x x
x x x x
142
Table 4-15 DIMM population order: Memory-mirroring mode with the mezzanine installed Sets of DIMMs Number of installed processors 2 2 2 2 DIMM connector population sequence on the system board 1, 8, 9, 16 3, 6, 11, 14 2, 7, 10, 15 4, 5, 12, 13 DIMM connector population sequence on the memory tray 17, 24, 25, 32 19, 22, 27, 30 18, 23, 26, 31 20, 21, 28, 29
DIMM 10
DIMM 11
DIMM 12
DIMM 13
DIMM 14
DIMM 15
DIMM 16
DIMM 17
DIMM 18
DIMM 19
DIMM 20
DIMM 21
DIMM 22
DIMM 23
DIMM 24
DIMM 25
DIMM 26
DIMM 27
DIMM 28
DIMM 29
DIMM 30 x x x x
DIMM 31 x x x x x x
4 8 12 16 20 24 28 32 x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x
x x x x x x x x
143
DIMM 32 x x
DIMM 1
DIMM 2
DIMM 3
DIMM 4
DIMM 5
DIMM 6
DIMM 7
DIMM 8
DIMM 9
These options are configured by using the UEFI during the boot sequence.
Sparing
62
Mirroring
50
Normal
100
20
40
60
80
100
120
144
4.9 Storage
The x3690 X5 has internal capacity of up to sixteen 2.5-inch drives, as shown in Figure 4-20. The server supports 2.5-inch disk drives or solid-state drives (SSDs), or 1.8-inch SSDs.
Figure 4-20 Front of the x3690 X5 with sixteen 2.5-inch drive bays
This section covers the following topics: 4.9.1, 2.5-inch SAS drive support on page 145 4.9.2, IBM eXFlash and SSD disk support on page 149 4.9.3, SAS and SSD controller summary on page 152 4.9.4, Battery backup placement on page 155 4.9.5, ServeRAID Expansion Adapter on page 157 4.9.6, Drive combinations on page 158 4.9.7, External SAS storage on page 162 4.9.8, Optical drives on page 163 See the IBM ServerProven website for the latest supported options: http://www.ibm.com/systems/info/x86servers/serverproven/compat/us/
Backplanes
Most standard models of the x3690 X5 include one SAS backplane supporting four 2.5-inch SAS disks, as listed in 4.3, Models on page 124. Additional backplanes can be added to increase the supported number of SAS disks to 16 (using part number 60Y0338 for a 8x backplane and part number 60Y0369 for a 4x backplane). The database model 7148-3Dx has two IBM eXFlash 8x 1.8-inch HS SAS SSD backplanes as standard. See 4.3, Models on page 124 for details. The standard backplanes are installed in the leftmost sections. Table 4-17 on page 146 lists the backplane options. These backplanes support both SAS and SSD 2.5-inch drives. The specific combinations of the backplanes that are supported are listed in 4.9.6, Drive combinations on page 158.
145
Table 4-17 x3690 X5 hard drive backplanes Part number 60Y0339 60Y0381 Feature code 9287 1790 Backplane IBM 4x 2.5 HS SAS HDD Backplane IBM 8x 2.5 HS SAS HDD Backplane Drives supported Four 2.5 SAS drives Eight 2.5 SAS drives SAS cables includeda 1 short, 1 long 2 short, 2 long
a. See the next paragraph for a description of short and long cables. The option part numbers include the cables. If you order a configuration by using feature codes, use Table 4-18.
As listed in Table 4-17, the backplane option part numbers include the necessary cables to connect the backplane to the SAS controller. The short SAS cable is needed when installing a hard drive backplane for 2.5-inch bays 1 - 8 (the left half of the drive bays in the server, when looking from the front). The long SAS cable is used for hard drive backplanes for 2.5-inch bays 9 - 16 (the right half of the drive bays, when looking from the front). When configuring an order by using feature codes, for example, with configure-to-order (CTO), the feature codes for the backplanes do not include the cables. You must order the cables separately, as listed in Table 4-18.
Table 4-18 x3690 X5 SAS cable options (not needed if ordering backplane part numbers) Part number 69Y2322 69Y2323 Feature code 6428 6429 Description x3690 X5 short SAS cable x3690 X5 long SAS cable When used For backplanes of bays 1 - 8 For backplanes of bays 9 - 16
Using the ServeRAID Expansion Adapter: When using this adapter, the adapter must be installed in PCIe slot 1. Short SAS cables are used to connect the two ports of the ServeRAID controller to the two controller I/O ports on the expander. All four backplane SAS cable connections are connected to the ServeRAID Expander using the long SAS cables that are shown in Table 4-18 on page 146.
146
Figure 4-21 and Figure 4-23 on page 150 show the backplanes and their cable connections.
Cofiguration connector SAS signal connector SAS signal connector
Cofiguration connector
Backplane used 4x/8x SAS HDD 4x/8x SAS HDD 4x/8x SAS HDD 4x/8x SAS HDD 4x/8x SAS HDD 4x/8x SAS HDD 4x/8x SAS HDD
a. Using the self-encrypting drive (SED) feature of these drives requires a ServeRAID M5014 or M5015 RAID controller, plus either of the ServeRAID M5000 keys, as listed in Table 4-22 on page 149, to add SED support.
Table 4-20 lists the 2.5-inch Nearline SATA 7.2K drives that are supported in the x3690 X5. These drives are supported with the SAS hard disk backplane, 60Y0338 and 60Y0369.
Table 4-20 Supported 2.5-inch Nearline SATA drives Part number 42D0709 Feature code 5409 Description IBM 500GB 7200 NL SATA 2.5'' SFF Slim-HS HDD
147
Self-encrypting drives (SEDs) are also an available option as listed in Table 4-19 on page 147. SEDs provide cost-effective advanced data security with Advanced Encryption Standard (AES) 128 disk encryption. To make use of the encryption capabilities, you must also use either a ServeRAID M5014 or M5015 RAID controller, plus either the ServeRAID M5000 Advance Feature Key or the Performance Accelerator Key. See Controller options with 2.5-inch drives on page 149 for details. For more information about SEDs, see the IBM Redbooks at-a-glance guide, Self-Encrypting Drives for IBM System x, TIPS0761, which is available at this website: http://www.ibm.com/redbooks/abstracts/tips0761.html
Figure 4-22 Location of the single SATA drive in the x3690 X5 Table 4-21 x3690 X5 Single SATA HDD Bay kit Option 60Y0333 Feature code 9284 Description IBM x3690 X5 Single SATA HDD Bay kit
The IBM x3690 X5 Single SATA HDD Bay kit includes the following components: 500 GB 7200 RPM 2.5-inch simple-swap SATA drive Simple-swap drive backplane and cable 4x4 drive bay filler panel Drive bay spacer filler
148
Follow these installation steps for the single SATA drive: 1. 2. 3. 4. Install the single SATA HDD bay assembly into the last bay for drive bays 12 - 15. Disconnect the optical drives cable from the planar connector. Plug the cable from the single SATA HDD bay into the connector on the planar. Install the SATA drive into drive bay 15.
The 2.5-inch drives require less space than the 3.5-inch drives, consume half the power, produce less noise, can seek faster, and offer increased reliability. Compatibility: As listed in Table 4-19 on page 147, the 2.5-inch 50 GB SSD is also supported with the standard SAS backplane and the option SAS backplane, part numbers 60Y0338 and 60Y0369. It is not compatible with the 1.8-inch SSD eXFlash backplane, 60Y0359. A typical configuration can be two 2.5-inch SAS disks for the operating system and two High IOPS disks for data. Only the 2.5-inch High IOPS SSD disk can be used on the SAS backplane. The 1.8-inch disks for the eXFlash cannot be used on the SAS backplane.
81Y4426
A10C
a. The battery is not included with the ServeRAID M5015 if ordered using the feature code, and it is not needed if using all SSD.
149
listed in Workload-optimized x3690 X5 models on page 125. You can add two more eXFlash backplanes to increase the supported number of SSDs to 24. The IBM eXFlash 8x 1.8-inch HS SAS SSD Backplane, part number 60Y0360, supports eight 1.8-inch High IOPS SSDs, as shown in Table 4-23. The eight drive bays require the same physical space as four SAS hard disk bays. A single eXFlash backplane requires two SAS x4 input cables and one custom power/configuration cable (shipped standard). Up to four SSD backplanes and 24 SSDs are supported in the x3690 X5 chassis. For more information regarding eXFlash and SSD information, including a brief overview of the benefits from using eXFlash, see 2.8, IBM eXFlash on page 47.
Table 4-23 x3690 X5 hard drive backplanes Part number 60Y0360 Feature code 9281 Backplane IBM eXFlash 8x 1.8 HS SAS SSD Backplane Drives supported Eight 1.8 solid-state drives SAS cables included 2 short, 2 long
Figure 4-23 shows the 8x 1.8-inch SSD backplane with its two SAS connectors.
The failure rate of SSDs is low because, in part, the drives have no moving parts. The 50 GB High IOPS SSD is a Single Level Cell (SLC) device with Enterprise Wear Leveling. As a consequence of both of these technologies, the additional layer of protection provided by a RAID controller might not always be necessary in every client environment. In certain cases, RAID-0 might even be an acceptable option. Table 4-25 on page 151 lists the controllers that support SSDs.
150
Table 4-25 Controllers supported with the eXFlash SSD backplane option Part number 46M0914 46M0829 46M0916 46M0969 81Y4426 Feature code 3876 0093 3877 3889 A10C Description IBM 6Gb SSD Host Bus Adapter (No RAID support) ServeRAID M5015 SAS/SATA Controllera ServeRAID M5014 SAS/SATA Controllera ServeRAID B5015 SSD IBM ServeRAID M5000 Performance Accelerator Key: Adds Cut Through I/O (CTIO) for SSD FastPath optimization on ServeRAID M5014, M5015, and M5025 controllers.
a. Add the Performance Accelerator Key to the ServeRAID M5015 or M5014 for use with SSDs.
Important: When ordering M5000 series controllers for use with SSD drives, you must not use the cache battery. If using M5000 series controllers in a mixed environment, order the cache battery along with Performance Accelerator Key. If you have already set up the ServeRAID controller that you plan to use and you want to leave the battery attached, you can still disable the write back cache by going into the MegaRAID web BIOS configuration utility and disabling Disk Cache and Default Write, as shown in Figure 4-24.
151
RAID support 0,1, and 1E 0,1,10,5, and 50b 0,1,10,5,50,6,60c 0,1,10,5,50,6,60c No 1 and 5
ServeRAID M1015 ServeRAID M5014 ServeRAID M5015 IBM 6Gb SSD HBA ServeRAID B5015 SSD
a. The BR10i is standard on most models. See 4.3, Models on page 124. b. M1015 support for RAID-5 and RAID-50 requires the M1000 Advanced Feature Key (46M0832, fc 9749). c. M5014 and M5015 support for RAID-6 and RAID-60 requires the M5000 Advanced Feature Key (46M0930, fc 5106). d. ServeRAID M5015 option part number 46M0829 includes the M5000 battery; however, the feature code 0093 does not contain the battery. Order feature code 5744 if you want to include the battery in the server configuration.
152
e. The IBM 6Gb SSD Host Bus Adapter is currently not available as a separately orderable option. Use the feature code to add the adapter to a customized order, using the CTO process. Part number 46M0914 is the L1 manufacturing part number. Part number 46M0983 is the pseudo option number, which is also used in manufacturing.
153
Performance optimization for SSDs LSI SafeStore: Support for self-encrypting drive services, such as instant secure erase and local key management (which requires the use of self-encrypting drives) Support for the optional M5000 Advanced Feature Key, which enables the following features: RAID levels 6 and 60 LSI SafeStore: Support for self-encrypting drive services, such as instant secure erase and local key management (which requires the use of self-encrypting drives) Performance Accelerator Key: Performance Accelerator Key uses the same features as the Advanced Feature Key but also includes performance enhancements to enable SSD support in a mixed HDD environment. For more information, see ServeRAID M5015 and M5014 SAS/SATA Controllers for IBM System x, TIPS0738, which is available at the following website: http://www.redbooks.ibm.com/abstracts/tips0738.html?Open
154
155
battery trays on top of the memory tray or the DIMM air baffle (whichever one is installed in the server). The battery trays are standard with the server. Each battery tray holds up to two batteries, to support a maximum of four RAID adapters with attached batteries in the x3690 X5. Table 4-27 lists the kit to order a remote battery cable.
Table 4-27 Remote battery cable ordering Option 44E8837 Feature code 5862 Description Remote Battery Cable Kit
The Remote Battery Cable kit, part number 44E8837, contains the following components: Remote battery cable Plastic interposer Plastic stand-off Two screws The screws and stand-off attach the interposer to the RAID controller after the battery is removed. Figure 4-25 shows these components.
The cable is routed through to the battery that is now installed in the RAID battery tray. This tray is either attached to the memory mezzanine if a memory mezzanine is installed, or the air baffle, which is in place of the mezzanine. Figure 4-26 shows how the battery trays are installed in the memory mezzanine. Each battery tray can hold two batteries.
RAID battery trays
Memory mezzanine
156
The card, which is shown in Figure 4-27, has two input connectors, which you connect to a supported RAID controller, plus four output connectors to go to each backplane, which allows you to connect to up to 16 drives. Important: You can use only the 2.5-inch hot-swap drive backplanes with this adapter (see Table 4-23).
You can use the Expansion Adapter only with the following SAS controllers: ServeRAID M1015 SAS/SATA adapter ServeRAID M5014 SAS/SATA adapter ServeRAID M5015 SAS/SATA adapter The Expansion Adapter must be installed in PCI Slot 1, and the ServeRAID adapter must be installed in PCI Slot 3.
157
Figure 4-28 x3690 with one IBM 4x 2.5-inch HS SAS HDD backplane
Figure 4-29 x3690 with two IBM 4x 2.5-inch HS SAS HDD backplanes
Figure 4-30 on page 159 shows a configuration that uses one 8x HDD backplane instead of two 4x HDD backplanes. Two SAS cables are needed.
158
Figure 4-30 x3690 X5 with one IBM 8x 2.5-inch HS SAS HDD backplane
Figure 4-31 illustrates the IBM eXFlash 8x SAS SSD backplane, which requires two SAS cables. With the eXFlash, eight drives can be used in the same space as four 2.5-inch drives.
Figure 4-31 x3690 with one IBM eXFlash 8x 1.8-inch HS SAS SSD backplane
Figure 4-32 x3690 with three IBM 4x 2.5-inch HS SAS HDD backplanes
Figure 4-33 shows one 8x and one 4x HDD backplane resulting in 12 drives. This configuration also requires three SAS cables.
Figure 4-33 x3690 with one 8x 2.5-inch HS SAS HDD and one 4x 2.5-inch HS SAS HDD backplane
Figure 4-34 on page 160 shows a mixture of 2.5-inch HDDs and 1.8-inch SSDs. This configuration requires three SAS cables.
159
Figure 4-34 x3690 with one 8x 2.5-inch backplane and one eXFlash 8x 1.8-inch SSD backplane
Figure 4-35 x3690 with four IBM 4x 2.5-inch HS SAS HDD backplanes
Figure 4-36 x3690 with two IBM 8x 2.5-inch HS SAS HDD backplanes
Figure 4-37 illustrates another 16-drive configuration with one 8x and two 4x backplanes. Also, you can configure this system with the two 4x backplanes for bays 0 - 7 and the 8x backplane for bays 8 - 15. Four SAS cables are required.
Figure 4-37 x3690 with one 8x 2.5-inch backplane and two 4x 2.5-inch backplanes
Figure 4-38 on page 161 shows two 4x backplanes and one eXFlash backplane. You can use one 8x backplane instead of the two 4x backplanes that are shown here. Four SAS cables are used in this configuration.
160
Figure 4-38 x3690 with two 4x 2.5-inch backplanes and one IBM eXFlash 8x 1.8-inch SSD backplane
Figure 4-39 shows two 8x eXFlash backplanes. Using these two backplanes requires four SAS cables. Figure 4-39 also shows the use of a single SATA drive.
Figure 4-39 x3690 X5 with two IBM eXFlash 8x 1.8-inch SSD backplanes and a single SATA drive
Figure 4-40 x3690 X5 with three 4x 2.5-inch backplanes and one IBM eXFlash 8x 1.8 SSD backplane
Figure 4-41 shows one 4x backplane and two 8x eXFlash backplanes. Five SAS cables are needed.
Figure 4-41 x3690 X5 with one 4x 2.5-inch backplane and two IBM eXFlash 8x 1.8 SSD backplanes
161
Figure 4-42 x3690 X5 with two 4x 2.5-inch backplanes and two IBM eXFlash 8x 1.8 SSD backplanes
Figure 4-43 shows the maximum number of 8x eXFlash backplanes supported in an x3690 X5. This configuration requires six SAS cables.
Figure 4-43 x3690 X5 with three IBM eXFlash 8x 1.8-inch SSD backplanes and a possible SATA drive
The M5025 has two external SAS 2.0 x4 connectors and supports the following features: Eight external 6 Gbps SAS 2.0 ports implemented through two four-lane (x4) connectors. Two mini-SAS external connectors (SFF-8088). 162
IBM eX5 Implementation Guide
6 Gbps throughput per SAS port. 800 MHz PowerPC processor with LSI SAS2108 6 Gbps RAID on Chip (ROC) controller. PCI Express 2.0 x8 host interface. 512 MB onboard data cache (DDR2 running at 800 MHz). Intelligent lithium polymer battery backup unit standard with up to 48 hours of data retention. Support for RAID levels 0, 1, 5, 10, and 50 (RAID 6 and 60 support with either M5000 Advanced Feature Key or M5000 Performance Key). Connections: Up to 240 SAS or SATA drives. Up to nine daisy-chained enclosures per port. SAS and SATA drives are supported, but the mixing of SAS and SATA in the same RAID array is not supported. Support for up to 64 logical volumes. Support for LUN sizes up to 64 TB. Configurable stripe size up to 1024 KB. Compliant with Disk Data Format (DDF) configuration on disk (COD). S.M.A.R.T. support. Support for the optional M5000 Advanced Feature Key, which enables the following features: RAID levels 6 and 60. LSI SafeStore: Support for self-encrypting drive services, such as instant secure erase and local key management (which requires the use of self-encrypting drives). Support for SSD drives in a mixed environment with SAS and SSD with the optional M5000 Series Performance Accelerator Key, which enables the following features: RAID levels 6 and 60. Performance optimization for SSDs. LSI SafeStore: Support for self-encrypting drive services, such as instant secure erase and local key management (which requires the use of self-encrypting drives). For more information, see ServeRAID M5025 SAS/SATA Controller for IBM System x, TIPS0739, which is available at the following website: http://www.redbooks.ibm.com/abstracts/tips0739.html?Open
163
DVD-ROM: The DVD-ROM drive uses the same connector on the system board as the single SATA drive; therefore, the DVD-ROM drive cannot be installed when using the SATA drive.
These slots are connected to the planar through two riser cards, both of which are installed as standard. Figure 4-45 on page 164 shows the locations of the two riser cards in the server.
Adapters
PCIe riser 1 with slots 1 and 2 PCIe riser 2 with slots 3, 4 and 5
164
4.10.1 Riser 1
In standard x3690 X5 models, riser slot 1 has the 2x8 riser card installed (60Y0329, feature 9285) which has the following slots: Slot 1, PCIe 2.0 x8 full height, full length slot Slot 2, PCIe 2.0 x8 full height, half length slot The 2x8 riser can be replaced by another riser with one PCIe 2.0 x16 slot, which is either a full-length slot or a 3/4-length slot, as listed in Table 4-31. This x16 slot is suitable for graphics processing unit (GPU) adapters. Additional power for the adapter is available from an onboard power connector if needed. Table 4-31 lists the riser card options for Riser 1. Only one of the risers listed in the table can be installed in the server at a time.
Table 4-31 x3690 X5 PCIe Riser 1 card options Part number 60Y0329 60Y0331 60Y0337 Feature code 9285 9282 9283 Riser card IBM System x3690 X5 PCI-Express (2x8) Riser Carda IBM System x3690 X5 PCI-Express (1x16) Riser Card - 3/4 length IBM System x3690 X5 PCI-Express (1x16) Riser Card - full lengthb
a. The 2x8 riser card is standard in all x3690 X5 models, including 7148-ARx. b. The 1x16 full-length riser cannot be used if the memory mezzanine is installed in the server.
4.10.2 Riser 2
Riser slot 2 has the 3x8 riser card installed in all standard models, except for model 7148-ARx (see 4.3, Models on page 124), and contains the following slots: Slot 3, PCIe 2.0 x8 low profile adapter. Slot 4, PCIe 2.0 x4 low profile adapter (x8 mechanical). Slot 5, PCIe 2.0 x8 low profile adapter. The Emulex 10Gb Ethernet adapter is installed in this slot if the adapter is part of the server configuration. Full-length adapters: Full-length adapters cannot be installed in any slots if the memory mezzanine is also installed. Instead, adapters up to 3/4 length are supported. Table 4-32 lists the option.
Table 4-32 x3690 X5 PCIe Riser 2 option Part number 60Y0366 Feature code 9280 Riser card option IBM System x3690 X5 PCI-Express (3x8) Riser Carda
a. The 3x8 riser card is standard in all x3690 X5 models, except 7148-ARx.
Note: The Emulex 10GbE Virtual Fabric Adapter that is standard in most models is installed in slot 5. See 4.10.3, Emulex 10Gb Ethernet Adapter on page 166 for details of the adapter.
165
Figure 4-46 The Emulex 10Gb Ethernet Adapter has a blue circuit board and a longer connector
The Emulex 10Gb Ethernet Adapter is a customer-replaceable unit (CRU). To replace the adapter (for example, under warranty), order the CRU number, as shown in Table 4-33. The table also shows the regular Emulex 10Gb Virtual Fabric Adapter (VFA) for IBM System x option, which differs only in the connector type (standard x8) and the color of the circuit board (green). Redundancy: The standard version of the Emulex VFA and the eX5 extended edge custom version can be used together as a redundant pair. This pair is a supported combination.
166
Table 4-33 Emulex adapter part numbers Option description Emulex 10Gb Ethernet Adapter for x3690 X5 Emulex 10Gb Virtual Fabric Adapter for IBM System x Part number None 49Y4250 Feature code 1648 5749 CRU number 49Y4202 Not applicable
General details about this card are in Emulex 10Gb Virtual Fabric Adapter for IBM System x, TIPS0762, which is available at the following website: http://www.redbooks.ibm.com/abstracts/tips0762.html Important: Although these cards are functionally identical, the availability of iSCSI and Fibre Channel over Ethernet (FCoE) upgrades for one card does not automatically mean availability for both cards. At the time of writing, the target availability of these features is the second quarter of 2011. Check availability of iSCSI and FCoE feature upgrades with your local IBM representative. The Emulex 10Gb Ethernet Adapter for x3690 X5 includes the following features: Dual-channel, 10 Gbps Ethernet controller Near line-rate 10 Gbps performance Two SFP+ empty cages to support either of the following items: SFP+ SR link is with SFP+ SR Module with LC connectors SFP+ twinaxial copper link is with SFP+ direct attached copper module/cable Note: Servers that include the Emulex 10Gb Ethernet Adapter do not include transceivers. You must order transceivers separately if needed, as listed in Table 4-34. TCP/IP stateless off-loads TCP chimney offload Based on Emulex OneConnect technology FCoE support as a future feature entitlement upgrade Hardware parity, CRC, ECC, and other advanced error checking PCI Express 2.0 x8 host interface Low-profile form-factor design IPv4/IPv6 TCP, User Datagram Protocol (UDP) checksum offload VLAN insertion and extraction Support for jumbo frames up to 9000 bytes Preboot eXecutive Environment (PXE) 2.0 network boot support Interrupt coalescing Load balancing and failover support Deployment and management of this adapter and other Emulex OneConnect-based adapters with OneCommand Manager Interoperable with BNT 10Gb Top of Rack (ToR) switch for FCoE functions Interoperable with Cisco Nexus 5000, Brocade 10Gb Ethernet switches for NIC/FCoE
167
SFP+ transceivers are not included with the server and must be ordered separately. Table 4-34 lists compatible transceivers.
Table 4-34 Transceiver ordering information Option number 49Y4218 49Y4216 46C3447 Feature code 0064 0069 5053 Description QLogic 10Gb SFP+ SR Optical Transceiver Brocade 10Gb SFP+ SR Optical Transceiver BNT SFP+ Transceiver
168
Description Brocade 8Gb FC Dual-port HBA Brocade 4Gb FC Single-port HBA QLogic 4Gb FC Dual-Port PCIe HBA Emulex 4GB FC Single-Port PCI-E HBA Emulex 4GB FC Dual-Port PCI-E HBA
1826
169
Serial Over LAN (SOL) Active Energy Manager Power/Reset control (power on, hard/soft shutdown, hard/soft reset, and schedule power control)
170
Light path diagnostics can alert the user to the following errors: Over current faults Fan faults Power supply failures PCI errors You can obtain the full details about the functions and operation of light path diagnostics in this system in the Installation and Users Guide - IBM System x3690 X5 at the following website: http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5085206
4.11.6 Cooling
The x3690 X5 has the following fans: Five, hot-swappable fans that are located in the front portion of the chassis Power supply internal fans that are located at the rear of each power supply Fans are numbered left to right, if you are looking at the front of the chassis. Fan 1 is nearest the power supplies, and Fan 5 is nearest the operator information panel. Figure 4-48 on page 172 shows the location of the fans. The individual fans are hot-swappable, as denoted by the orange release latches. The complete fan housing unit is not hot swappable. Fans 1 - 5 are accessible through an opening in the server top cover, the hot-swap fan access panel. You do not have to remove the server top cover to access the fans. Attention: If you release the cover latch and remove the server top cover while the server is running, the server is automatically powered off immediately. This powering off is required for electrical safety reasons.
171
Fan 5 Fan 1
Figure 4-49 shows the top of the server and the hot-swap fan access panel.
The following conditions affect system fan-speed adjustments: Inlet Ambient temperature CPU temperatures
172
a. Use 4782 for the power supplies, 9279 for the power supply interposer, and 6406 for the Y-cable.
The IBM 675W Redundant Power Supply Kit, option 60Y0327, includes the following items: Two 675W power supplies Two Y-cord power cables (2.8 m, 10A/200-250V, 2xC13 to IEC 320-C14) Two power cables (2.8 m, 10A/100-250V, C13 to IEC 320-C14) One power interposer card The Redundant Power Supply Kit includes a power supply interposer (power backplane). The interposer is a small circuit board that routes power from the power supply outputs to the system planar. Table 4-37 on page 174 lists the ac power input requirements.
173
Table 4-37 Power Supply ac input requirements Minimum Low range High range 90 V ac 180 V ac Maximum 137 V ac 265 V ac Nominal 100-127 V ac 50/60 Hz 200-240 V ac 50/60 Hz Maximum input current 7.8A RMS 3.8A RMS
A fan that is located inside each power supply cools the power modules. MAX5 has five redundant hot-swap fans, all in one cooling zone. The MAX5 fan speed is controlled by the IMM of the attached host, based on altitude and ambient temperature. Fans also respond to certain conditions and come up to speed accordingly: If a fan fails, the remaining fans spin up to full speed. As the internal temperature rises, all fans spin up to full speed.
174
Riser card 2 with slots 3, 4, and 5 (slots 3 and 4 are on the back)
PCIe slot 5
Figure 4-50 Low profile x8 riser card with hypervisor flash USB connector
The IBM USB Memory Key for virtualization is included in the virtualization-optimized models that are listed in 4.3, Models on page 124, but it can be added to any x3690 X5 system.
Table 4-39 USB key for embedded hypervisor Option 41Y8278 41Y8287 Feature code 1776 2420 Description IBM USB Memory Key for VMware ESXi 4 IBM USB Memory Key for VMware ESXi 4.1 with MAX5
For additional information and setup instructions for VMware ESXi software, see the VMware ESXi Embedded and vCenter Server Setup Guide that is available at the following website: http://www.vmware.com/pdf/vsphere4/r41/vsp_41_esxi_e_vc_setup_guide.pdf Also, before installing VMware, see the installation guide at 7.9.5, VMware vSphere ESXi 4.1 on page 358.
175
VMware ESXi 4.1 Memory and processor limits: Certain operating systems have upper limits to the amount of memory that is supported (for example, over 1 TB) or the number of processor cores that are supported (over 64 cores). See the ServerProven Plan for x3690 X5 for details and the full list of supported operating systems at the following website: http://ibm.com/systems/info/x86servers/serverproven/compat/us/nos/matrix.shtml
VMware ESX and two processors: If you plan to install VMware ESX or ESXi on the x3690 X5 with two installed processors, you must also install and populate the memory mezzanine. Failure to do so results in the following error: NUMA node 1 has no memory
Cable management arms (CMAs) are optional but useful because they help prevent cables from becoming tangled and causing server downtime. Table 4-41 lists the available cable management arms.
Table 4-41 Cable management arms Part number 69Y2347 69Y2344 69Y4390 Feature code 6473 6474 6458 Name IBM System x3690 X5 Cable Management Arm for Ball Bearing Slides IBM System x3690 X5 2U Cable Management Arm Friction CMA Use with rail kit 69Y2345 69Y4403 69Y4389
176
Chapter 5.
177
5.1 Introduction
The IBM BladeCenter HX5 supports up to two processors using Intel Xeon 6500 (Nehalem EX) 4-core, 6-core, or 8-core processors or four processors using the Intel Xeon 7500 Nehalem EX 4-core, 6-core, or 8-core processors. The HX5 supports up to 40 dual inline memory modules (DIMMs) with the addition of the MAX5 memory expansion blade when using Xeon 7500 Nehalem EX. Figure 5-1 shows the following three configurations: Single-wide HX5 with two processor sockets and 16 DIMM sockets. Double-wide HX5 with four Xeon 7500 series processors and 32 DIMM sockets. Double-wide HX5 with two Xeon 7500 or 6500 series processors and 40 DIMM sockets: 16 in the HX5 server and 24 in the attached MAX5 memory expansion blade. MAX5: MAX5 can only connect to a single HX5 server.
HX5 2-socket
HX5 4-socket
12 MB, 18 MB, or 24 MB (shared between cores) (processor-dependent) 978 or 800 MHz (processor SMI link speed dependent) 16 32 HX5: Up to 800 MHz MAX5: Up to 1066 MHz 40
178
Features Maximum RAM (using 8 GB DIMMs) Memory type DIMMs per channel Internal storage Maximum number of drives Maximum internal storage I/O expansion
DDR 3 error checking and correction (ECC) Very Low Profile (VLP) Registered DIMMs 1 1 HX5: 1; MAX5: 2
Optional 1.8-inch solid-state drives (SSDs); Non-hot-swap (require an additional SSD carrier) Two Up to 100 GB using two 50 GB SSDs One CIOv One CFFh Four Up to 200 GB using four 50 GB SSDs Two CIOv Two CFFh Two Up to 100 GB using two 50 GB SSDs One CIOv One CFFh
Figure 5-2 shows the components on the system board of the HX5. CIOv and CFFh Expansion slots Memory
Memory Buffers Two 30 mm nodes QPI wrap card connector Information panel LEDs
Figure 5-2 Layout of HX5 (showing a 2-node 4-socket configuration)
The MAX5 memory expansion blade, which is shown in Figure 5-3 on page 180, is a device with the same dimensions as the HX5. When the MAX5 is attached to the HX5, the combined
179
unit occupies two blade bays in the BladeCenter chassis. The MAX5 cannot be removed separately from the HX5.
Figure 5-3 MAX5 memory slots
MAX5 HX5
4, 6, or 8 cores
12 MB, 18 MB, or 24 MB (shared between cores) 978 or 800 MHz (scalable memory interconnects (SMI) link-speed dependent) HX5: Up to 978 MHz MAX5: Up to 1066MHz (SMI link-speed dependent)
180
Feature DIMMs per channel DIMM sockets Maximum installable RAM (8 GB DIMMs) Memory type
HS22 Two 12 96 GB
HX5 One 30 mm: 16 60 mm: 32 30 mm: 128 GB 60 mm: 256 GB DDR3 ECC VLP RDIMMs Two or four non-hot-swap 1.8 SSDs (require the SSD Expansion Card) Per 30mm blade: One CIOv One CFFh LSI 1064 controller on the optional SSD Expansion Card Internal USB socket for VMware ESXi Broadcom 5709S
DDR3 ECC VLP RDIMMs 2x Hot-swap 2.5 drive SAS, SATA, or SSD One CIOv One CFFh Onboard LSI 1064 Optional ServeRAID MR10ie (CIOv) Internal USB socket for VMware ESXi Broadcom 5709S
DDR3 ECC VLP RDIMMs Two non-hot-swap 1.8 SSDs (require the SSD Expansion Card) Per 60mm blade: One CIOv One CFFh LSI 1064 controller on the optional SSD Expansion Card Internal USB socket for VMware ESXi Broadcom 5709S
I/O expansion
One CIOv One CFFh Onboard LSI 1064 Optional ServeRAID MR10ie (CIOv) Internal USB socket for VMware ESXi Broadcom 5709S
serial-attached SCSI (SAS) controller Embedded Hypervisor Onboard Ethernet Chassis supported
181
a. One-node and 2-node HX5 configurations with 130W processors are not supported in chassis with standard cooling modules. See Table 5-4. b. Support for the BC-HT dc model can be granted for specific configurations with the SPORE process.
The number of HX5 servers supported in each chassis depends on the thermal design power of the processors that are used in the HX5 servers. Table 5-4, which uses the following conventions, shows the HX5 servers: A green square in a cell means the chassis can be filled with HX5 blade servers up to the maximum number of blade bays in the chassis (for example, 14 blades in the BladeCenter H). A yellow square in a cell means that the maximum number of HX5 blades that the chassis can hold is fewer than the total available blade bays (for example, 12 in a BladeCenter H). All other bays must remain empty. The empty bays must be distributed evenly between the two power domains of the chassis (bays 1 - 6 and bays 7 - 14).
Table 5-4 HX5 chassis compatibility Maximum number of servers supported in each chassis BC-S (8886) BC-H (models other than 4Tx) 2900W supplies Std. blower 5 4 2 2 2 2 14 Nonec 7 Nonec 7 6 Enh. blowerb 14 10 7 5 7 6 2980W suppliesa Std. blower 14 Nonec 7 Nonec 7 7 Enh. blowerb 14 12 7 6 7 7 BC-H (-4Tx) 2980W Enh. blowerb 14 12 7 6 7 7 10 8 5 4 5 5 BC-HT AC (8750)
Thermal design power (TDP) of the CPUs 95W, 105W 130W 95W, 105W
HX5 2-node (60 mm) 130W HX5 1-socket + MAX5 (60 mm) 95W, 105W 130W
a. IBM BladeCenter H 2980W AC Power Modules, 68Y6601 (standard in 4Tx, optional with all other BC-H chassis models) b. IBM BladeCenter H Enhanced Cooling Modules, 68Y6650 (standard in 4Tx, optional with all other BC-H chassis models) c. Not supported
182
Network Equipment Building System (NEBS): The HX5 is currently not a NEBS-compliant offering in the BC-HT.
5.4 Models
The base models of the BladeCenter HX5, with and without the MAX5 memory expansion blade, are shown in Table 5-5. In the table, Opt indicates optional and Std indicates standard.
Table 5-5 Models of the HX5 Modela Intel Xeon model and cores 1x E7520 4C 1x L7555 8C 1x E7530 6C 1x E7540 6C 1x E7540 6C 2x E6540 6C 2x E6540 6C 2x X6550 8C 2x X7560 8C 1x X7560 8C 1x X6550 8C 1x X7540 6C Clock speed 1.86 GHz 1.86 GHz 1.86 GHz 2.00 GHz 2.00 GHz 2.00 GHz 2.00 GHz 2.00 GHz 2.26 GHz 2.26 GHz 2.26 GHz 2.00 GHz TDP HX5 max memory speed 800 MHz 978 MHz 978 MHz 978 MHz 978 MHz 978 MHz 978 MHz 978 MHz 978 MHz 978 MHz 978 MHz 978 MHz MAX5 memory speed 800 MHz 978 MHz 978 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz 1066 MHz MAX5 Scale to four socket Yes Yes Yes Yes Yes No No No No Yes No Yes 10 GbE cardb Opt Opt Opt Opt Std Opt Std Opt Opt Std Std Std Standard memoryc 2x 4 GB 2x 4 GB 2x 4 GB 2x 4 GB 2x 4 GB HX5: 4x 4GB MAX5: None HX5: 4x 4GB MAX5: None HX5: 4x 4GB MAX5: None HX5: 4x 4GB MAX5: None 2x 4 GB 2x 4 GB 2x 4 GB
7872-42x 7872-82x 7872-61x 7872-64x 7872-65x 7872-63x 7872-6Dx 7872-83x 7872-84x 7872-86x 7872-E8x 7872-E6x
95W 95W 105W 105W 105W 105W 105W 130W 130W 130W 130W 105W
Opt Opt Opt Opt Opt Std Std Std Std Opt Opt Opt
a. This column lists worldwide, generally available variant (GAV) model numbers. They are not orderable as listed and must be modified by country. The US GAV model numbers use the following nomenclature: xxU. For example, the US orderable part number for 7870-A2x is 7870-A2U. See the product-specific official IBM announcement letter for other country-specific GAV model numbers. b. Emulex Virtual Fabric Adapter Expansion Card (CFFh). c. The HX5 has 16 DIMM sockets and can hold 128 GB using 8 GB memory DIMMs. The MAX5 has 24 DIMM sockets and can hold 192 GB using 8 GB memory DIMMs. A 1-node HX5 + MAX5 support 320 GB total using 8 GB DIMMs.
Also available is a virtualization workload-optimized model of the HX5. This model is a preconfigured, pretested model that is targeted at large-scale consolidation. Table 5-6 on page 184 shows the model.
183
Table 5-6 Workload-optimized models of the HX5 Model Intel Xeon model and cores/max Clock speed TDP HX5 max memory speeda MAX5 Scalable to four socket 10GbE cardb Standard memory (max 320 GB)c
Virtualization workload-optimized models (includes VMware ESXi 4.1 on a USB memory key) 7872-68x 2x E6540 6C/2 2.00 GHz 105 W 978 MHz Std No Std 160 GB HX5: 16x 4GB MAX5: 24x 4GB
a. Memory speed of the HX5 is dependent on the processor installed; however, the memory speed of the MAX5 is up to 1066 MHz irrespective of the processor installed in the attached HX5. b. Emulex Virtual Fabric Adapter Expansion Card (CFFh). c. HX5 has 16 DIMM sockets and can hold 128 GB using 8 GB memory DIMMs. MAX5 has 24 DIMM sockets and can hold 192 GB using 8 GB memory DIMMs. A 1-node HX5 + MAX5 support 320 GB total using 8 GB DIMMs.
Model 7872-68x is a virtualization-optimized model and includes the following features in addition to standard HX5 and MAX5 features: Forty DIMM sockets, all containing 4 GB memory DIMMs for a total of 160 GB of available memory. VMware ESXi 4.1 on a USB memory key is installed internally in the server. See 5.15, Integrated virtualization on page 214 for details. Emulex Virtual Fabric Adapter Expansion Card (CFFh).
Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer
Intel Xeon Processor 1 QPI x16 SMI links QPI links Intel I/O Hub ESI x4 x16 x4 CIOv I/O connector Scalability connector QPI Intel Xeon Processor 2 x8 CFFh I/O connector x8 SSD connector
QPI
Figure 5-5 on page 186 shows a block diagram of the Speed Burst Card attachment to the system. Speed Burst Card: The Speed Burst Card is not required for an HX5 with only one processor installed. It is also not needed for a 2-node configuration (a separate card is available for a 2-node configuration, as described in 5.8, Scalability on page 188).
185
Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer
SMI links
QPI links
Figure 5-6 shows where the Speed Burst Card is installed on the HX5.
186
Attachment to a single HX5 using the IBM HX5 MAX5 1-node Scalability kit, part number 59Y5877, as described in 5.8.3, HX5 with MAX5 on page 190 Communication with the processors on the HX5 using high-speed QPI links MAX5 is standard with certain models, as listed in 5.4, Models on page 183. For other models, MAX5 is available as an option, as listed in Table 5-8.
Table 5-8 IBM MAX5 for BladeCenter Part number 46M6973 59Y5877 Feature code 1740 1742 Description IBM MAX5 for BladeCenter IBM HX5 MAX5 1-node Scalability kit
MAX5 consists of the EX5 node controller chip, six memory buffers, and 24 DIMM sockets. The MAX5 has three power domains: A, B, and C. Each power domain includes two memory controllers and eight DIMM sockets. Figure 5-7 shows the layout of the MAX5.
Logical connectors to HX5 blades QPI QPI EXA EXA EXA
Power domain A
2 4 6 8 1 3 5 7
10
12 11
Memory buffer
Memory buffer
23 24 21 22
Power domain C
In the next section, 5.8, Scalability on page 188, we describe how the MAX5 connects to the HX5. We explain the memory options and rules in 5.10, Memory on page 194.
187
5.8 Scalability
This section explains how the HX5 can be expanded to increase the number of processors and the number of memory DIMMs. The HX5 blade architecture allows for a number of scalable configurations, including the use of a MAX5 memory expansion blade, but the blade currently supports three configurations: A single HX5 server with two processor sockets. This server is a standard 30 mm blade, which is also known as single-wide server or single-node server. Two HX5 servers connected to form a single image 4-socket server. This server is a 60 mm blade, which is also known as a double-wide server or 2-node server. A single HX5 server with two processor sockets, plus a MAX5 memory expansion blade attached to it, resulting in a 60 mm blade configuration. This configuration is sometimes referred to as a 1-node+MAX5 configuration. We describe each configuration in the following sections. We list the supported BladeCenter chassis for each configuration in 5.3, Chassis support on page 182.
188
Scalability card
Figure 5-8 2-node HX5 with the 2-node scalability card indicated
The double-wide configuration consists of two connected HX5 servers. This configuration consumes two blade slots and has the 2-node scalability card attached. The scaling is performed through QPI scaling. The 2-node scalability card is not included with the server and must be ordered separately, as listed in Table 5-9.
Table 5-9 HX5 2-Node Scalability Kit Part number 46M6975 Feature code 1737 Description IBM HX5 2-Node Scalability Kit
The IBM HX5 2-Node Scalability Kit contains the 2-node scalability card, plus the necessary hardware to physically attach the two HX5 servers to each other.
189
Figure 5-9 shows the block diagram of a 2-node HX5 and the location of the HX5 2-node scalability card.
HX5
HX5
16 DIMMs (1 DPC)
QPI links
QPI links
16 DIMMs (1 DPC)
SMI links
Ensure that all firmware is up-to-date before attaching the blades together. For the minimum firmware requirements, see 8.2.3, Required firmware of each blade and the AMM on page 379. For configuring software when using a 2-node for creating the partition in a scalable complex, using Flexnode to toggle between a single blade and a 2-node, and deleting the partition, see 8.6, Creating an HX5 scalable complex on page 402. Important: When a blade is attached, it does not automatically become a 2-node single-image system. You must create the 2-node single-image system by using the scalable complex in the Advanced Management Module (AMM) in the chassis. See 8.6, Creating an HX5 scalable complex on page 402 for more information.
190
The card that is used to connect the MAX5 to the HX5 is the IBM HX5 MAX5 1-Node Scalability Kit, which is extremely similar in physical appearance to the 2-Node Scalability Kit that was shown in Figure 5-8 on page 189. Table 5-10 lists the ordering information.
Table 5-10 HX5 1-Node Scalability Kit Part number 59Y5877 Feature code 1742 Description IBM HX5 MAX5 1-Node Scalability Kit
Figure 5-11 shows the block diagram of the single-node HX5 with MAX5. Important: The MAX5 can be connected only to a single HX5 server. A configuration of two MAX5 units connected to a 2-node HX5 is not supported.
HX5
MAX5
EXA 24 DIMMs (2 DPC)
SMI links
QPI
16 DIMMs (1 DPC)
QPI links
EXA
EXA
CPU 1
EXA
QPI
191
Having only one processor installed in the HX5 instead of two processors is supported; however, the recommendation is for two processors to maximize memory performance. Ensure that all firmware is up-to-date before attaching the MAX5 to the blade. For minimum firmware requirements, see 8.2.3, Required firmware of each blade and the AMM on page 379. When inserting an HX5 with MAX5, there is no partition information to set up in a scalable complex. MAX5 ships ready to use when attached.
a. The first feature code is for the first processor. The second feature code is for the second processor.
Table 5-12 lists the capabilities of each processor option that is available for the HX5.
Table 5-12 Intel Xeon 6500 and 7500 features Processor model/cores Scalable to four socket Processor frequency Turboa HTb L3 cache Power QPI speed HX5 memory speed MAX5 memory speed
Standard processors (E) Xeon E6540 6C Xeon E7520 4C Xeon E7530 6C Xeon E7540 6C No Yes Yes Yes 2.0 GHz 1.86 GHz 1.86 GHz 2.0 GHz Yes +2 No Yes +2 Yes +2 Yes Yes Yes Yes 18 MB 18 MB 12 MB 18 MB 105 W 95 W 105 W 105 W 6.4 GT/s 4.8 GT/s 5.86 GT/s 6.4 GT/s 978 MHz 800 MHz 978 MHz 978 MHz 1066 MHz 800 MHz 978 MHz 1066 MHz
192
Processor model/cores
Processor frequency
Turboa
HTb
L3 cache
Power
QPI speed
Low-power processors (L) Xeon L7555 8C Yes 1.86 GHz Yes +2 Yes 24 MB 95 W 5.86 GT/s 978 MHz 978 MHz
Advanced processors (X) Xeon X6550 8C Xeon X7542 6C Xeon X7550 8C Xeon X7560 8C No Yes Yes Yes 2.0 GHz 2.66 GHz 2.0 GHz 2.26 GHz Yes +3 Yes +3 Yes +3 Yes +3 Yes No Yes Yes 18 MB 18 MB 18 MB 24 MB 130 W 130 W 130 W 130 W 6.4 GT/s 5.86 GT/s 6.4 GT/s 6.4 GT/s 978 MHz 978 MHz 978 MHz 978 MHz 1066 MHz 978 MHz 1066 MHz 1066 MHz
a. Intel Turbo Boost technology. The number that is listed is the multiple of 133 MHz by which the processor base frequency can be increased. For example, if the base frequency is 2.0 GHz and the Turbo value is +2, the frequency can increase to as high as 2.266 GHz. b. Intel Hyper-Threading technology.
Xeon E6510: As shown in Table 5-12, the Xeon E6510 processor does not support scaling to four sockets, and it does not support the MAX5. This limit is a technical limitation of this particular processor. Follow these processor configuration rules: All installed processors must be identical. In a 2-node configuration, both two and four processors are supported. That is, each node can have one or two processors installed. All processors must be identical. If you only have two processors installed (two processors in a single-node, or one processor in each node in a 2-node), Xeon 6500 series processors are supported. However, you cannot add two additional processors later to form a 4-processor system, because the Xeon 6500 series processors do not support 4-way. A MAX5 configuration (HX5 server with MAX5 attached) supports one or two installed processors. Memory speed in the HX5 depends on the SMI link of the processor (the SMI link speed is listed in Table 5-12 on page 192 as a GT/s value). It also depends on the limits of the low-power scalable memory buffers that are used in the HX5: If the SMI link speed is 6.4 GT/s or 5.86 GT/s, the memory in the HX5 can operate at a maximum of 978 MHz. If the SMI link speed is 4.8 GT/s, the memory in the HX5 can operate at a maximum of 800 MHz. Table 5-12 on page 192 indicates these memory speeds. If a MAX5 memory expansion blade is installed, the memory in the MAX5 can operate as high as 1066 MHz, depending on the DIMMs installed. The MAX5 memory speed is independent of the HX5 memory speed. For more information about calculating memory speed, see 2.3.1, Memory speed on page 22.
193
5.10 Memory
The HX5 2-socket has eight DIMM sockets per processor (a total of 16 DIMM sockets), supporting up to 128 GB of memory when using 8 GB DIMMs. With the addition of the MAX5 memory expansion blade, a single HX5 blade has access to a total of 40 DIMM sockets supporting up to 320 GB of RAM when using 8 GB DIMMs. The HX5 and MAX5 use registered Double Data Rate 3 (DDR3), very low profile (VLP) DIMMs and provide reliability, availability, and serviceability (RAS) and advanced Chipkill memory protection. For more information about Chipkill memory protection, see Chipkill on page 29. For information about RAS, see 2.3.6, Reliability, availability, and serviceability (RAS) features on page 28. This section has the following topics: 5.10.1, Memory options on page 194 5.10.2, DIMM population order on page 196 5.10.3, Memory balance on page 199 5.10.4, Memory mirroring on page 200 5.10.5, Memory sparing on page 202 To see a full list of the supported memory features, such as Hemisphere Mode, Chipkill, nonuniform memory access (NUMA), and memory mirroring, and an explanation of each memory feature, see 2.3, Memory on page 22.
a. Although the speed of the supported memory DIMMs is as high as 1333 MHz, the actual memory bus speed is a function of the processor and the memory buffers used in the HX5 server. In the HX5, memory speed is up to 978 MHz and in the MAX5, memory speed is up to 1066 MHz. See Table 5-12 on page 192 for specifics. b. This DIMM supports redundant bit steering (RBS) when used in the MAX5, as described in Redundant bit steering on page 29.
Two GB DIMM option: The 2 GB DIMM option that is listed in Table 5-13 is not supported in the MAX5, because the MAX5 does not support mixing DIMMs with various DRAM technologies, such as 1 Gb versus 2 Gb.
194
For optimal performance, populate all DIMM slots on the HX5 before filling the MAX5. Each processor controls eight DIMMs and four memory buffers in the server, as shown in Figure 5-12. To make use of all 16 DIMM sockets, you must install both processors. If only one processor is installed, you can only install eight DIMM sockets.
DDR-3 DIMM 1 DDR-3 DIMM 2 DDR-3 DIMM 3 DDR-3 DIMM 4 DDR-3 DIMM 5 DDR-3 DIMM 6 DDR-3 DIMM 7 DDR-3 DIMM 8 Memory channels
Memory buffer
Scalability Connector
Memory buffer
DDR-3 DIMM 9 DDR-3 DIMM 10 DDR-3 DIMM 11 DDR-3 DIMM 12 DDR-3 DIMM 13 DDR-3 DIMM 14 DDR-3 DIMM 15 DDR-3 DIMM 16
Memory buffer Intel Xeon CPU 1 Memory buffer Memory buffer QPI Links Intel Xeon CPU 2
Memory buffer
SMI links
Figure 5-12 Portion of the HX5 block diagram showing the processors, memory buffers, and DIMMs
Figure 5-13 shows the physical locations of the 16 memory DIMM sockets.
Scalability connector Control panel connector DIMM 5 DIMM 6 DIMM 7 DIMM 8
TIGHTEN SCREWS ALTERNATELY
Microprocessor 1
Microprocessor 2
DIMM 9 DIMM 10 DIMM 11 DIMM 12 Battery DIMM 14 DIMM 16 DIMM 15 DIMM 13 Blade expansion connector
The MAX5 memory expansion blade has 24 memory DIMM sockets, as shown in Figure 5-14 on page 196. The MAX5, which must be connected to an HX5 system (only the 1-node HX5 supports the MAX5), has one memory controller and six SMI-connected memory buffers.
195
MAX5 memory runs at 1066, 978, or 800 MHz DDR3 speeds. The memory speed is dependent on the processor QPI speed in the HX5: A QPI speed of 6.4 GHz means the speed of the MAX5 memory is 1066 MHz. A QPI speed of 5.8 GHz means the speed of the MAX5 memory is 978 MHz. A QPI speed of 4.8 GHz means the speed of the MAX5 memory is 800 MHz. Table 5-12 on page 192 indicates these memory speeds for each processor. To see more information about how memory speed is calculated with QPI, see 2.3.1, Memory speed on page 22.
196
Table 5-14 NUMA-compliant DIMM installation for a single-node HX5 Hemisphere Modea Number of DIMMs Processor 1 Buffer Buffer Buffer Buffer Buffer DIMM 10 Processor 2 Buffer DIMM 11 DIMM 12 Buffer DIMM 13 DIMM 14 Buffer DIMM 15 x DIMM 15 x DIMM 16 x x x x Number of CPUs
DIMM 1
DIMM 2
DIMM 3
DIMM 4
DIMM 5
DIMM 6
DIMM 7
DIMM 8
2 2 2 2
4 8 12 16
N Y N Y
x x x x x x x x
x x x x x x x x x x x x
DIMM 9 x x x x
x x x x x x x x x x x
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26.
In a 2-node (4-socket) configuration with two HX5 servers, follow the memory installation sequence in both nodes. You must populate memory to have a balance for each processor in the configuration. For best performance, use the following general guidelines: Install as many DIMMs as possible. You can get the best performance by installing DIMMs in every socket. Each processor needs to have identical amounts of RAM. Spread out the memory DIMMs to all memory buffers. That is, install one DIMM to a memory buffer before beginning to install a second DIMM to that same buffer. See Table 5-14 for DIMM placement. You must install memory DIMMs in the order of the DIMM size with largest DIMMs first, then next largest DIMMs, and so on. Placement must follow the DIMM socket installation that is shown in Table 5-14. To maximize performance of the memory subsystem, select a processor with the highest memory bus speed (as listed in Table 5-12 on page 192). The lower value of the processors memory bus speed and the DIMM speed determine how fast the memory bus can operate. Every memory bus operates at this speed.
Table 5-15 NUMA-compliant DIMM installation for a 2-node HX5 Hemisphere Modea Number of DIMMs Processor 1 Buffer Buffer Buffer Buffer Buffer DIMM 10 Processor 2 Buffer DIMM 11 DIMM 12 Buffer DIMM 13 DIMM 14 Buffer DIMM 16 x x x x Number of CPUs
DIMM 1
DIMM 2
DIMM 3
DIMM 4
DIMM 5
DIMM 6
DIMM 7
DIMM 8
4 4 4 4
8 16 24 32
N Y N Y
x x x x x x x x
x x x x x x x x x x x x
DIMM 9 x x x x
x x x x x x x x x x x
197
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26
Based on the two DIMM options that are currently supported in the MAX5 (listed in Table 5-13 on page 194), this step means that all DIMMs in each of the quads listed here must be either 4 GB or 8 GB. You cannot mix 4 GB and 8 GB DIMMs in the same quad. Memory must be installed in matched pairs of DIMMs in the MAX5. Memory DIMMs must be installed in the order of DIMM size with largest DIMMs first. For example, if you plan to install both 4 GB and 8 GB DIMMs into the MAX5, use the population order that is listed in Table 5-16 on page 199. Install all 8 GB DIMMs first, and then install the 4 GB DIMMs. The DIMM sockets in the MAX5 are arranged in three power domains (A, B, and C), as shown in Figure 5-15. Each power domain includes two memory controllers and eight DIMM sockets.
Logical connectors to HX5 blades QPI QPI EXA EXA EXA
Power domain A
2 4 6 8 1 3 5 7
10
12 11
Memory buffer
Memory buffer
23 24 21 22
Power domain C
198
This list shows the correct DIMMs in each power domain: Power domain A: 1 - 4 and 5 - 8 Power domain B: 13 - 16 and 17 - 20 Power domain C: 9 - 12 and 21 - 24 For the best memory performance, install the DIMMs by spreading them among all six memory buffers and all three power domains. Table 5-16 shows the installation order.
Table 5-16 DIMM installation for the MAX5 for IBM BladeCenter Number of DIMMs Power domain A Buffer DIMM 1 DIMM 2 DIMM 3 DIMM 4 DIMM 5 Buffer DIMM 6 DIMM 7 DIMM 8 DIMM 9 Domain C () Buffer DIMM 10 DIMM 11 DIMM 12 DIMM 13 Power domain B Buffer DIMM 14 DIMM 15 DIMM 16 DIMM 17 Buffer DIMM 18 DIMM 19 DIMM 20 DIMM 21 Domain C () Buffer DIMM 22 DIMM 23 x x x x DIMM 24 x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
2 4 6 8 10 12 14 16 18 20 22 24
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
MAX5 and VMware ESX: When using a MAX5 with VMware ESX 4.1 or ESXi 4.1, a boot parameter is required to access the MAX5 memory expansion unit. The MAX5 memory expansion unit utilizes NUMA technology, which needs to be enabled within the operating system. Without enabling NUMA technology, you might see the following message: The system has found a problem on your machine and cannot continue. Interleaved Non-Uniform Memory Access (NUMA) nodes are not supported. See the RETAIN tip H197190 for more information and the necessary parameters: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5084842
199
fashion to provide a balanced system and populating all processors identically is also required by VMware. Looking at Figure 5-16 as an example, Processor 0 has DIMMs populated, but no DIMMs are populated that are connected to Processor 1. In this case, Processor 0 has access to low-latency local memory and high-memory bandwidth. However, Processor 1 has access only to remote or far memory. So, threads executing on Processor 1 have a longer latency to access memory as compared to threads on Processor 0. This result is due to the latency penalty incurred to traverse the QPI links to access the data on the other processors memory controller. The bandwidth to remote memory is also limited by the capability of the QPI links. The latency to access remote memory is more than 50% higher than local memory access. For these reasons, we advise that you populate all of the processors with memory, remembering the necessary requirements to ensure optimal interleaving and Hemisphere Mode.
LOCAL
REMOTE
QPI links
Memory controller
Memory controller
Buffer DIMM
Buffer
Buffer
DIMM DIMM
Figure 5-16 Memory latency when not spreading DIMMs across both processors
200
Table 5-17 DIMM installation for memory mirroring: One processor Processor 1 Number of processors Buffer Number of DIMMs DIMM 1 DIMM 2 Buffer DIMM 3 DIMM 4 Buffer DIMM 5 DIMM 6 Buffer DIMM 7 DIMM 8 Buffer DIMM 10 DIMM 9 Processor 2 Buffer DIMM 11 DIMM 12 Buffer DIMM 13 DIMM 14 Buffer DIMM 15 DIMM 15 x Buffer DIMM 20 DIMM 21 DIMM 22 DIMM 23 DIMM 24 DIMM 16
Table 5-18 lists the DIMM installation sequence for memory-mirroring mode when two processors are installed.
Table 5-18 DIMM installation for memory mirroring: Two processors Processor 1 Number of processors Buffer Number of DIMMs DIMM 1 DIMM 2 Buffer DIMM 3 DIMM 4 Buffer DIMM 5 DIMM 6 Buffer DIMM 7 DIMM 8 Buffer DIMM 10 DIMM 9 Processor 2 Buffer DIMM 11 DIMM 12 Buffer DIMM 13 DIMM 14 Buffer DIMM 16 x
1 1
8 16
x x
x x
x x
x x
x x
x x
x x
x x x x x x x x
4 8 12 16
x x x x x x x x x x x x x x x x
x x x x
x x x x x x x x x x
x x x x x x x x x x
Domain C must be empty: Memory mirroring is only supported using two domains. You must remove all DIMMs from Domain C. If there is memory in Domain C, you get the following error in the AMM logs and all memory in the MAX5 is disabled: Group 1, (memory device 1-40) (All DIMMs) memory configuration error
201
1 1
4 8
x x
x x
x x
x x x x x x
Table 5-21 shows the installation order when two processors are installed.
Table 5-21 DIMM installation for the HX5 memory sparing: Two processors Processor 1 Buffer Number of processors Number of DIMMs DIMM 1 DIMM 2 Buffer DIMM 3 DIMM 4 Buffer DIMM 5 DIMM 6 Buffer DIMM 7 DIMM 8 Buffer DIMM 10 DIMM 9 Processor 2 Buffer DIMM 11 DIMM 12 Buffer DIMM 13 DIMM 14 x Buffer
2 2 2 2
4 8 12 16
x x x x
x x x x
x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x
Redundant bit steering: Redundant bit steering (RBS) is not supported on the HX5 because the integrated memory controller of the Intel Xeon 7500 processors does not support the feature. See Redundant bit steering on page 29 for details. The MAX5 memory expansion blade supports RBS, but only with x4 memory and not x8 memory. As shown in Table 5-13 on page 194, the 8 GB DIMM, part number 49Y1554, uses x4 DRAM technology. RBS is automatically enabled in the MAX5 memory port, if all DIMMs installed to that memory port are x4 DIMMs.
202
Sparing
62
Mirroring
50
Normal
100
20
40
60
80
100
120
5.11 Storage
The storage system on the HX5 blade is based on the use of the SSD Expansion Card for IBM BladeCenter HX5, which contains an optional LSI 1064E SAS Controller and two 1.8-inch micro SATA drive connectors. The SSD Expansion Card allows the attachment of two 1.8-inch solid-state drives (SSDs). If two SSDs are installed, the HX5 supports RAID-0 or RAID-1 capability. Installation of the SSDs in the HX5 requires the SSD Expansion Card for IBM BladeCenter HX5. Only one SSD Expansion Card is needed for either one or two SSDs. Table 5-22 lists the ordering details.
Table 5-22 SSD Expansion Card for IBM BladeCenter HX5 Part number 46M6908 Feature code 5765 Description SSD Expansion Card for IBM BladeCenter HX5
203
Top side of the SSD Expansion Card with one SSD installed in drive bay 0
Underside of the SSD Expansion Card showing PCIe connector and LSI 1064E controller
Figure 5-18 SSD Expansion Card for the HX5 (left: top view; right: underside view)
The SSD Expansion Card can be installed in the HX5 in combination with a CIOv I/O expansion card and CFFh I/O expansion card, as shown in Figure 5-19.
Figure 5-19 Placement of SSD expansion card in combination with a CIOv card and CFFh card
ServeRAID MR10ie support: The HX5 does not currently support the ServeRAID MR10ie RAID controller.
204
Table 5-23 Supported SSDs Part number 43W7734 Feature code 5314 Description IBM 50GB SATA 1.8-inch NHS SSD
For more information about SSD drives and their advantages, see 2.8.1, IBM eXFlash price-performance on page 49.
In a 2-node configuration, each controller operates independently, and each controller maintains its own configuration for the sets of drives that are installed in that node. One controller cannot cross over to the other node to perform a more complex RAID solution. Using independent controllers allows for several configuration options. Each LSI 1064 controller has an option of RAID-1, RAID-0, or JBOD (just a bunch of disks). No redundancy exists in a JBOD configuration, and each drive runs independently. The blade uses JBOD, by default, if no RAID array is configured. Figure 5-21 on page 206 shows the three options in the LSI 1064 setup page, LSI Logic MPT Setup Utility. Only two options are supported with this blade, because there are only two drives maximum that are installable in the HX5 and an Integrated Mirroring Enhanced (IME) volume requires three drives minimum.
205
The following options are presented: Create IM Volume: Creates a RAID-1 array RAID 1 drives are mirrored on a 1 to 1 ratio. If one drive fails, the other drive takes over automatically and keeps the system running. However, in this configuration, you lose 50% of your disk space with one of the drives being used as a mirrored image. The stripe size is 64 Kb and cannot be altered. This option also affects the performance on the drives, because all data has to be written twice, once per drive. See the performance chart in Figure 5-22 on page 207 for details. Create IME Volume: Creates a RAID-1E array This option requires three drives, so it is not available in the HX5. Create IS Volume: Creates a RAID-0 array RAID 0 or the Integrated Striping (IS) volume, as shown in LSI, is one of the faster performing disk arrays because read and write sectors of data are interleaved between multiple drives. The downside to this configuration is immediate failure if one drive fails. There is no redundancy. In a RAID-0, you also keep the full size of both drives. Identical size drives are recommended for performance, as well as data storage efficiency. The stripe size is 64 Kb and cannot be altered. We provide the instructions to create an array in 8.4, Local storage considerations and array setup on page 385.
206
Tests: These tests are not certified tests. We performed these tests to illustrate the performance differences among the three configuration options. We ran these tests on an HX5 2-node system running four 2 GHz Intel Xeon 7500 series 6-core processors and 16 GB of memory. The SSDs were IBM 50 GB SATA 1.8-inch NHS SSDs. Results might vary depending on the size and type of installed drives. The results show that when choosing a configuration for setting up hard drives, the performance difference between a RAID-1, RAID-0, and JBOD configuration is minimal. JBOD was always the fastest performer, followed by RAID-1, and then RAID-0.
207
SAS 1
PCIe x4
DS3512
Figure 5-24 Connecting SAS Connectivity Card and external SAS solution
208
The expansion blade has the following features: Support for PCIe 2.0 adapters in an expansion blade The expansion blade lets you install one or two standard form factor PCIe 2.0 adapter cards in a BladeCenter environment, enabling the use of specialized adapters or adapters that otherwise are not available to BladeCenter clients. Each of the two adapters can consume up to 75 W. Ability to stack up to four expansion blades on a single base blade You can attach up to two, three, or four expansion blades (depending on the attached server), therefore, maintaining the BladeCenter density advantage while still giving you the option to install PCIe cards as needed without the need for each expansion blade having to be attached to a server and the added complexity and cost that brings. The first expansion blade connects to the server blade using the CFFh expansion slot of the server blade. The second expansion blade attaches to the CFFh connector on the first expansion blade, and so on. The following maximums are for the attached expansion blades: Single-node HX5: Up to three expansion blades Two-node HX5: Up to two expansion blades MAX5 support: The HX5 with an attached MAX5 does not also support the attachment of the PCI Express Gen 2 Expansion Blade. CFFh slot still available The CFFh expansion connector is accessible on the topmost expansion blade, even with four expansion blades attached. This design lets you maintain the integrated networking capabilities of the blade server when it is installed in a BladeCenter S, H, or HT chassis. For details about the supported PCI Express adapter cards, see the IBM Redbooks at-a-glance guide, IBM BladeCenter PCI Express Gen 2 Expansion Blade, TIPS0783, which is available at this website: http://www.ibm.com/redbooks/abstracts/tips0783.html
5.13.1 CIOv
The CIOv I/O expansion connector provides I/O connections through the midplane of the chassis to modules located in bays 3 and 4 of a supported BladeCenter chassis. The CIOv slot is a second-generation PCI Express 2.0 x8 slot. A maximum of one CIOv I/O expansion card is supported per HX5. A CIOv I/O expansion card can be installed on a blade server at the same time that a CFFh I/O expansion card is installed in the blade.
Chapter 5. IBM BladeCenter HX5
209
Table 5-25 lists the CIOv expansion cards that are supported in the HX5.
Table 5-25 Supported CIOv expansion cards Part number 44X1945 46M6065 46M6140 43W4068 44W4475 Feature code 1462 3594 3598 5093 5477 Description QLogic 8Gb Fibre Channel Expansion Card (CIOv) QLogic 4Gb Fibre Channel Expansion Card (CIOv) Emulex 8Gb Fibre Channel Expansion Card (CIOv) SAS Connectivity Card (CIOv) Ethernet Expansion Card (CIOv)
See the IBM ServerProven compatibility website for the latest information about the expansion cards that are supported by the HX5: http://ibm.com/servers/eserver/serverproven/compat/us/ CIOv expansion cards are installed in the CIOv slot in the HX5 2-socket, as shown in Figure 5-25.
CIOv card
Figure 5-25 The HX5 type 7872 showing the CIOv I/O expansion card position
5.13.2 CFFh
The CFFh I/O expansion connector provides I/O connections to high-speed switch modules that are located in bays 7, 8, 9, and 10 of a BladeCenter H or BladeCenter HT chassis, or to switch bay 2 in a BladeCenter S chassis. The CFFh slot is a second-generation PCI Express x16 (PCIe 2.0 x16) slot. A maximum of one CFFh I/O expansion card is supported per blade server. A CFFh I/O expansion card can be installed on a blade server at the same time that a CIOv I/O expansion card is installed in the server.
210
See the IBM ServerProven compatibility website for the latest information about the expansion cards that are supported by the HX5: http://ibm.com/servers/eserver/serverproven/compat/us/ CFFh expansion cards are installed in the CFFh slot in the HX5, as shown in Figure 5-26.
CFFh card
Figure 5-26 The HX5 type 7872 showing the CFFh I/O expansion card position
A CFFh I/O expansion card requires that a supported high-speed I/O module or a Multi-switch Interconnect Module is installed in bay 7, 8, 9, or 10 of the BladeCenter H or BladeCenter HT chassis.
Chapter 5. IBM BladeCenter HX5
211
In a BladeCenter S chassis, the CFFh I/O expansion card requires a supported switch module in bay 2. When used in a BladeCenter S chassis, a maximum of two ports are routed from the CFFh I/O expansion card to the switch module in bay 2. See the IBM BladeCenter Interoperability Guide for the latest information about the switch modules that are supported with each CFFh I/O expansion card at the following website: http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073016
5.14.1 UEFI
The HX5 2-socket uses an integrated Unified Extensible Firmware Interface (UEFI) next-generation BIOS. The UEFI provides the following capabilities: Human readable event logs; no more beep codes Complete setup solution by allowing adapter configuration function to be moved to UEFI Complete out-of-band coverage by Advanced Settings Utility to simplify remote setup Using all of the features of UEFI requires an UEFI-aware operating system and adapters. UEFI is fully backward-compatible with BIOS. For more information about UEFI, see the IBM white paper, Introducing UEFI-Compliant Firmware on IBM System x and BladeCenter Servers, which is available at the following website: http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5083207 For the UEFI menu setup, see 8.5, UEFI settings on page 396.
212
213
As shown in Figure 5-27, the flash drive plugs into the Hypervisor Interposer, which, in turn, attaches to the system board near the processors. The Hypervisor Interposer is included as standard with the HX5. Hypervisor Interposer
See 5.17, Operating system support on page 215 for details about VMware ESX and other operating system support.
214
For information about setup options and common errors, see 8.6, Creating an HX5 scalable complex on page 402.
215
SUSE Linux Enterprise Server 11 with Xen for AMD64/EM64T VMware ESX 4.0 Update 1 VMware ESXi 4.0 Update 1 VMware ESX 4.1 VMware ESXi 4.1 Key information regarding VMware ESX: ESXi 4.0 support is single-node HX5 only. ESX 4.0 supports single-node and 2-node HX5. ESXi 4.1 and ESX 4.1 are required for MAX5 support. See the ServerProven website for the most recent information: http://ibm.com/systems/info/x86servers/serverproven/compat/us/nos/ematrix.shtml
216
Part 2
Part
Implementing scalability
In this second part of the book, we provide detailed configuration and setup information to get your servers operational. We particularly focus on setting up MAX5 configurations of all three eX5 servers, as well as 2-node configurations of the x3850 X5 and HX5. This part consists of the following chapters: Chapter 6, IBM System x3850 X5 and x3950 X5 on page 219 Chapter 7, IBM System x3690 X5 on page 301 Chapter 8, IBM BladeCenter HX5 on page 373 Chapter 9, Management on page 447
217
218
Chapter 6.
219
6.1 Before you apply power for the first time after shipping
Before you begin populating your server with all of its processors, memory, and PCI adapters and before you install an operating system, perform the recommendations described in the following sections.
220
CMOS battery
3. Use your finger to pry up the battery on the side closest to the neighboring IC chip. The battery easily lifts out of the socket. Tip: The light path diagnostics (LPD) lights are powered from a separate power source (a capacitor) than the CMOS memory. LPD lights remain lit for a period of time after ac power and the CMOS memory battery have been removed. 4. After 30 seconds, insert one edge of the battery, with the positive side up, back into the holder. 5. Push the battery back into the socket with your finger and it clips back into place, as shown in Figure 6-1.
6.1.3 Verify that the server completes POST before adding options
When you have ordered options for your server that have not yet been installed, it is a good idea to ensure that the server completes the POST properly before you start to add your options. Performing this task makes it easier to compartmentalize a potential problem with an installed option rather than having to look at the entire server to try to identify a good starting point during problem determination.
221
same server. Any operational differences in stepping levels are handled by the microcode of the processor, Integrated Management Module (IMM), and UEFI.
Only processor 1 is installed as a minimum configuration for testing Regardless of the amount of memory installed. Memory installed only on memory cards 1 and 2. Minimum of 2 DIMMs on memory card 1. Memory installed on memory cards other than memory cards 1 and 2. Memory installed on both the system board and the MAX5. PCIe slots 1 - 4 are not functional. Performance improves as DIMMs are added to evenly populate all ranks on each memory card. None of the memory installed on other memory cards is accessible by processor 1. Performance improves significantly when more active memory calls use local memory on the system board.
Both processors 1 and 4 are installed Memory installed only on memory cards 1 and 2. Minimum of two DIMMs on memory card 1. Performance improves as DIMMs are added to evenly populate all ranks on each memory controller. Processing threads assigned to processor 4 always have a significant drop in performance for memory-intensive tasks. This configuration is not an operational configuration for operating systems, such as VMware.
222
Operational characteristic Performance improves as DIMMs are added to evenly populate all ranks on each memory card. All memory on memory cards 7 and 8 is not visible if processor 4 fails. The operational configuration for VMware requires that the memory cards of the installed processor have the exact same total memory installed. For ease of maintenance, the best practice is to have an identical memory configuration on each of the installed processors memory cards. Performance improves as DIMMs are added to evenly populate all ranks on each memory controller for local memory. Performance improves significantly when more active memory calls use local memory on the system board instead of the memory that is located in the MAX5. In the rare instance of a memory card failure, VMware operational requirements can be satisfied if all of the memory is installed in the MAX5 or each processor has the same amount of memory disabled. Memory access is significantly slower if all of the memory is in the MAX5.
Processors 2 and 3 are added to the existing processors 1 and 4 Memory is installed only on memory cards 1, 2, 7, and 8. Minimum of two DIMMs on memory card 1. Performance improves as DIMMs are added to evenly populate all ranks on each memory controller. Processing threads assigned to processors without memory cards have a significant drop in performance for memory-intensive tasks. This configuration is not an operational configuration for operating systems, such as VMware. Performance improves as DIMMs are added to evenly populate all ranks on each memory card. Memory that is installed on memory cards that are associated with a failed or uninstalled processor is not seen by the other processors. The operational configuration for VMware requires the memory cards of the installed processors have the exact same total memory installed. For ease of maintenance, the best practice is to have an identical memory configuration on each of the installed processors memory cards. Memory latency time increases by 50% for memory access calls between processors 1 and 2 when the QPI wrap card when the QPI wrap card that is installed across QPI slot 1 and 2 has failed or is not installed. The same effect occurs with memory access calls between processors 3 and 4 when the QPI wrap card installed across QPI slot 3 and 4 has failed or is not installed. Failure of a QPI wrap card will be represented in the hardware event log as a loss of a redundant QPI lane for two of the processors.
223
Memory Installed Memory installed on all memory cards for installed processors and the MAX5.
Operational characteristic QPI wrap cards must be removed from the server to accept QPI cables that are used to link the MAX5 memory expansion unit to the server. Performance improves as DIMMs are added to evenly populate all ranks on each memory controller for local memory. Performance improves significantly when more active memory calls use local memory on the system board instead of the memory that is located in the MAX5. In the rare instance of a memory card failure, VMware operational requirements can be satisfied if all of the memory is installed in the MAX5 or each processor has the same amount of memory disabled. Memory access is significantly slower if all of the memory is in the MAX5.
224
Pin 1 indicator Tool lock position indicator Tool unlock position indicator
To use the tool, perform the following steps: 1. When the processor is not installed in the server, ensure that you place the processor flat on an antistatic bag or mat, with the contact pads facing down. 2. Rotate the blue handle at the top of the tool so that the tool is in the unlocked position. 3. Place the tool over the top of the processor with the pin 1 indicator of the processor lined up with the pin 1 indicator on the tool. 4. Rotate the blue handle to the locked position to securely hold the processor inside the processor removal tool. 5. When the processor is being removed from the processor socket, lift the processor straight up out of the processor socket. 6. When placing the processor into the processor socket, ensure the processor socket has its protective cover removed and is open to receive the processor. Line up the pin 1 indicator on the tool with the pin 1 indicator on the processor socket and carefully lower the processor into the socket. The two holes on the processor removal tool need to line up with the two bolt heads that secure the processor socket to the system board. 7. When you have placed the processor in the desired location, rotate the blue handle back to the open position to release the tool from the placed processor. Carefully lift the tool from the processor.
225
processing nodes in a high-performance cluster, database servers, or print servers for graphics printers require the best memory performance possible. Remember these considerations when installing memory in the System x3850 X5: Due to the fact that no single processor has direct access to all PCIe slots, the QPI links that are used to communicate between the two processors are not only used for memory calls, but also for interfacing to PCIe adapters. When installing memory for multiple processors and not all memory cards are installed, only processing threads that are assigned to processor without local memory experience a 50% increase in memory latency. For a server with heavy I/O processing, the additional memory traffic also affects the efficiency of addressing PCIe adapters. Any memory that is installed on memory cards for an uninstalled or defective processor is not seen by the other processors on the server. For nonuniform memory access (NUMA)-aware operating systems when multiple processors are installed, you must install the same amount of memory on each memory card of each installed processor. See 3.8.2, DIMM population sequence on page 79 for details. You can achieve the best processor performance when memory is installed to support
Hemisphere Mode. Hemisphere Mode is the required memory configuration to allow two
3850 X5 servers to scale. It is possible to have processor 1 and 4 in Hemisphere Mode, and not processor 2 and 3, to permit scaling. In this type of installation, having processor 1 and 4 in Hemisphere Mode improves the memory access latency for all processors. To determine the DIMM population for Hemisphere Mode, see 3.8.2, DIMM population sequence on page 79. You might consider installing all DIMMs of the same field-replaceable unit (FRU) to avoid conflicts that might prevent NUMA compliance or Hemisphere Mode support. These problems are most likely to occur when multiple DIMM sizes are installed. The server supports installing multiple-sized DIMMs in each rank, but the configuration becomes complex and difficult to maintain. Operating systems that depend on NUMA compliance inform you when the server is not NUMA-compliant. However, nothing informs you that the processors are not in Hemisphere Mode. It is better to install more smaller DIMMs than fewer larger DIMMs to ensure that all of the memory channels and buffers of each processor have access to the same amount of memory. This approach allows the processors to fully use their interleave algorithms to access memory faster by spreading access over multiple paths.
226
The DIMM configuration is invalid. See the DIMM placement tables in x3690 X5 memory population order on page 133 for details. A nonfunctional DIMM, failed DIMM slot, bent processor pin, or resource conflict with a PCIe adapter. Swap the DIMM with a functional DIMM, reactivate the DIMM by pressing F1-Setup, and retest: When the problem follows the DIMM, replace the DIMM. When the problem stays with the memory slot location, remove any non-IBM PCIe adapters, reactivate, and retest. When the failure is in the same slot on the memory board, verify that the memory board is completely seated or swap memory boards with a known functional memory board location, reactivate, and retest. When the problem remains with the memory board location, a bent processor pin might be the cause. Contact IBM support for replacement parts. The firmware of the server contains a a powerful diagnostic tool. Use the following steps to perform a simple test of a new memory configuration before placing the server into production: 1. During POST, at the IBM System x splash panel, press F2-Diagnostics, as shown in Figure 6-3.
2. When the built-in diagnostics start, they start the Quick Memory Test for all of the memory in the server, as shown in Figure 6-4 on page 228. You can stop the Quick Memory Test at any time and run a Full Memory Test, which runs the same test patterns multiple times and takes five times longer than the Quick Memory Test. The only time that you want to run the 227
Full Memory Test is if you have an intermittent memory problem that you are trying to isolate. Because the server identifies which specific DIMMs are experiencing excessive single-bit failures, it is far more efficient to swap reported DIMMs with similar DIMMs inside the server to see if the problem follows the DIMMs, stays with the memory slots, or simply goes away because the DIMMs were reseated.
3. The quick diagnostics continue to run, reporting the test that the quick diagnostics are performing currently and the length of time that it will take to complete that phase. If an error occurs, the quick diagnostics stop and indicate the memory errors that were encountered before progressing into more advanced diagnostics, as shown in Figure 6-5.
228
Important: Never warm-boot the server while running the built-in diagnostics. Several built-in functions that are used in the normal operation of the server are disabled during diagnostics to get direct results from the hardware. Only a normal exit from diagnostics or a cold boot of the server will re-enable those functions. Failure to perform this task correctly causes the server to become unstable. To correct this problem, simply power off the server and power it back on.
229
Figure 6-6 on page 229 shows the Memory configuration panel with your choice of performing memory sparing (Memory Spare Mode) or memory mirroring (memory mirror mode), but not both options at the same time. Select the desired option and reboot the server. If your memory population order does not support the requested option, the server reports a memory configuration error during the next reboot. See 3.8.4, Memory mirroring on page 87 for the correct memory population order to support memory mirroring or sparing.
The MAX5 also has Field Programmable Gate Array (FPGA) firmware that resides on it. The MAX5 firmware is updated at the same time that the server to which it is attached has its 230
IBM eX5 Implementation Guide
FPGA firmware flashed. The recommendation is to reflash the FPGA on your system x3850 X5 after connecting the MAX5 to ensure that the FPGA levels on the MAX5 and server match. The server might fail to complete POST because of a significant difference in FPGA code between the MAX5 and the server. Correct this situation by flashing the FPGA through the IMM while the server is plugged into power but not powered off. An entry in the hardware error log reports when the FPGA levels of the MAX5 and the server do not match. The order of the flash types in Table 6-2 on page 230 is the sequence in which to perform the flashes. To update the server to this firmware or higher, you can use one of the following choices: When you have a compatible operating system installed, you can use the UpdateXpress System Packs Installer (UXSPI) tool that is described in 9.11, UpdateXpress System Pack Installer on page 511. Regardless of the version of operating system or when an operating system is not installed, you can use the Bootable Media Creator (BOMC) tool that is described in 9.12, Bootable Media Creator on page 514. After establishing a network connection and logging into the IMM, use the firmware update function, as described in 9.13, MegaRAID Storage Manager on page 521. UEFI update is a two-step process: The UEFI flash is a two-step process. When the UEFI flash completes, warm-reboot the server to at least the F1-Setup panel to allow the second half of the UEFI flash to complete. Failure to perform the required warm-reboot results in a corrupt version of the UEFI, which forces the server to boot into the recovery page of the UEFI until the situation is corrected by following the UEFI recovery process in the Problem Determination and Service Guide (PDSG). For the latest firmware requirements for using the MAX5, see RETAIN tip H197572: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5085756
231
keep a set of these plastic boots for times when you want to remove the QPI cables for the movement of equipment from rack to rack or for servicing the unit.
There are four QPI cables connecting all of the QPI ports of the x3850 X5 to all of the QPI ports on the MAX5. Figure 6-8 shows how to connect the cables between the QPI ports of the two units.
Rack rear
x3850 X5 MAX5
The QPI cables are keyed to only be inserted one way. A quick visual check for cable orientation is to look for the 4U QPI or 1U QPI labels on the cable. The labels, along with the blue retainer release tab, are placed on what becomes the visible top of the cable when the cables are installed correctly. The ends of the cables are labeled to indicate which end to insert into the correct equipment. The 4U QPI end of the cables plugs into the x3850 X5. The 1U QPI end of the cables plugs into the MAX5. The cable end slides into the port until it clicks into place. You can disengage the retainer that holds the cable in place by pressing on the blue tab on the top of the QPI cable connection. QPI cables for connecting QPI ports 1 to 1 and 4 to 4 must be installed, even when only one processor is installed, to allow the MAX5 to be controlled by the server. If one of the cables is detached, the server does not power on or complete POST. You must install QPI cables for each of the processors installed to ensure full memory access to the MAX5. Table 6-3 on page 233 describes the QPI port on the back of the server with the corresponding processor socket in the server.
232
Table 6-3 QPI port relationship to the processor socket QPI port number 1 2 3 4 Processor socket number 4 3 2 1
Important information about FPGA firmware: When attaching the MAX5 for the first time, reapply the FPGA firmware using the IMM firmware update tool after the server is plugged into ac power but prior to powering up the server. Do not use Bootable Media Creator or USXPI to update the FPGA until the FPGA firmware is a match between the server and the MAX5. Mismatched FPGA makes the server unstable and the server can power off during the flash, corrupting the FPGA. This situation results in hardware replacement, because there is no recovery for corrupt FPGA. Figure 6-9 shows the back of the x3850 X5 with the attached MAX5.
Important power considerations: With the server or MAX5 plugged into ac power and the scaled units power turned off, the QPI cables still have active dc power running through them. Only unplug or plug in the QPI cable when neither the MAX5 nor the server is still plugged into ac power. Failure to follow this rule results in damaged circuits on either the MAX5 memory board or the processor board of the server.
chassis, it is not a requirement to remove the QPI cables before removing the memory board. 2. Remove the front bezel of the MAX5 by pressing in on the tab buttons on both sides of the bezel. The bezel can be pulled away from the MAX5. 3. As shown in Figure 6-10, there are two blue release tabs that, when pressed to the sides of the enclosure, allow you to pull out the cam levers that are used to begin to pull out the memory tray.
Figure 6-10 Blue release tabs for the memory tray cam levers
4. You can pull the memory tray out about 30% before it stops. This design allows you to get a better grip of the tray on either side and then use a finger to push in another set of blue release tabs on either side of the tray, as shown in Figure 6-11.
Figure 6-11 Final release tab for removing the MAX5 memory tray
5. Slide the memory tray completely out and place it on a flat work surface to work on the memory. Important: You must remove the ac power from both the server and the attached MAX5 before removing the memory tray. The FPGA of both the server and the MAX5 is still an active component when the server is powered down but not removed from utility power. FPGA components exist on both the server and the MAX5. Removing the memory board with the ac power still active damages the FPGA components of both the server and the MAX5. For memory population order, see 4.9, Storage on page 145.
234
After you install and configure the MAX5 properly, a successful link between the server and the MAX5 can be confirmed when both units power on and off as one complete unit and the memory in the MAX5 can be seen by the server. You also see the following message during POST: System initializing memory with MAX5 memory scaling
You might experience these situations if code levels do not match: Severe FPGA mismatch: Neither node can power on. The hardware event log, accessed through the IMM management port, reports an FPGA mismatch. Minor FPGA mismatch: The system powers on, but it reports an FPGA mismatch during POST, prior to the F1 Splash panel, and in the hardware event log of the IMM. Expect spontaneous reboots, memory failures, processor failures, QPI link failures, and machine check errors. Severe IMM mismatch: The secondary node does not power up. The IMM of the primary node is unable to access events occurring in the secondary node. Minor IMM mismatch: The primary IMM is unable to collect all of the hardware information from the secondary node. A newer version of the IMM prevents both nodes from powering up. UEFI mismatch: When two scaled x3850 X5 servers are powered up, the IMM loads the UEFI from only the primary node. The UEFI of the secondary node is never used. The firmware level of the primary node becomes the running firmware level of both nodes.
Chapter 6. IBM System x3850 X5 and x3950 X5
235
236
There are four QPI cables connecting all of the QPI ports of both x3850 X5 servers. Figure 6-13 shows the cable routing.
Rack rear
Consider this important information: The QPI cables are keyed to be inserted only one way. A quick visual check for cable orientation is to look for the blue retainer release tab on the top of each cable end. The blue retainer release tab is placed on what becomes the visible top of the cable when the cable is installed correctly. The ends of the cable are labeled to indicate which end can be inserted into the correct equipment. The server that becomes the primary node has the end of the QPI plugged into it with the labels J2 and J1, as shown in Figure 6-14. The secondary node has the end of the QPI cable plugged into it with the labels J4 and J3. All four QPI cables must be oriented in the exact same manner. Neither server can power on if the cables are not consistent in their orientation. Figure 6-14 shows how the cable ends are labeled.
Figure 6-14 QPI cable end labels: J1 and J2 attach to the primary node
237
The cable end must slide into the port until it clicks into place. You can disengage the retainer that holds the cable in place by pressing on the blue tab on the top of the QPI cable connection. QPI cables for connecting QPI ports 1 to 1 and 4 to 4 must be installed, regardless of whether processors 2 and 3 are installed, to allow the secondary node to be controlled by the primary node. If one of the cables is detached, the server cannot power on. QPI cables must be installed for each of the processors installed. Table 6-5 describes the QPI port on the back of the server with the corresponding processor socket in the server.
Table 6-5 QPI port relationship to the processor socket QPI port number 1 2 3 4 Processor socket number 4 3 2 1
After you attach the cables, connect the power to both nodes of the complex. When the power status light changes from a rapidly blinking light to a slow blinking green light, press the power button on the primary node. Both nodes power up. If both nodes do not power up, double check that the required matching firmware is on both nodes of the server. You must disconnect the QPI cables to determine if the required matching firmware is on both nodes of the server. If neither node powers up or the secondary node does not power up, consider swapping QPI cables between port 1 on both nodes to one of the cross-port cables between ports 2 and 3. If this action proves unsuccessful, do the same task with the QPI cables between port 4 of both nodes and again one of the two crossed QPI cables between ports 2 and 3. There are additional communication lanes on port 1 for synchronization of the CPU frequencies and additional communication lines on port 4 for communication between the FPGA of the primary node and the secondary node. If this approach works, check the IMM event log for any QPI link failures between processors to see if one of the cross-linked QPI cables between ports 2 and 3 needs to be replaced. Frequently, one of the critical QPI cables was not completely seated. After the complex is up and running, double check the IMM event log for any QPI link failures. The error message reports which processors experience the error. Based on Table 6-5, the processor reporting the QPI link error points to a QPI cable on a specific port. Provided the two nodes are running matched IMM and FPGA, the problem can be the processor or QPI port on either end of the QPI cable or the cable itself. The cable is the most likely point of failure. After the two nodes boot as a single entity, any firmware flashes applied to the primary node are automatically applied to the secondary node.
238
stand-alone database server with built-in fault tolerance for power and cooling, processors, memory, and PCIe buses.
PCIe adapters connect to the processors via the I/O hub. The purpose of the I/O hub is to combine the various data streams from each of the PCIe slots into a single aggregate link to the processors using a dedicated QPI link to each processor. The x3850 X5 has two I/O hubs. The data transfer rate of the QPI link is negotiated between the processor and the I/O hub. Table 3-9 on page 74 shows the QPI link speeds based on the type of installed processors. The I/O hub supports the highest QPI link speed that is shown in the table, 6.4 gigatransfers per second (GT/s). You can also adjust the QPI link speed to conserve power by booting into F1-Setup, selecting System Settings Operating Modes, and setting QPI Link Frequency to values other than the default Max Performance, as shown in Figure 6-15 on page 240.
239
240
Figure 6-16 PCIe slot speed selection panel to force Gen1 compliance
241
Another possible solution to this problem is booting the server in Legacy Only mode. This mode allows both non-UEFI-aware operating systems and PCIe adapters to function as they function on a non-UEFI server. Many of the advanced memory addressing features of the UEFI environment are not available to the operating system in this mode. Booting the server to an operating system in this mode allows you the ability to apply firmware updates to the Gen1 adapter that is not recognized in an UEFI environment. To enable this feature from within F1-Setup, select Boot Manager Add Boot Option, scroll down, and select Legacy Only when it is displayed in the list of options. Figure 6-18 on page 243 shows where Legacy Only is located in the list of available boot options. If Legacy Only is not listed, it has already been added to the Boot Manager.
242
Besides adding Legacy Only to the Boot Manager, you must also change the sequence to place Legacy Only at the top of the boot sequence. To change the boot order from within the Boot Manager panel, select Change Boot Order. Figure 6-19 shows how the boot order panel is activated by pressing Enter and then selecting the item to move by using the arrow keys. When Legacy Only is selected, use the plus (+) key to move it up the panel.
243
Microsoft Windows 2008 x64: When installed on an UEFI server, Microsoft Windows 2008 x64 will install Microsoft Boot Manager as part of the boot sequence. Regardless of how you change the boot sequence in the boot manager, Microsoft Boot Manager will always be at the top of the sequence. When you install this same operating system with Legacy Only enabled, Microsoft Boot Manager is not installed as part of the boot manager. Removing the Legacy Only option from the boot manager will prevent the server from booting into the installed Windows 2008 x64.
3 and 4
When installing teamed network adapters or multipath fibre host bus adapters (HBAs), each member of the team must be placed in separate zones to maximize the throughput potential of this server and minimize the loss of a single processor. Consider these key points about these slots: PCIe slot 1 is an x16 slot to support possible future I/O adapters. It can be used to hold any PCIe adapter. PCIe slot 2 is actually an x4 with an x8 mechanical connection. Do not use this slot for an SSD RAID controller. PCIe slot 7 is an x8 slot and is specifically designed to support an Emulex 10GbE adapter that is designed for this server. The RAID SAS port is an x8 PCIe slot that supports a wide variety of ServeRAID and SAS HBAs. This slot is not suitable for adapters, such as SSD controllers, that generate a lot of heat; these adapters need to be installed in a slot at the rear of the server with better airflow. While true performance on a given PCIe adapter can largely depend on the configuration of the environment on which it is used, there are general performance considerations with respect to the x3850 X5 server. The I/O hub supports 36 lines of PCIe traffic with a combined bandwidth of 36 GBs. Each processors QPI link to the I/O hub is capable of a maximum throughput of 26 GBs, depending on the processors installed on the server. With only one processor installed, the maximum combined bandwidth of all the PCIe lanes is reduced to the maximum bandwidth of a single
244
QPI link. If two matched processors of any QPI link speed are installed, this limit is no longer an issue. Of all of the I/O adapters that can be installed on the server, the ServeRAID and 6 Gb SAS controllers managing SSD drives are the only adapters that can approach the theoretical limits of the x8 PCIe slot. Therefore, when you use SSD drives, connect no more than eight SSD drives to a single controller for the best performance of the SSD drives. Also, only use x8 slots to host the controllers that manage your SSD drives. A single ServeRAID controller managing a single 4-drive SAS HDD array will function within the theoretical limits of an x4 PCIe slot. In this case, the mechanical nature of the HDD drives will limit the maximum throughput of data that passes through the PCIe slot. The dual-port 8 Gbs Fibre Channel, 10Gbs Ethernet, and 10 Gbs Converge Network adapters (CNA) are all capable of approaching the theoretical limits of an x4 Gen2 PCI slot and might perform better in an X8 Gen2 slot.
245
When booting from SAN with multiple paths for redundancy, you will need to enable the legacy option ROM for both HBAs.
246
Other PCI adapters can have their boot option ROM disabled from within their configuration panels. To access individual adapter configuration panels from F1-Setup, select System Settings Adapters and UEFI Drivers and press Enter. Figure 6-22 shows the selections that are presented by this panel when accessed.
247
To enter the configuration of a specific adapter, select the PciROOT directly under the adapter name. When you have multiple controllers of the same type, selecting the PciROOT of any of the same adapter types will select all of them and display a panel that allows you to select the specific adapter from within the configuration routine of the adapter type. Figure 6-23 demonstrates this process when two ServeRAID adapters are installed on the server.
Figure 6-23 LSI adapter selection panel from within the LSI configuration panel
The controller configuration properties for adapters that are similar to the ServeRAID or Fibre Channel HBAs you are looking for are displayed, as shown in Figure 6-24.
Figure 6-24 ServeRAID BIOS Config Utility Controller Properties: Disabling Boot ROM
Although it is not stated or obvious from the description, disabling the Controller BIOS only disables the Boot ROM execution during POST. All of the other operating characteristics of the adapter BIOS and firmware remain intact.
248
installed in the server. To control the boot sequence, from within F1-Setup, select System Settings Device and I/O Protest Option ROM Execution Order. Figure 6-25 shows what the panel looks like after pressing Enter on the list of possible choices. Use the up and down arrow keys to select a specific entry to move and use the plus or minus keys to move the item up or down the list.
249
This precaution will eliminate the possibility of a single breaker or circuit fault from taking down the entire server. Think of the power supplies in your server as shock absorbers in your car. They are designed to absorb and overcome a wide variety of power conditions that can occur from an electric utility company, but like shock absorbers on a car, they will eventually begin to fail when fed a steady diet of unstable power. The time of their failure will most likely not coincide with a planned maintenance panel. For this reason, ensure that the two halves of the power supplies are plugged into two separate UPS sources to filter out all of the moderate to severe power fluctuations that occur.
Tip: If you have a server but you do not know the logon credentials, you can reset it by going to the panel that is shown in Figure 6-26, from F1-Setup, and then restore the IMM configuration to the factory defaults by selecting Reset IMM to Defaults.
250
When configured as Dedicated, you are connecting to the network via the system management port. As shown in Figure 6-28, the port is located from the rear of the server on the left side of the video port. Using this port allows for easier separation of public and
Chapter 6. IBM System x3850 X5 and x3950 X5
251
management network traffic. Separating the traffic is done when you connect your public network port to switch ports that belong to a public access virtual LAN (VLAN) and the management port is connected to a switch port defined by a separate management VLAN.
When configured as Shared, you are sharing network traffic on the second onboard Ethernet port, the one closest to the power supply, as shown in Figure 6-29 on page 252. While this configuration eliminates a physical switch port and patch cable configuration, both the media access control (MAC) address for the second Ethernet port and the MAC address for the IMM will address through this single network port. This situation means at least two separate IP addresses for the same physical port, which prevents you from configuring the onboard Ethernet ports in a network team using 802.3ad load balancing. Using this type of load balancing scheme will result in dropped packets for the IMM MAC address. Smart load balancing and failover are still available network teaming options. However, keeping the public traffic from the management traffic becomes more difficult. To maintain separation between public and management traffic, network teaming software must be used to establish a VLAN to be used by the server to send public-tagged traffic to the network switch. The switch port must be configured as a trunk port to support both the public-tagged VLAN traffic, plus the untagged traffic for the management. The management VLAN must be defined as the native VLAN on the switch port, so that its untagged traffic from the switch will be accepted by the IMM MAC and dropped by the second Ethernet ports MAC.
Figure 6-29 The onboard Ethernet port used when IMM Network interface is Shared
While the IMM uses a dedicated RISC processor, there are limitations as to the amount of network traffic that the IMM can be exposed to before complex functions, such as booting from a remote DVD, or USB storage will become unreliable because of timing issues. While the operating system has all of the necessary drivers in place to deal with these timing issues, the UEFI is not as tolerant. For this reason (maintaining secured access), the IMM must be kept on a separate management network.
252
System Status
The first panel that you will see after completing the login and the session timeout limits panel is the System Status panel, as shown in Figure 6-30 on page 254. This panel provides a quick summary review of the hardware status of the server. A green circle indicates that all is working from a strict hardware point of view. The IMM can check on the status of server components, the ServeRAID controllers, and PCIe interfaces to most PCIe adapters. It does not check on the functional status of most PCIe adapters with regard to their hardware connections to external devices. You will need to refer to the system event log from within the operating system or the switch logs of the network and fibre switches to which the server is connected to resolve connectivity issues.
Chapter 6. IBM System x3850 X5 and x3950 X5
253
When an actual hardware error is detected in the server, the system status is represented by a red X. The System Health Summary will provide information about the errors presently unresolved in the server, as shown in Figure 6-31 on page 255.
254
255
Most servers are not physically located near the people who manage them. To help you see the event from a remote location, the IMM provides the capability of looking at all tiers of light path diagnostics, as shown in Figure 6-33.
256
Remote control
Certain problems require that you get into the operating system or F1-Setup to detect them or fix them. For remotely managed servers, you can use the Remote Control feature of the Integrated Management Module. Figure 6-35 shows the options available for starting a remote control session.
Figure 6-35 Integrated Management Module Remote Control session start-up panel
257
IMM Remote Control provides the following features: The remote control provides you with the same capability that you have with a keyboard, mouse, and video panel directly connected to the server. You have the ability to encrypt the session when used over public networks. You have the ability to use local storage or ISO files as mounted storage resources on the remote server you are using. These storage resources can be unmounted, changed, and remounted throughout the session, as needed. When combined with the Power/Restart functions of the IMM, you can power down, reboot, or power on the server while maintaining the same remote control session. Depending on the application that you are accessing through the IMM Remote Control, you might notice that the mouse pointer is difficult to control. Fix this problem in the Video Viewer by selecting Tools Single Cursor, as shown in Figure 6-36.
Figure 6-36 Fixing the mouse pointer in the Remote Control Video Viewer
258
If you use the factory defaults UEFI settings, the machine works in a either a 1-node, 2-node, or 1-node with MAX5 configuration. You can also change UEFI settings to meet your system requirements. In this section, we provide an overview of the UEFI settings for tuning your system for performance. For an explanation of each setting, see 2.7, UEFI system settings on page 36. You can use the Advanced Settings Utility (ASU) tool to change the UEFI settings values. ASU exposes more settings than the settings accessed using the F1-Setup panel. For more information about ASU, see 9.7, Advanced Settings Utility (ASU) on page 495. Table 6-8 on page 260 provides an overview of the most common UEFI settings for optimizing system performance. 259
Table 6-8 UEFI settings, ASU values, and default settings UEFI value Processor settings TurboMode TurboBoost Power Optimization Processor Performance States CPU C-States C1 Enhanced Mode Processor Data Prefetcher Hyper-Threading Execute Disable Bit Intel Virtualization Technology QPI Link Frequency uEFI.TurboModeEnable uEFI.TurbBoost uEFI.ProcessorEistEnable uEFI.ProcessorCcxEnable uEFI.ProcessorC1eEnable uEFI.ProcessorDataPrefetch uEFI.ProcessorHyperThreading uEFI.ExecuteDisableBit uEFI.ProcessorVmxEnable uEFI.QPISpeed Enable Disable Traditional PowerOptimized Enable Disable Enable Disable Enable Disable Enable Disable Enable Disable Enable Disable Enable Disable Max Performance Power Efficiency Minimal Power Enable PowerOptimized Enable Disable Enable Enable Enable Enable Enable Max Performance ASU value ASU settings Default
Memory Settings N/A CKE Low Power Memory Speed IMM.ThermalModePolicy uEFI.CkeLowPolicy uEFI.DdrSpeed Normal Performance Enable Disable Max Performance Power Efficiency Minimal Power Static Trade Off Static Read Primary Static Write Primary Adaptive Open Closed Enable Disable Enable Disable Normal Disable Max Performance
uEFI.SchedulerPolicy
Adaptive
260
The MAX5 Memory Scaling Affinity setting provides two options to determine how the system will present the memory capacity in the MAX5 unit to the OS: Non-Pooled The default option. Non-Pooled splits the memory in the MAX5 and assigns it to each of the installed processors. Configure VMware and Microsoft OSs to use the Non-Pooled setting.
261
Pooled This option presents the additional memory in the MAX5 as a pool of memory without being assigned to any particular processor. Use this setting with Linux OSs.
Disabled Disabled Enabled Enabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled
Enabled Enabled Enabled Enabled Enabled Enabled Max Performance Performance Disabled Max Performance Closed Closed Disabled
Disabled Disabled Enabled Disabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled
Enabled Enabled Enabled Enabled Enabled Enabled Power Efficiency Normal Disabled Max Performance Closed Closed Disabled
Enabled Enabled Enabled Enabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled
262
Virtualizationb Enabled
HPC Disabled
Depending on the processor workload, enabling TurboMode might also increase power consumption. The actual processing performance boost that you get from TurboMode depends on the environment that the server is in terms of temperature and humidity, because the processor will only boost performance up to the environmental limits set by the processor. b These Virtualization settings are recommended for a stand-alone host only. For multiple virtualized hosts, in clustered workloads, use the Maximum performance settings instead.
6.10 Installing an OS
This section provides an overview of the options that you have when installing an OS on the x3850 X5. We provide instructions for VMware ESX/ESXi installation when a MAX5 is attached to the x3850 X5 and for installing an OS with a USB key are provided, as well as additional installation topics. We recommend that you use ServerGuide for your OS install procedure. For more information about ServerGuide, see 9.8, IBM ServerGuide on page 501. Topics in this section: 6.10.1, Installing without a local optical drive on page 263 6.10.2, Use of embedded VMware ESXi on page 271 6.10.3, Installing the ESX 4.1 or ESXi 4.1 Installable onto x3850 X5 on page 275 6.10.4, OS installation tips and instructions on the web on page 288 6.10.5, Downloads and fixes for x3850 X5 and MAX5 on page 293 6.10.6, SAN storage reference and considerations on page 294
IMM
A remote control feature is available through the IMM web interface. You must log in to the IMM with a user ID that has Supervisor access. You can also assign to the server a CD or DVD drive, diskette drive, USB flash drive, or disk image that is on your computer. The following server OSs have USB support, which is required for the Remote Disk feature: Microsoft Windows Server 2008 Microsoft Windows Server 2003 Red Hat Linux versions 4.0 and 5.0 SUSE Linux version 10.0
Chapter 6. IBM System x3850 X5 and x3950 X5
263
Novell NetWare 6.5 For more information, see the Integrated Management Module User's Guide at the following web page: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5079770 Follow these steps to mount a drive through IMM: 1. Connect to the IMM with your web browser. 2. Click Task Remote Control. 3. If you want to allow other users remote control access during your session, click Start Remote Control in Multi-user Mode. Otherwise, click Start Remote Control in Single User Mode. 4. Two Java application windows open, as shown in Figure 6-39 and Figure 6-40.
264
5. Select the Virtual Media Session window. 6. Click Add Image if you want to map an IMG or ISO image file. 7. Select the check box next to the drive that you want to map and click Mount Selected, as shown in Figure 6-41.
8. The image drive is now accessible by the system. Closing the session: Closing the Virtual Media Session window when a remote disk is mapped to the machine causes the machine to lose access to the remote disk.
Internet Explorer: If you use Internet Explorer 7 or 8 and the remote control window does not open, see RETAIN tip H196657 for steps to solve the problem: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5083262
265
For Linux, look on the vendor websites. They contain information about installation with a USB flash drive. For example, the following web pages provide details for using a USB key as an installation medium: Installing Red Hat Linux from a USB flash drive: http://ibm.com/support/techdocs/atsmastr.nsf/WebIndex/WP101131 How to create a bootable USB drive to install SLES: http://www.novell.com/support/php/search.do?cmd=displayKC&docType=kc&externalId =3499891 You can also use the ServerGuide Scripting Toolkit to create a bootable USB flash drive, as explained in the next section.
Windows installation
This section describes the process to install the ServerGuide Scripting Toolkit to create a deployment image for Windows 2008 R2 Enterprise Edition and then to copy this image to a USB key for deployment. To configure a USB key for deployment, you need the following requirements: A system running Windows Vista, Windows Server 2008, Windows 7, Windows Server 2008 R2, Windows 2.1 Preinstallation Environment (PE), or a Windows 3.0 PE session A USB key with a storage capacity at least 64 MB larger than your Windows PE image, but not less than 4 GB. We follow this procedure: 1. Install the ServerGuide Scripting Toolkit. 2. Create a deployment image. 3. Prepare USB key.
266
3. Decompress the ibm_utl_sgtkwin_X.XX_windows_32-64.zip file to the directory that you have created, for example, C:\sgshare\sgdeploy.
3. Choose the OS type that you want and click Next, as shown in Figure 6-43.
4. Insert the correct OS installation media or select the folder that contains the installation files for the source, as shown in Figure 6-44 on page 268. If necessary, modify the target and click Next.
267
5. When the copy process is finished, as shown in Figure 6-45, click Finish.
6. Open a command prompt and change to the C:\sgshare\sgdeploy\SGTKWinPE directory. Use the following command to create the Windows installation image: SGTKWinPE.cmd ScenarioINIs\Local\Win2008_R2_x64_EE.ini 7. When the process is finished, as shown in Figure 6-46 on page 269, your media creation software is started to create bootable media from the image. Cancel this task.
268
18:26:21 - Creating the WinPE x64 ISO... 18:27:07 - The WinPE x64 ISO was created successfully. *** WinPE x64 ISO: 4_EE\WinPE_x64.iso c:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x6
18:27:07 - Launching the registered software associated with ISO files... *** Using ISO File: 64_EE\WinPE_x64.iso c:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x
18:27:08 - The WinPE x64 build process finished successfully. SGTKWinPE complete. c:\sgshare\sgdeploy\SGTKWinPE> Figure 6-46 Build process is finished
269
2. Enter diskpart to format the USB key using FAT32. All files on the USB key will be deleted. At the command prompt, type the commands that are listed in Figure 6-47.
C:\>diskpart Microsoft DiskPart version 6.1.7600 Copyright (C) 1999-2008 Microsoft Corporation. On computer: DISKPART> list disk Disk ### -------Disk 0 Disk 1 Disk 2 Status ------------Online Online Online Size ------271 GB 135 GB 7839 MB Free Dyn ------- --0 B 0 B 0 B Gpt --*
DISKPART> select disk 2 Disk 2 is now the selected disk. DISKPART> clean DiskPart succeeded in cleaning the disk. DISKPART> create partition primary DiskPart succeeded in creating the specified partition. DISKPART> select partition 1 Partition 1 is now the selected partition. DISKPART> active DiskPart marked the current partition as active. DISKPART> format fs=fat32 100 percent completed DiskPart successfully formatted the volume. DISKPART> assign DiskPart successfully assigned the drive letter or mount point. DISKPART> exit Figure 6-47 Using diskpart to format the USB memory key
3. Copy the contents from C:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x64_EE\ISO to the USB key. The USB key includes the folders and files that are shown in Figure 6-48.
4. Boot the target system from the USB key. The deployment executes automatically.
270
RAID controller: If the target system contains a RAID controller, RAID is configured as part of the installation.
271
4. Select Change Boot Order. 5. Change the boot order to Legacy Only followed by Embedded Hypervisor, as shown in Figure 6-50.
272
2. In the Loading VMware Hypervisor panel, press Shift+O when the progress bar is displayed. 3. Enter the following command at the prompt: esxcfg-advcfg -k TRUE allowInterleavedNUMAnodes 4. After the system boots, connect to the system using the vSphere Client. 5. Select the Configuration tab of the host and click Advanced Settings under Software, as shown in Figure 6-51.
6. Click VMkernel and select the check box next to VMkernel Boot.allowInterleavedNUMAnodes, as shown in Figure 6-52 on page 274.
273
Figure 6-52 Editing the WMkernel settings in vSphere Client Advanced Settings
Updating ESXi
You can install the latest version of ESXi 4 on IBM Hypervisor keys and it is supported by IBM. Use the following VMware upgrade mechanisms for the update: VMware Upgrade Manager Host Update Utility For more information, see the VMware Documentation website: http://www.vmware.com/support/pubs
ESXi recovery
You can use the IBM recovery CD to recover the IBM USB Memory Key to a factory-installed state. Table 6-10 provides the part numbers and versions for the CDs.
Table 6-10 VMware ESXi recovery CD Part number 68Y9634 49Y8747 68Y9633 46M9238 46M9237 46M9236 46D0762 Description VMware ESXi 4.0 U1 VMware ESXi 4 VMware ESX Server 3i v 3.5 Update 5 VMware ESX Server 3i v 3.5 Update 4 VMware ESX Server 3i v 3.5 Update 3 VMware ESX Server 3i v 3.5 Update 2 VMware ESX Server 3i version 3.5
274
To order a recovery CD, contact your local support center at the following website: http://www.ibm.com/planetwide/region.html
6.10.3 Installing the ESX 4.1 or ESXi 4.1 Installable onto x3850 X5
Before installing any VMware OS, always refer to the latest OS support information that is contained on the IBM ServerProven site. You can obtain IBM ServerProven information at the following website: http://ibm.com/systems/info/x86servers/serverproven/compat/us/nos/vmwaree.html
Tip: If your System has 1 Terabyte (TB) or greater of memory, see the following web page: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5084935
Installing the ESX/ESXi 4.1 Installable edition onto x3850 X5 with MAX5
To correctly configure and install ESX/ESXi 4.1 Installable editions of ESX/ESXi Server, follow these instructions: 1. Connect the MAX5, as described in 6.4, Attaching the MAX5 memory expansion unit on page 230. 2. Configure memory scaling. Systems running VMware ESX/ESXi Server must use Non-Pooled mode in the MAX5 Memory Scaling option. See 1-node with MAX5 on page 261 for instructions to configure the MAX5 Memory Scaling option. 3. Configure your RAID arrays. 4. Use the following procedures to install ESX/ESXi 4.1 Installable on an x3850 X5 with MAX5: Installing the ESX 4.1 OS on page 276 Installing the ESXi 4.1 OS on page 278
275
6. Select Commit Changes and press Enter to save the changes. 7. Boot the host from the ESX installation media. 8. Press F2 when you see the ESX 4.1 installation options panel, as shown in Figure 6-54.
276
9. The Boot Options line will appear on the panel. Type the following parameter at the end of the Boot Options line: allowInterleavedNUMAnodes=TRUE The edited result looks like Figure 6-55. Press Enter to proceed.
10.Proceed through the installer until you reach the Setup Type page. Click Advanced setup and clear Configure bootloader automatically (leave checked if unsure) as shown in Figure 6-56. Click Next.
277
11.Proceed through the installer to the Set Bootloader Options page and type the following parameter in the Kernel Arguments text box, as shown in Figure 6-57: allowInterleavedNUMAnodes=TRUE
12.Complete the remainder of the ESX installation and reboot the host.
6. Select Commit Changes and press Enter to save the changes. 7. Boot from the ESXi Installable installation media. 8. Press the Tab key when the blue boot panel appears, as shown in Figure 6-59 on page 279.
278
VMware VMvisor Boot Menu ESXi Installer Boot from local disk
9. Add the following line after vmkboot.gz: allowInterleavedNUMAnodes=TRUE Ensure that you leave a space at the beginning and the end of the text that you enter (as shown in Figure 6-60); otherwise, the command will fail to execute at a later stage during the installation. Press Enter to proceed after the line has been edited correctly.
VMware VMvisor Boot Menu ESXi Installer Boot from local disk
> mboot.c32 vmkboot.gz allowInterleavedNUMAnodes=TRUE --- vmkernel.gz --- sys.vg z --- cim.vgz --- ienviron.vgz --- install.vgz
10.Complete the ESXi installation and reboot when prompted. Ensure that you remove the media or unmount the installation image before the system restarts. 11.In the Loading VMware Hypervisor panel, press Shift+O when the progress bar is displayed, as shown in Figure 6-61 on page 280.
279
IMPORTANT: If you do not press Shift+O during the Loading VMware Hypervisor panel, you will receive the error that is shown in Figure 6-62: The system has found a problem on your machine and cannot continue. Interleaved NUMA nodes are not supported.
12.Enter the following command at the prompt after you have pressed Shift+O: esxcfg-advcfg -k TRUE allowInterleavedNUMAnodes Your output looks like Figure 6-63 on page 281.
280
13.Press Enter after the command has been entered. Press Enter again to continue to boot. 14.After the system boots, connect to it using the vSphere Client. 15.Select the Configuration tab of the host and click Advanced Settings under the Software panel, as shown in Figure 6-64.
16.Click VMkernel in the left pane and select VMkernel.Boot.allowInterleavedNUMAnodes, as shown in Figure 6-65 on page 282. Click OK when finished. This step concludes the installation process.
281
ESX installation
Use the following steps to change your ESX 4.1 installation: 1. Use the vSphere Client to connect to the system. 2. Select the Configuration tab of the host and click Advanced Settings under the Software panel. 3. Click VMkernel in the left pane and select VMkernel.Boot.allowInterleavedNUMAnodes, as shown in Figure 6-66 on page 283. Click OK.
282
4. Power off the x3850 X5 and unplug all power cords. 5. Attach the MAX5 to the x3850 X5. For more information, see 6.4, Attaching the MAX5 memory expansion unit on page 230. 6. Replug the power cords on x3850 X5 and MAX5 but do not power on the complex
configuration.
7. Log in to the IMM web interface of the x3850 X5. 8. Click Firmware Update under Tasks, as shown in Figure 6-67.
9. Click Browse and select the Field Programmable Gate Array (FPGA) update file. 10.Click Update to start the update process. A progress indicator opens as the file is transferred to the temporary storage of IMM. 11.When the transfer is complete, click Continue to complete the update process. A progress indicator opens as the firmware is flashed. A confirmation page opens to verify that the update was successful. 12.Unplug the power cords from the x3850 X5 and MAX5. Wait 1 minute and replug the power cords. This procedure activates the new FPGA code on the x3850 X5 and MAX5. 13.Wait approximately 10 minutes to power on the x3850 X5. There cannot be any LEDs lit on the Light Path panel. The System Health Status in the IMM web interface looks like Figure 6-68 on page 284.
283
14.Power on the x3850 X5. 15.At the UEFI panel, press F1 for the UEFI menu. 16.Select System Settings Memory and change the MAX5 Memory Scaling Affinity to Non-Pooled, as shown in Figure 6-69.
17.Press Esc two times to go back to the Main page. Select Save Settings and press Enter to save the UEFI settings. 18.Exit the UEFI and boot the system to ESX. Important: If you do not select VMkernel.Boot.allowInterleavedNUMAnodes under VMkernel, Figure 6-70 shows the error that will appear during boot.
If you see this error, follow these steps to change the boot code: 1. At the VMware bootloader panel, press a to modify the kernel arguments. Be sure that VMware ESX 4.1 is highlighted, as shown in Figure 6-71 on page 285.
284
2. Add the following line at the beginning: allowInterleavedNUMAnodes=TRUE Ensure that you leave a space at the beginning and the end of the text that you enter (as shown in Figure 6-72); otherwise, the command will fail to execute at a later stage during the boot process. Press Enter to proceed after the line has been edited correctly.
ESXi installation
Use the following steps to change your ESXi 4.1 installation: 1. Use the vSphere Client to connect to the system. 2. Select the Configuration tab of the host and click Advanced Settings under the Software panel.
285
3. Click VMkernel in the left pane and select VMkernel.Boot.allowInterleavedNUMAnodes. Click OK.
4. Power off the x3850 X5 and unplug all power cords. 5. Attach the MAX5 to the x3850 X5. For more information, see 6.4, Attaching the MAX5 memory expansion unit on page 230. 6. Replug the power cords on x3850 X5 and MAX5 but do not power on the complex
configuration.
7. Log in to the IMM web interface. 8. Click Firmware Update under Tasks, as shown in Figure 6-74.
Tasks Power/Restart Remote Control PXE Network Boot Firmware Update
9. Click Browse and select the FPGA update file. 10.Click Update to start the update process. A progress indicator opens as the file is transferred to the temporary storage of IMM. 11.When the transfer is completed, click Continue to complete the update process. A progress indicator opens as the firmware is flashed. A confirmation page opens to verify that the update was successful. 12.Unplug the power cords from the x3850 X5 and MAX5. Wait 1 minute and replug the power cords. This procedure activates the new FPGA code on the x3850 X5 and MAX5.
286
13.Wait approximately 10 minutes to power on the x3850 X5. There must be no LEDs lit on the light path diagnostics panel. The System Health Status in the IMM web interface looks like Figure 6-75.
14.Power on the x3850 X5. 15.At the UEFI panel, press F1 for the UEFI menu. 16.Select System Settings Memory and change the MAX5 Memory Scaling Affinity to Non-Pooled, as shown in Figure 6-76.
17.Press Esc two times to go back to the Main page. Select Save Settings and press Enter to save the UEFI settings. 18.Exit the UEFI and boot the system to ESXi. Important: If you forgot to select VMkernel.Boot.allowInterleavedNUMAnodes under VMkernel, Figure 6-77 shows the error that will appear during boot.
287
If this error occurs, follow these steps to change the boot code: 1. In the Loading VMware Hypervisor panel, press Shift+O when the progress bar is displayed, as shown in Figure 6-78.
2. Enter the following command at the prompt after you have pressed Shift+O: esxcfg-advcfg -k TRUE allowInterleavedNUMAnodes Your output looks like Figure 6-63 on page 281.
3. Press Enter after the command has been entered. Press Enter again to continue to boot. 4. Do not forget to select VMkernel.Boot.allowInterleavedNUMAnodes as active with the vSpere Client.
288
If installing Windows 2008 Enterprise Edition on an 8-socket system, Hyper-Threading must be turned off prior to installation or the system will stop with a blue window. The default installation installs the OS in UEFI mode. UEFI mode requires a separate imaging methodology, because UEFI mode requires the OS to reside on a GUID Partition Table (GPT) disk. The system can re-enumerate the boot drive when fibre devices are connected. This situation will cause the OS to not boot properly. Use Windows Boot Manager from the UEFI Boot Manager to fix this problem. You must set the MAX5 memory configuration in UEFI to Non-pooled mode. The recommendation is to review all RETAIN tips that are related to this system and the OS that you are installing. See the following website: http://ibm.com/support/entry/portal/Problem_resolution/Hardware/Systems/System_ x/System_x3850_X5 For further tuning information, see the Performance Tuning Guidelines for Windows Server 2008 at the following page: http://www.microsoft.com/whdc/system/sysperf/Perf_tun_srv.mspx
289
Consider these useful tips: The Dual-port Emulex 10GbE NIC driver is not included on the Windows Server 2008 DVD and must be downloaded separately (or use ServerGuide instead). You must ensure that the Flow Control is enabled under network adapter properties after upgrading the network device driver. When Flow Control is enabled, it allows the receipt or transmission of PAUSE frames. PAUSE frames enable the adapter and the switch to control the transmission rate. If installing Windows 2008 R2 Enterprise Edition on an 8-socket system, Hyper-Threading must be turned off prior to installation or the systems will show a blue screen. To correct an issue with the GPT disks, Windows 2008 R2 requires the Microsoft Hotfix that can be found at the following website: http://support.microsoft.com/kb/975535 The default installation media installs the OS in UEFI mode unless Legacy Only is the first entry in the boot order. In 2-node configurations, edit the Windows Boot Configuration Data (BCD) Store to enable the High Precision Event Timer (HPET). For more information, see the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5084072 The system can re-enumerate the boot drive when fibre devices are hooked up. This action will cause the OS to fail to boot properly. Use Windows Boot Manager from the UEFI Boot Manager to fix this problem. If your system has one TB or more of memory installed, it is a requirement to apply the Microsoft Hotfix that is available at the following website: http://support.microsoft.com/kb/980598 You must update with the latest BroadCom Ethernet device driver when your system has 64 or more processors (cores). You must set the MAX5 memory configuration in UEFI to Non-pooled mode. If your system has greater than 128 GB of memory and you plan to enable Hyper-V after installing Windows 2008 R2, you must first apply the Microsoft Hotfix that is available at the following website: http://support.microsoft.com/kb/979903/en-us The recommendation is to review all Retain Tips related to this system and the OS that you are installing. See the following website: http://ibm.com/support/entry/portal/Problem_resolution/Hardware/Systems/System_ x/System_x3850_X5 For further tuning information, see Performance Tuning Guidelines for Windows Server 2008 R2 at the following website: http://www.microsoft.com/whdc/system/sysperf/Perf_tun_srv-R2.mspx
290
The following websites contain additional OS installation guides: Installing VMware ESX Server 4.1 - IBM System x3850 X5 (7145): http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5086213 Installing VMware ESXi Server 4.1 Installable - IBM System x3850 X5 (7145): http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5086214 Consider these useful tips: VMware 4.0 U1 is the minimum supported OS level for a single-node x3850X5. The addition of a MAX5 or a second node requires at least VMware 4.1. Neither the included Emulex 10GbE NIC or the Emulex VFA, nor the Qlogic CNA driver, is included in VMWare ESX 4.0 U1. Download the drivers from VMWares site: http://downloads.vmware.com/d/info/datacenter_downloads/vmware_vsphere_4/4#driv ers_tools VMware 4.x does not include the driver for the Intel PRO/1000 Ethernet Adapter. Download the drivers from the VMware site: http://downloads.vmware.com/d/info/datacenter_downloads/vmware_vsphere_4/4#driv ers_tools For a 2-node configuration or x3850 X5 with MAX5, you need to edit your grub.conf file with the following line: allowInterleavedNUMAnodes=TRUE For more information, see Installing the ESX/ESXi 4.1 Installable edition onto x3850 X5 with MAX5 on page 275. ESX 4.1 generates a warning with 1 TB RAM, because 1 TB RAM is the maximum RAM for an ESX 4.1 host. For more information, see the Configuration Maximums for VMware vSphere 4.1 at the following website: http://www.vmware.com/pdf/vsphere4/r41/vsp_41_config_max.pdf ESX requires the amount of memory per CPU and per node to be balanced or it does not start. For more information, go to the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5069570 ESX requires the MAX5 memory configuration in UEFI to be in Non-pooled mode. The recommendation is that you review all Retain Tips related to this system and the OS that you are installing. See the following website: http://ibm.com/support/entry/portal/Problem_resolution/Hardware/Systems/System_ x/System_x3850_X5 For further tuning information, see VMware vCenter Server Performance and Best Practices for vSphere 4.1 at the following website: http://www.vmware.com/resources/techresources/10145
291
The following websites contain additional OS installation guides: Installing Red Hat Enterprise Linux Version 6 - IBM System x3850 X5 (Type 7145) http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5086423 Installing Red Hat Enterprise Linux Version 5 Update 4 - IBM System x3850 X5, x3950 X5 http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5083917 Consider these useful tips: All system drivers are included in RHEL 5.4. With RHEL 5.4 and if the server has 1 TB of installed physical memory, see the Red Hat Knowledgebase article at the following website: http://kbase.redhat.com/faq/docs/DOC-25412 RHEL 5.5 is required for 8-socket scaling. RHEL has a tendency to assign eth0 to the Emulex 10GbE NIC, which might be undesirable. The recommended action is to either blacklist the module during PXE installs or from within the operating system, hard-code the Ethernet device names based on their MAC addresses. RHEL requires the MAX5 memory configuration in UEFI to be in Pooled mode. We recommend that you review all RETAIN tips that relate to this system and the OS. See the following website: http://ibm.com/support/entry/portal/Problem_resolution/Hardware/Systems/System_ x/System_x3850_X5
292
The recommendation is to review all Retain Tips related to this system and the OS you are installing. See the following website: http://ibm.com/support/entry/portal/Problem_resolution/Hardware/Systems/System_ x/System_x3850_X5
Server firmware
Software that resides on flash memory and controls the lower-level function of server hardware is called server firmware. An IBM System, such as the x3850 X5, runs a number of firmware images that control separate components of the server. This list shows the primary firmware for the x3850 X5: Unified Extensible Firmware Interface (UEFI) Integrated Management Module (IMM) Field-Programmable Gate Array (FPGA) Preboot Dynamic System Analysis (DSA) Additional devices, such as network cards and RAID controllers, also contain their own firmware revisions. Firmware updates are provided by IBM and can be downloaded from the IBM website, including proven firmware from other manufacturers to be applied on IBM systems. We describe several methods of performing firmware updates on IBM eX5 servers in Chapter 9, Management on page 447. Tip: It is a recommended practice to update all System x firmware to the latest level prior to performing an OS or application installation.
Tip: IBM Bootable Media Creator (BoMC) is a tool that simplifies the IBM System x firmware update process without an OS running on the system. More information about this tool is available at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=TOOL-BOMC
Device drivers
Device drivers are software that controls hardware server components on the OS level. They are specific to the OS version, and therefore, critical device drivers are included with the installation media. Device driver updates are provided by IBM, OS vendors, and component device vendors. They are mostly downloadable from each companys support website. Whenever possible, we recommend using the tested and approved driver updates from IBM.
293
Tip: IBM UpdateXpress is a tool that allows the IBM System x firmware and drivers to be updated via the OS. More information about this tool is available at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-XPRESS
Fibre Channel over Ethernet (FCoE) FCoE is the transport, or mapping, of encapsulated FC frames over the Ethernet. The Ethernet provides the physical interface, and FC provides the transport protocol. System setup for FCoE requires Converged Network Adapter (CNA) to pass both network and storage data connected to 10Gb converged network infrastructure. Internet SCSI (iSCSI) iSCSI is an Internet Protocol (IP)-based storage networking standard for linking data storage facilities. The protocol allows clients (called initiators) to send SCSI commands to SCSI storage devices (targets). A hardware initiator might improve the performance of the server. iSCSI is often seen as a low-cost alternative to Fibre Channel. serial-attached SCSI (SAS) SAS uses point-to-point connection and the typical SAS throughput is 6 Gbps full duplex. Where a complex SAN configuration is not necessary, SAS is a good choice, although performance and distance will be limited compared to the other solutions.
295
If you do not have internal drives, disable the onboard SAS RAID Controller by selecting System Settings Devices and IO ports Enable/Disable Onboard Devices and disabling the SAS Controller or Planar SAS. Set the HBA as the first device in the Option ROM Execution Order by selecting System Settings Devices and IO Ports Set Option ROM Execution Order. For legacy OSs only (all OSs except Windows 2008 and SLES 11 SP1), set Legacy Only as the first boot device. Remove all devices, which might not host an OS, from the boot order. The optimal minimum configuration is CD/DVD and Hard Disk 0. For legacy OSs only, set Legacy Only as the first boot device. Enable the BIOS from your HBA. Verify that your HBA can see a logical unit number (LUN) from your storage. Make sure that the LUN is accessible through only one path, either Zoning or LUN masking. After installation, do not forget to install the multipath driver before you set more than one path, if you have more than one path to the LUN.
296
IBM XIV Storage System: Architecture, Implementation, and Usage, SG24-7659 This book describes the concepts, architecture, and implementation of the IBM XIV Storage System: http://www.redbooks.ibm.com/abstracts/sg247659.html IBM Midrange System Storage Hardware Guide, SG24-7676 This book consolidates, in one document, detailed descriptions of the hardware configurations and options offered as part of the IBM Midrange System Storage servers, which include the IBM System Storage DS4000 and DS5000 families of products: http://www.redbooks.ibm.com/abstracts/sg247676.html For more information regarding HBA storage-specific settings and zoning, contact your SAN vendor or storage vendor.
297
System Monitors System Status Virtual Light Path Event Log Vital Product Data
From the System Status pages, you can perform these tasks: Monitor the power status of the server and view the state of the OS. View the server temperature readings, voltage thresholds, and fan speeds. View the latest server OS failure screen capture. View the list of users who are logged in to the IMM. From the Virtual Light Path page, you can view the name, color, and status of any LEDs that are lit on a server. From the Event Log page, you can perform these tasks: View certain events that are recorded in the event log of the IMM. View the severity of events. For more information about the IMM, see 9.2, Integrated Management Module (IMM) on page 449.
298
For more information and to download Electronic Service Agent, go to the following website: https://www-304.ibm.com/support/electronic/portal/navpage.wss?category=5&locale=en _US
299
5. If the working node does not have hard drives installed, add the hard drives with the operating system from the second node. At the boot sequence, you will get a prompt to import the RAID configuration. 6. Reconnect the power cords. 7. Power on the system.
300
Chapter 7.
This chapter provides assistance for making configuration, monitoring, and maintenance decisions for the x3690 X5 server. The information provided is meant to help you make informed decisions; however, the information provided must not be considered as an absolute implementation process. This chapter covers the following topics: 7.1, Before you apply power for the first time after shipping on page 302 7.2, Processor considerations on page 304 7.3, Memory considerations on page 306 7.4, MAX5 considerations on page 311 7.5, PCIe adapters and riser card options on page 316 7.6, Power supply considerations on page 326 7.7, Using the Integrated Management Module on page 327 7.8, UEFI settings on page 337 7.9, Operating system installation on page 346 7.10, Failure detection and recovery on page 369
301
7.1 Before you apply power for the first time after shipping
Before you begin populating your server with all of its processors, memory, and PCI adapters, and before you install an operating system (OS), perform the recommendations provided in this section.
302
Figure 7-2 Location of switch block SW2 and the CMOS memory battery
c. The numbers on the switch block represent the OFF side of the switch. They are located on the side of the switch block that is closest to the front of the server. To clear CMOS, slide switch 1 (shown in Figure 7-3) to the ON position closest to the rear of the server.
Pull the CMOS memory battery for 30 seconds: a. Disconnect the ac power from the server and remove the cover. b. Locate the CMOS battery (shown in Figure 7-2). c. Use your finger to pry up the battery on the side closest to the neighboring IC chip. The battery will easily lift out of the socket. Note: The light path diagnostic (LPD) lights are powered from a separate power source (capacitor) than the CMOS memory. LPD lights will remain lit for a period of time after ac power and the CMOS memory battery have been removed. d. After 30 seconds, insert one edge of the battery, with the positive side up, back into the holder.
303
e. Push the battery back into the socket with your finger and clip it back into place.
7.1.3 Verify that the server will complete POST before adding options
When you have ordered options for your server that have not yet been installed, it is a good idea to ensure that the server will complete POST properly before you start to add your options. Performing this task makes it easier to compartmentalize a potential problem with an installed option rather than having to look at the entire server to try to find a good starting point during problem determination.
304
Battery
Microprocessor 2
Microprocessor 1
305
Memory installed
Operational characteristics
Both processors 1 and 2 are installed Memory installed only on the system board (no memory mezzanine). Minimum of 2 DIMMs. Performance improves as DIMMs are added to evenly populate all ranks on each memory controller. Processing threads assigned to processor 2 will always have a significant drop in performance for memory-intensive tasks. Not an operational configuration for OSs, such as VMware. Performance improves as DIMMs are added to evenly populate all ranks on each memory controller. If processor 2 fails, memory on the memory mezzanine board is unusable. This configuration is only an operational configuration for VMware when both the system board and mezzanine memory board have the exact same total memory installed. For ease of maintenance, the best practice is to have an identical memory configuration on both the system and mezzanine memory board. Performance improves as DIMMs are added to evenly populate all ranks on each memory controller for local memory. Performance is significantly improved when more active memory calls are using local memory on the system board instead of the memory located in the MAX5. In the rare instance of a mezzanine failure, VMware operational requirements can be satisfied if all of the memory is installed in the MAX5. Memory access will be significantly slower if all of the memory is in the MAX5.
Memory installed on both the system board and the mezzanine memory board.
Memory installed on the system board, the mezzanine memory board, and the MAX5.
306
Any memory installed on the optional mezzanine board will not be seen by processor 1 when processor 2 is not installed or operational. For nonuniform memory access (NUMA)-aware OSs, when two processors are installed, you must install the same amount of memory on both the system board and the mezzanine. See Table 4-9 on page 136 for details. The best processor performance can be achieved when memory is installed to support
Hemisphere Mode. When the mezzanine is not installed and both processors are installed,
it is possible to have processor 1 in Hemisphere Mode and not processor 2. In this type of installation, having processor 1 in Hemisphere Mode will improve the memory access latency for both processors. To determine the DIMM population for Hemisphere Mode without the mezzanine board installed, see Table 4-9 on page 136. Consider installing all DIMMs of the same field-replaceable unit (FRU) to avoid conflicts that might prevent NUMA compliance or support Hemisphere Mode. The problems are most likely to occur when various-sized DIMMs are installed. The server supports various-sized DIMMs installed in each rank, but the configuration becomes complex and difficult to maintain. Although OSs that depend on NUMA compliance will inform you when the server is not NUMA compliant, there is nothing to inform you when the processors are not in Hemisphere Mode. It is better to install more smaller DIMMs than fewer larger DIMMs to ensure that all of the memory channels and buffers of each processor have access to the same amount of memory. This approach allows the processors to fully use their interleave algorithms to access memory faster by spreading it over multiple paths.
307
2. When the built-in diagnostics start, the diagnostics start the Quick Memory Test for all of the memory in the server, as shown in Figure 7-6 on page 309. You can stop the Quick Memory Test at any time and run a Full Memory Test, which runs the same test patterns multiple times and takes five times longer than the Quick Memory Test. The only time that you want to run the full memory test is if you have an intermittent memory problem that you are trying to isolate. Because the server identifies which specific DIMMs are experiencing excessive single-bit failures, it is far more efficient to swap reported DIMMs with similar DIMMs inside the server and see if the problem follows the DIMMs, stays with the memory slots, or simply goes away because the DIMMs were reseated.
308
3. The quick diagnostics continue to run, reporting each test that the quick diagnostics are currently performing and the length of time that it will take to complete that test. If an error occurs, the quick diagnostics stop and indicate the memory errors encountered before progressing into more advanced diagnostics, as shown in Figure 7-7.
You can terminate the diagnostics at any point by pressing Esc. Important: Never warm-boot the server while running the built-in diagnostics. Several built-in functions that are used in the normal operation of the server are disabled during diagnostics to get direct results from the hardware. Only a normal exit from diagnostics or a cold boot of the server will re-enable those functions. Failure to perform this task correctly will cause the server to become unstable. To correct this problem, simply power off the server and power it back on.
309
Figure 7-8 shows the Memory configuration panel with selections for performing memory sparing or memory mirroring, but not both at the same time. Select the desired option and reboot the server. If your memory population order does not support the requested option, the server will report a memory configuration error during the next reboot. See 4.8.2, x3690 X5 memory population order on page 133 for the correct memory population order to support memory mirroring or sparing.
310
Field Programmable Gate Array (FPGA) on the MAX5: The MAX5 also has FPGA firmware that resides on it. The FPGA firmware is updated automatically when the server to which it is attached has its FPGA firmware flashed. The default FPGA loaded onto the MAX5 from the factory is for the System x3850 X5. You are required to reflash the FPGA on your system x3690 X5 after the MAX5 is connected to it. The server might fail to complete POST because of a significant difference in FPGA code between the MAX5 and the server. Correct this problem by flashing the FPGA through the IMM while the server is plugged into power but not powered on.
311
The recommended sequence of updates is shown in Table 7-2 on page 311. To update the server to this firmware level or a higher level, you can use one of the following choices: When you have a compatible OS installed, you can use the UpdateXpress System Packs Installer (UXSPI) tool that is described in 9.11, UpdateXpress System Pack Installer on page 511. Regardless of the version of OS or when an OS is not installed, you can use the Bootable Media Creator (BOMC) tool that is described in 9.12, Bootable Media Creator on page 514. After establishing a network connection and logging in to the IMM, use the firmware update function, as described in 9.13, MegaRAID Storage Manager on page 521. UEFI update is a two-step process: The UEFI flash is a two-step process. When the UEFI flash has completed, the server must be warm-booted to at least the F1-Setup panel to allow the second half of the UEFI flash to complete. Failure to perform the required warm boot will result in a corrupt version of the UEFI, which will force the server to boot into the recovery page of the UEFI until corrected by following the UEFI recovery process in the Problem Determination and Service Guide (PDSG). For the latest firmware requirements using the MAX5, see RETAIN tip H197572: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5085756
312
There are only two QPI cables connecting the two QPI ports of the x3690 X5 to two of the four QPI ports on the MAX5. Figure 7-10 shows how the cables must be connected between the QPI ports of the two units.
Rack rear
x3690 X5 MAX5
The QPI cables are keyed to only be inserted one way. A quick visual check for cable orientation is to look for the 2U QPI or 1U QPI labels on the cable. The labels, along with the blue retainer release tab, are placed on what will become the visible top of the cables when they are installed correctly. The ends of the cables are labeled to indicate which end to insert into the correct equipment. The 2U QPI end of the cable plugs into the x3690 X5. The 1U QPI end of the cable plugs into the MAX5. The cable end must slide into the port until it clicks into place. You can disengage the retainer that holds the cable in place by pressing on the blue tab on the top of the QPI cable connection. Both cables must be installed, even when only one processor is installed, to allow the MAX5 to be controlled by the server. If one of the cables is detached, the server will not power on or complete POST.
313
Updating FPGA firmware: When attaching the MAX5 for the first time, you must reapply the FPGA firmware using the IMM firmware update tool after the server is plugged into ac power but prior to powering up the server. Do not use Bootable Media Creator or USXPI to update the FPGA until the FPGA firmware is a match between the server and the MAX5. Mismatched FPGA will make the server unstable and it might power off during the flash, corrupting the FPGA. This situation will
result in requiring a hardware replacement, because there is no recovery for corrupt FPGA.
Figure 7-11 shows the back of the x3690 X5 with the MAX5 attached.
Figure 7-11 Lab photo of an x3690 X5 (top) attached to an eX5 MAX5 (bottom)
Important: With the server or MAX5 plugged into ac power and the scaled units power turned off, the QPI cables still have active dc power running through them. You must only
plug in or unplug the QPI cable when the MAX5 and the server are not plugged into ac power. Failure to follow this rule will result in damaged circuits on either the MAX5 memory
board or the system board of the server.
314
Figure 7-12 Blue release tabs for the memory tray cam levers
4. You can then pull the memory tray out about 30% before it stops. This design allows you to get a better grip on the tray on either side and then use your fingers to push in another set of blue release tabs on either side of the tray, as shown in Figure 7-13.
Figure 7-13 Final release tab for removing the MAX5 memory tray
5. Slide the memory tray completely out and place it on a flat work surface to work on the memory. Power: You must remove the ac power from both the server and the attached MAX5 before removing the memory tray. The FPGAs of both the server and the MAX5 are still active components when the server is powered down but not removed from utility power. FPGA components exist on both the server and the MAX5. Removing the memory board with ac power still active damages the FPGA components of both the server and the MAX5. For the memory population order, see 4.8, Memory on page 131. After the MAX5 is installed and configured properly, you can confirm a successful link between the server and the MAX5 when both units power on and off as one complete unit and the memory in the MAX5 can be seen by the server. You will also see the following message during POST: System initializing memory with MAX5 memory scaling
315
PCIe adapters connect to the processors via the I/O hub. The purpose of the I/O hub is to combine the various data streams from each of the PCIe slots into a single aggregate link to the processors by using a dedicated QPI link to each processor. The data transfer rate of the QPI link is negotiated between the processor and the I/O hub. Table 4-5 on page 130 shows the QPI link speeds based on the types of installed processors. The I/O hub supports the highest QPI link speed on the table, which is 6.4 gigatransfers per second (GT/s). You can adjust the QPI link speed to conserve power by booting into F1-Setup, then selecting System Settings Operating Modes, and setting QPI Link Frequency to values other than the default Max Performance, as shown in Figure 7-14 on page 317.
316
317
Figure 7-15 PCIe slot speed selection panel to force Gen1 compliance
318
Another possible solution to this problem is booting the server in Legacy Only mode. This mode allows both non-UEFI-aware OSs and PCIe adapters to function as though they are on a non-UEFI server. Many of the advanced memory addressing features of the UEFI environment will not be available to the OS in this mode. Booting the server to an OS in this mode allows you the ability to apply firmware updates to a Gen1 adapter that it not recognized in an UEFI environment. To enable this feature from within F1-Setup, select Boot Manager Add Boot Option, scroll down until Legacy Only is displayed in the list of options, and select it, as shown in Figure 7-17 on page 320. If Legacy Only is not listed, it has already been added to the boot manager.
319
If you add Legacy Only to the boot manager, you must also change the boot sequence to place Legacy Only at the top of the boot sequence. To change the boot order from within the Boot Manager panel, select Change Boot Order. Figure 7-18 shows the Change Boot Order panel. Use the arrow keys to select Legacy Only and move it to the top of the list. Press Enter to confirm your changes.
320
Microsoft Windows 2008 x64: When installed on an UEFI server, Microsoft Windows 2008 x64 will install Microsoft Boot Manager as part of the boot sequence. Regardless of how you change the boot sequence in the boot manager, Microsoft Boot Manager will always be at the top of the sequence. When you install this same OS with Legacy Only enabled, Microsoft Boot Manager is not installed as part of the boot manager. Removing the Legacy Only option from the boot manager will prevent the server from booting into the installed Windows 2008 x64.
321
Of all of the I/O adapters that can be installed on the server, the ServeRAID and 6 Gbps SAS controllers, when managing solid-state drives (SSDs), are the only adapters that can approach the theoretical limits of the x8 PCIe 2.0 slot. When you use SSD drives, connect no more than 8 SSD drives to a single controller for the best performance of the SSD drives. Only use x8 slots to host the controllers that manage your SSD drives (that is, not slot 4, because it is an x4 slot). A single ServeRAID controller managing a single 4-drive SAS hard disk drive (HDD) array will function within the theoretical limits of an x4 PCIe slot. In this case, the mechanical nature of the HDD drives will limit the maximum throughput of data that passes through the PCIe slot. The dual-port 8 Gbs Fibre Channel, 10Gbs Ethernet, and 10 Gbs Converge Network Adapters (CNAs) are all capable of approaching the theoretical limits of an x4 PCIe 2.0 slot and might perform better in an x8 PCIe 2.0 slot.
322
When booting from SAN with multiple paths for redundancy, you need to enable the legacy option ROM for both HBAs.
323
Other PCI adapters can have their boot option ROM disabled from within their configuration panels. To access individual adapter configuration panels from F1-Setup, select System Settings Adapters and UEFI Drivers. Figure 7-21 shows the available selections on this panel.
324
To enter the configuration of a specific adapter, select the PciROOT directly beneath the adapter name. When you have multiple controllers of the same type, selecting the PciROOT of any of the same adapter types will select all of them and present a panel that allows you to select the specific adapter from within the configuration routine of the adapter type. For example, Figure 7-22 demonstrates this process when two ServeRAID adapters are installed in the server.
Figure 7-22 LSI Adapter Selection panel from within the LSI configuration window
For adapters that are similar to the ServeRAID or Fibre Channel HBAs, look for the controller configuration or properties. Inside the controller properties for the ServeRAID controllers, you see a panel similar to the panel that is shown in Figure 7-23.
Figure 7-23 ServeRAID BIOS Configuration Utility Controller Properties: Disabling Boot ROM
Although not obvious by the description, disabling the Controller BIOS only disables the Boot ROM execution during POST. All of the other operating characteristics of the adapter BIOS and firmware remain intact.
325
326
Ensure that half of the power supplies of both units are plugged into one utility power source and the remaining half are plugged into a separate utility power source. This approach eliminates the possibility of a single breaker or circuit fault taking down the entire server. Think of the power supplies in your server like shock absorbers in a car. They are designed to absorb and overcome a wide variety of power conditions that can occur from an electric utility company, but like shock absorbers on a car, they will eventually begin to fail when fed a steady diet of unstable power. The times of their failures will most likely not coincide with a planned maintenance window. Therefore, ensure that the two halves of power supplies are plugged into two separate UPS sources to filter out all of the moderate to severe power fluctuations that occur. Figure 7-25 depicts the power supplies of the combined x3690 X5 and the MAX5 and to which UPS power sources they must be attached to be fully redundant.
Two optional power supplies for the x3690 X5. Plug the left three power supplies into UPS source 1. The two power supplies of the attached eX5 MAX5.
Plug the right three power supplies into UPS source 2. Two primary power supplies for the x3690 X5.
Figure 7-25 Recommended power cabling for the x3690 X5 and the MAX5
327
328
panel of F1-Setup. In F1-Setup, you can access this panel by selecting System Setting Integrated Management Module Network Configuration, as shown in Figure 7-27.
When configured as Dedicated, you connect to the network via the system management port. As shown in Figure 7-28, the port is located from the rear of the server on the left side of the video port. Using this port allows for easier separation of public and management network traffic. You can separate public and management network traffic when you connect your public network port to switch ports that belong to a public access virtual LAN (VLAN) and the management port is connected to a switch port that is defined by a separate management VLAN.
When configured as Shared, the server shares network traffic on the second onboard Ethernet port, which is the port that is closest to the power supply, as shown in Figure 7-29 on page 330.
329
Figure 7-29 Onboard Ethernet port used when IMM Network Interface is Shared
Although this design eliminates a physical switch port and patch cable configuration, both the media access control (MAC) address for the second Ethernet port and the MAC address for the IMM will address through this network port. Therefore, at least two separate IP addresses are assigned to the same physical port. Sharing the port also prevents you from configuring the two onboard Ethernet ports in a network team using 802.3ad load balancing. Using 802.3ad load balancing results in dropped packets for the IMM MAC address. Smart load balancing and failover are still available network teaming options. However, keeping the public traffic from the management traffic become more difficult. To maintain this separation, you must use network teaming software to establish a VLAN to be used by the server to send public tagged traffic to the network switch. You must configure the switch port as a trunk port to support both the public-tagged VLAN traffic, plus the untagged traffic for the management. You must define the management VLAN as the native VLAN on the switch port, so that its untagged traffic from the switch will be accepted by the IMM MAC and dropped by the second Ethernet ports MAC. Although this configuration is more complex, it is a common practice in virtualized servers running on a common set of host computers. Although the IMM uses a dedicated RISC processor, there are limitations to the amount of network traffic to which the IMM can be exposed before complex functions, such as booting from a remote DVD or USB storage, become unreliable because of timing issues. Although the OS has all of the necessary drivers in place to deal with these timing issues, the UEFI is not as tolerant. Therefore, for maintaining secured access, keep the IMM on a separate management network.
330
auto-negotiate medium-dependent interface crossover (MDIX) when they auto-negotiate speed and duplex settings. If you can ping the IMM, you have a good direct network link. If the web session fails, perform the following steps: 1. Try a separate web browser. 2. Directly access the Integrated Management Module configuration panel and reset the IMM in F1-Setup by selecting System Settings Integrated Management Module Reset IMM. Wait about 5 minutes for the IMM to complete enough of its reboot to allow you to ping it. This IMM reset will have no effect on the OS that is running on the server. 3. Try clearing your web browser cache. 4. Load factory default settings back on the IMM through the F1-Setup by selecting System Settings Integrated Management Module Reset IMM to Defaults. The IMM will need to be reset again after the defaults are loaded. 5. Contact IBM support.
System Status
The first panel that you see after completing the login and the session time-out limits panel is the System Status panel, as shown in Figure 7-30 on page 332. This panel provides a quick summary review of the hardware status of the server. A green circle indicates all of the hardware is functioning well from a strictly hardware point of view. The IMM can check on the status of server components, ServeRAID controllers, and PCIe interfaces to most PCIe adapters. It does not check on the functional status of most PCIe adapters with regard to their hardware connections to external devices. You need to refer to the system event log from within the OS or the switch logs of the network and Fibre Channel switches to which the server is connected to resolve connectivity issues there.
331
When an actual hardware error is detected in the server, the system status is represented by a red X and the System Health Summary provides information about the errors that are presently unresolved in the server, as shown in Figure 7-31 on page 333.
332
333
Servers are not typically located near the people who manage them. To help you see the event from a remote location, the IMM provides the capability to view all tiers of light path diagnostics, as shown in Figure 7-33.
334
Remote control
Certain problems require that you view OS logs or enter UEFI via F1-Setup to detect them or fix them. For remotely managed servers, you have the Remote Control feature of the Integrated Management Module. Figure 7-35 shows the available options for starting a Remote Control session.
Figure 7-35 Integrated Management Module remote control session start-up window
335
The Remote Control feature provides the following advantages: It gives you the same functionality that you have with a keyboard, mouse, and video panel directly connected to the server. You can encrypt the session when using the Remote Control feature over public networks. You have the ability to use local storage or ISO files as mounted storage resources on the remote server that you are using. These storage resources can be unmounted, changed, and remounted throughout the session, as needed. Combined with the Power/Restart functions of the IMM, you can power down, reboot, or power on the server while maintaining the same remote control session. Depending on the application that you are accessing through the IMM Remote Control feature, you might find that the mouse pointer is difficult to control. You can fix this problem in the Video Viewer by selecting Tools Single Cursor, as shown in Figure 7-36.
Figure 7-36 Fixing the mouse pointer in the Remote Control Video Viewer
336
Figure 7-38 on page 338 shows the UEFI System Configuration and Boot Management panel.
337
Figure 7-38 UEFI System Configuration and Boot Management settings panel
338
The MAX5 Memory Scaling setting provides two options to determine how the system will present the memory capacity in the MAX5 unit to the running OS: Non-Pooled The default option splits the memory in the MAX5 and assigns part of it to each of the installed processors. Pooled This option presents the additional memory in the MAX5 as a separate pool of memory without its being assigned to any particular processor.
339
You can obtain a complete list of supported Windows Server OSs at IBM ServerProven: http://ibm.com/systems/info/x86servers/serverproven/compat/us/nos/microsoft.html
340
As seen in Figure 7-40, the Operating Modes settings provide the user with several pre-configured Processor and Memory settings that are defined by system characteristics, whether it is Efficiency, Acoustic, Performance, or Custom. The default is Custom. Acoustic Mode emphasizes power-saving operations generating less heat and noise, at the expense of performance. Efficiency Mode provides the best balance of performance and power consumption. Performance Mode provides the best system performance at the expense of power efficiency. Custom Mode allows individual settings of the performance-related options. Table 7-4 on page 342 displays comparisons of the operating modes for the IBM System x3690 X5. We recommend the available default Custom Mode setting, because it provides high system performance with acceptable power consumption. Also, consider Efficiency Mode to gain the best performance per watt operations. Acoustic Mode is an available option when it is necessary to operate the system with minimal power consumption. Use Performance Mode when you want to get the best possible performance from the system.
341
Table 7-4 Comparison of operating modes Settings Memory Speed CKE Low Power Proc Performance States C1 Enhance Mode CPU C-States QPI Link Frequency Turbo Mode Turbo Boost Power Optimization Efficiency Power Efficiency Enabled Enabled Enabled Enabled Power Efficiency Enabled Power Optimized Acoustic Minimal Power Enabled Enabled Enabled Enabled Minimal Power Disabled None Performance Max Performance Disabled Enabled Disabled Disabled Max Performance Enabled Traditional Custom (default) Max Performance Disabled Enabled Enabled Enabled Max Performance Enabled Power Optimized
A complete description of power-related and performance-related system settings is available in 2.7.3, Performance-related individual system settings on page 43.
342
Setting Hyper-Threading Execute Disable Bit Intel Virtualization Technology QPI Link Frequency IMM Thermal Mode CKE Low Power Memory Speed Page Policy Mapper Policy Patrol Scrub Demand Scrub
Virtualization Enabled Enabled Enabled Max Performance Performance Disabled Max Performance Closed Closed Disabled Enabled
Low latency Disabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled Disabled
HPC Enabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled Disabled
343
3. Select Delete Boot Option from the Boot Manager menu. 4. Select all items that you do not want to boot from by using the Spacebar to select them. When you have selected all of the items, scroll down to the end of the page using the down arrow key and select Commit Changes, as shown in Figure 7-42 on page 345. In the example in Figure 7-42 on page 345, Hard Disk 0 has also been selected for removal. The system is running Windows Server 2008 R2 with GUID Partition Table (GPT) disk support from UEFI, and therefore, Windows Boot Manager is used to boot the OS, not Hard Disk 0.
344
5. Press Esc when you have finished to return to the Boot Manager menu. 6. Select Change Boot Order from the Boot Manager menu. 7. Press Enter to make the device list active. 8. Use the up and down arrow keys to navigate to the device for which you want to change the order. After you highlight the device, use the - or Shift and + keys to move the device up or down the list. You can then perform the same actions to move other devices up or down the list. Press Enter when done. 9. Use the down arrow key to highlight Commit Changes and press Enter to commit the changes that you have made. 10.Press Esc to return to the Boot Manager menu. 11.Press Esc again to exit to the System Configuration and Boot Management menu. 12.Press Esc again to exit the UEFI and press the Y key to exit and save any changes that you have made. The x3690 X5 then proceeds to boot normally. Tip: It is common practice to place CD/DVD Rom higher than the default OS in the boot order list to accommodate system tools media that require booting from CD/DVD. More information about editing and cleaning the boot options is available in 7.5.3, Cleaning up the boot sequence on page 322.
345
346
In a typical configuration, these server functions are required to allow the server to perform PXE network boot and installation: DHCP server, providing IP address Trivial File Transfer Protocol (TFTP) server, providing initial boot and installation environment File server, providing OS installation files Use the guides that are provided in the following list to perform the PXE network boot installation that is specific to each OS: Windows Server 2008 R2 http://technet.microsoft.com/en-us/library/cc772106(WS.10).aspx Red Hat Enterprise Linux 6 http://docs.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/6/html/Installation_ Guide/index.html SUSE Linux Enterprise Server 11 http://www.novell.com/documentation/sles11/book_sle_deployment/?page=/documenta tion/sles11/book_sle_deployment/data/pre_sle.html VMware vSphere ESXi 4.1 http://www.vmware.com/pdf/vsphere4/r41/vsp_41_esxi_i_vc_setup_guide.pdf
347
Tip: When using Microsoft Internet Explorer (IE) version 7 or 8, if the Remote Control window does not open and you get no warning that a pop-up was blocked, it is possible that IE advanced security settings are blocking the Java application from starting. A work-around for this issue is to hold the Ctrl key while clicking Start Remote Control in Single User Mode or Start Remote Control in Multi-User Mode in the IMM Remote Control window. RETAIN tip H196657 describes other work-arounds: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5083262 Figure 7-45 on page 349 shows the Virtual Media Session interface where we can mount a selected installation image.
348
Follow this sequence to mount an image as virtual CD/DVD media at the controlled system: 1. 2. 3. 4. Click Add Image. Choose an image in .iso or .img format. Check the box under Map of the image to mount. Click Mount Selected.
Simply click Unmount All to unmount the image. You can also mount a local Floppy, CD/DVD, or Removable Disk to a remote system using similar methods.
349
For the full details, see the IBM ServerGuide Scripting Toolkit, Windows Edition Users Reference and IBM ServerGuide Scripting Toolkit, Linux Edition Users Reference at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-TOOLKIT
Windows installation
This section describes the process to install the ServerGuide Scripting Toolkit, create a deployment image for Windows Server 2008 R2 Enterprise Edition, and copy this image to a USB key for deployment. To configure a USB key for deployment, you need the following devices: A system running Windows Vista, Windows Server 2008, Windows 7, Windows Server 2008 R2, Windows 2.1 Preinstallation Environment (PE), or a Windows 3.0 PE session A USB key with a storage capacity at least 64 MB larger than your Windows PE image, but not less than 4 GB
Procedure
We use this procedure: 1. Install the ServerGuide Scripting Toolkit. 2. Create a deployment image. 3. Prepare the USB key.
350
3. Choose the OS type that you want and click Next, as shown in Figure 7-47.
4. Insert the correct OS installation media or select the folder that contains the installation files for the source, as shown in Figure 7-48 on page 352. If necessary, modify the target and click Next.
351
5. When the file copy process completes successfully, as shown in Figure 7-49, click Finish.
6. Open a command prompt and change to the C:\sgshare\sgdeploy\SGTKWinPE directory. Use the following command to create the Windows installation image: SGTKWinPE.cmd ScenarioINIs\Local\Win2008_R2_x64_EE.ini 7. When the process completes, as shown in Figure 7-50 on page 353, your media creation software is started to create a bootable media from the image. Cancel this task.
352
18:26:21 - Creating the WinPE x64 ISO... 18:27:07 - The WinPE x64 ISO was created successfully. *** WinPE x64 ISO: 4_EE\WinPE_x64.iso c:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x6
18:27:07 - Launching the registered software associated with ISO files... *** Using ISO File: 64_EE\WinPE_x64.iso c:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x
18:27:08 - The WinPE x64 build process finished successfully. SGTKWinPE complete. Figure 7-50 Build process is finished
353
2. Use diskpart to format the USB key using FAT32. All files on the USB key will be deleted. At the command prompt, type the commands that are listed in Figure 7-51.
C:\>diskpart Microsoft DiskPart version 6.1.7600 Copyright (C) 1999-2008 Microsoft Corporation. On computer: DISKPART> list disk Disk ### -------Disk 0 Disk 1 Disk 2 Status ------------Online Online Online Size ------271 GB 135 GB 7839 MB Free Dyn ------- --0 B 0 B 0 B Gpt --*
DISKPART> select disk 2 Disk 2 is now the selected disk. DISKPART> clean DiskPart succeeded in cleaning the disk. DISKPART> create partition primary DiskPart succeeded in creating the specified partition. DISKPART> select partition 1 Partition 1 is now the selected partition. DISKPART> active DiskPart marked the current partition as active. DISKPART> format fs=fat32 100 percent completed DiskPart successfully formatted the volume. DISKPART> assign DiskPart successfully assigned the drive letter or mount point. DISKPART> exit Figure 7-51 Using diskpart to format the USB memory key
3. Copy the contents from C:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x64_EE\ISO to the USB key. The USB key now contains the folders and files that are shown in Figure 7-52.
4. Boot the target system from the USB key. The deployment will execute automatically.
354
RAID controller: If the target system contains a RAID controller, RAID will be configured as part of the installation.
355
See 4.13, Integrated virtualization on page 174 for ordering information for the integrated virtualization option.
356
We recommend that you map SAN storage logical drives after the OS installation process completes, except in Boot from SAN configurations. You can access more information about the Windows Server 2008 R2 installation and configuration at the following website: http://www.microsoft.com/windowsserver2008/en/us/product-documentation.aspx If Windows Server 2008 R2 cannot be started and the system has 1 TB or more of memory installed, try to apply the Microsoft KB98059 that can be downloaded from the following website: http://support.microsoft.com/kb/980598 If a system with 128 GB or more of memory installed and running Windows Server 2008 R2 with Hyper-V receives error message STOP 0x000000A, try to apply the Microsoft KB979903 Hotfix that can be downloaded from the following website: http://support.microsoft.com/kb/979903 We recommend that you set Flow Control to Enabled in the Broadcom NetXtreme II Properties to ensure that no data transmission issue exists between servers with large performance gaps. The following steps describe this task: a. Select Start Administrative Tools Computer Management Device Manager. b. Expand Network adapters. c. Double-click Broadcom BCM5709C NetXtreme II GigE. d. Select Advanced Flow Control. e. Set Value to Rx & Tx Enabled and click OK. You can download Windows Server 2008 R2 Trial Software from the following website: http://www.microsoft.com/windowsserver2008/en/us/trial-software.aspx
357
7.9.4 Red Hat Enterprise Linux 6 and SUSE Linux Enterprise Server 11
IBM System x3690 X5 with MAX5 memory expansion fully supports the latest version of two major Linux server distributions: RHEL6 from Red Hat and SLES11 from Novell. The installation of RHEL6 and SLES11 is a straightforward process, because all of the required drivers for System x3690 X5 are included in the kernel. However, use the following general tips when installing either of these OSs: The x3690 X5 supports only the 64-bit version of RHEL6 and SLES11. RHEL6 and SLES11 are UEFI-compliant OSs. Each product adds its boot file on top of the systems UEFI boot order. We recommend not to attach SAN storage to the server during the installation process. Attach the SAN storage after the OS is up and running. You can access more information about the installation and configuration of Red Hat Enterprise Linux at the following website: https://access.redhat.com/knowledge/docs/manuals/Red_Hat_Enterprise_Linux/ You can download Red Hat Enterprise Linux 6 evaluation software from the following website: https://access.redhat.com/downloads/ You can access more information about the installation and configuration of SUSE Linux Enterprise Server at the following website: http://www.novell.com/documentation/suse.html You can download SUSE Linux Enterprise Server 11 evaluation software from the following website: http://www.novell.com/products/server/eval.html
358
4. Press Enter after doing the modification. Figure 7-54 shows the modified boot menu that allows the ESXi installation to proceed with the enabled interleaved NUMA nodes parameter.
VMware VMvisor Boot Menu ESXi Installer Boot from local disk
> mboot.c32 vmkboot.gz allowInterleavedNUMAnodes=TRUE --- vmkernel.gz --- sys.vg z --- cim.vgz --- ienviron.vgz --- install.vgz
Without this additional parameter, the VMware ESXi 4.1 installation fails with the Interleaved NUMA nodes are not supported error message, as shown in Figure 7-55.
Figure 7-55 Interleaved NUMA nodes are not supported error panel
5. Continue with the ESXi 4.1 installation. 6. To successfully first boot, the allowInterleavedNUMAnodes parameter must be set to TRUE in the Loading VMware Hypervisor panel. Perform the following steps: a. Press Shift+O while the gray bar progresses to add a boot parameter. b. Type the following command at the prompt: esxcfg-advcfg -k TRUE allowInterleavedNUMAnodes c. Press Enter to continue the boot process. Figure 7-56 on page 360 shows the edited ESXi Loading VMware Hypervisor panel.
Chapter 7. IBM System x3690 X5
359
7. Set the allowInterleavedNUMAnodes parameter to TRUE in all subsequent system reboots. You are required to have VMware vSphere Client to perform the following steps: a. After the ESXi hypervisor is running, connect to the System x3690 X5 using VMware vSphere Client. b. Click the Configuration tab. c. Choose Advanced Settings in the Software section. d. Click VMkernel. e. Select the VMkernel.Boot.allowInterleavedNUMAnodes check box to enable this parameter and click OK. Figure 7-57 on page 361 shows the required option to set in the Software Advanced Settings panel.
360
In addition to these required steps, use the tips in the following list for the VMware vSphere ESXi installation on the x3690 X5: VMware vSphere 4.1 is the last release to support ESX; all subsequent releases of vSphere will only support ESXi. Therefore, we recommend that you use the ESXi hypervisor rather than ESX. The following websites provide more information about this topic, as well as a comparison between ESX and ESXi: http://www.vmware.com/products/vsphere/esxi-and-esx/ http://kb.vmware.com/kb/1023990 Separate VMware vSphere editions support separate amounts of physical memory. To support the maximum memory capability of the memory-expanded x3690 X5, which is 1 TB, the Enterprise Plus Edition is necessary. Other editions only support up to 256 GB of physical memory. You can see a comparison of vSphere editions at this website: http://www.vmware.com/products/vsphere/buy/editions_comparison.html Installing VMware ESXi on System x3690 X5 requires that an equal amount of memory is installed for each system processor. See MAX5 memory population order on page 137 for guidelines about memory installation on the x3690 X5 scaled with MAX5. ESXi installation fails with the NUMA node 1 has no memory error message if a processor in the system has no memory installed.
361
Create a drive RAID array before installation. The ServeRAID management interface is accessible via UEFI settings, by selecting System Settings Adapters and UEFI Drivers. When running the ESXi hypervisor from a USB flash drive, plug it in to the internal USB slot to prevent physical disruption. The internal USB slot is located on the x8 low profile PCI riser card. ESXi formats all local disks in the server when it is initially booted. Remove any local and SAN-attached disks from the system if they contain important data. If ESXi is installed to a USB flash drive, you must add USB Storage to the system boot option. Placing it on the top of the boot order list decreases system boot time. VMware vSphere 4.1 offers no support for UEFI and therefore is categorized as Legacy OS. Moving Legacy Only up in the boot order list decreases system boot time. If a USB keyboard stops functioning at any time during the ESXi installation, try plugging it in to a separate USB port. You can access more information about the VMware vSphere installation and configuration at the following website: http://www.vmware.com/support/pubs/
362
3. The Boot Options line appears on the panel. Add the allowInterleavedNUMAnodes=TRUE parameter at the boot options line, so that the line reads in the following manner: Boot Options initrd=initrd.img debugLogToSerial=1 vmkConsole=false mem=512M quiet allowInterleavedNUMAnodes=TRUE The edited result looks like Figure 7-59 on page 364. Press Enter to proceed.
363
4. Proceed with ESX installation until you get to the Setup Type page, as shown in Figure 7-60. Click Advanced setup, clear Configure bootloader settings automatically (leave checked if unsure), and click Next.
5. Continue with the installation until you get to the Set Bootloader Options page, as shown in Figure 7-61 on page 365, and type the following parameter in the Kernel Arguments field: allowInterleavedNUMAnodes=TRUE
364
Server firmware
Software that resides on flash memory and controls the lower-level functions of the server hardware is known as server firmware. An IBM System x server, such as the x3690 X5, can run a number of types of firmware that are in charge of the server components. The following types of firmware make up the primary firmware for the IBM System x3690 X5: Unified Extensible Firmware Interface (UEFI) Integrated Management Module (IMM) Field-Programmable Gate Array (FPGA) Preboot Dynamic System Analysis (DSA) Other firmware controls each server component and corresponds to the manufacturer and model of the device, such as firmware for NIC, RAID controller, and so on. IBM provides the firmware updates. You can download the firmware updates from the IBM website, including proven firmware from other manufacturers to apply on IBM systems. We describe several methods of performing firmware updates on IBM eX5 servers in 9.10, Firmware update tools and methods on page 509. Tip: It is a recommended practice to update all System x firmware to the latest level prior to performing OS and application installation.
365
Useful links: IBM Bootable Media Creator (BoMC) is a tool that simplifies the IBM System x firmware update process without an OS running on the system. You can obtain more information about this tool at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=TOOL-BOMC You can access the IBM System x3690 X5 firmware and drivers quickly at the following website: http://www.ibm.com/support/fixcentral/systemx/quickorder?parent=ibm~Systemx3690 X5&product=ibm/systemx/7148&&platform=All&function=all&source=fc The following website contains a guide that describes firmware update best practices for UEFI-based IBM System x servers: http://ibm.com/support/entry/myportal/docdisplay?lndocid=MIGR-5082923
Device drivers
Device drivers are software that controls hardware server components on the OS level. They are specific to the OS version, so critical device drivers are included with the installation media. Device driver updates are provided by IBM, OS vendors, and component device vendors. They are primarily downloadable from each companys support website. Whenever possible, we recommend acquiring tested and approved driver updates from IBM. You can get more information at these useful links: The Windows Server installation process using IBM ServerGuide performs IBM driver updates after the OS installation is complete. You can access this tool at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-GUIDE IBM UpdateXpress is a tool that simplifies the IBM System x firmware and driver update running a supported OS. You can access more information about this tool at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-XPRESS You can access the IBM System x3690 X5 firmware and drivers quickly at the following website: http://www.ibm.com/support/fixcentral/systemx/quickorder?parent=ibm~Systemx3690 X5&product=ibm/systemx/7148&&platform=All&function=all&source=fc
366
Table 7-6 Internet links to support and downloads Vendor IBM IBM Product Systems support x3690 X5 support Address http://www.ibm.com/systems/support/ http://www.ibm.com/support/entry/portal/Ove rview/Hardware/Systems/System_x/System_x369 0_X5 http://www.ibm.com/support/entry/portal/doc display?lndocid=SERV-GUIDE http://www.ibm.com/support/entry/portal/doc display?lndocid=SERV-XPRESS http://www.ibm.com/support/entry/portal/doc display?lndocid=TOOL-BOMC http://support.microsoft.com/ph/14134 https://www.redhat.com/support/ http://www.novell.com/support/ http://downloads.vmware.com/d/ Installation tool Firmware and driver update tool Firmware update tool Docs, drivers, and OS update Docs, drivers, and OS update Docs, drivers, and OS update Docs, drivers, and OS update Available support Docs, firmware, and drivers
ServerGuide UpdateXpress Bootable Media Creator Windows Server RHEL SLES vSphere
367
Fibre Channel over Ethernet (FCoE) This relatively new interface was introduced to provide an easier method for FC storage environments to take advantage of the high performance 10 Gb Ethernet interface. Data communication uses FC protocol over a 10 Gb Ethernet interface. System setup for FCoE requires a Converged Network Adapter (CNA) on the servers to pass both network and FC storage data via 10 Gb network infrastructure. The attached storage system, for example, IBM System Storage DS3400, DS4000 series, DS5000 series, and so on, requires the available FC interface. FCoE competes with iSCSI to be the predominant SAN protocol that works on the Ethernet interface. Serial Attached SCSI (SAS) For smaller environments not requiring complex SAN configuration, using a SAS cluster is a more cost-effective alternative to an FCoE or iSCSI. Although performance, manageability, and distance are limited compared to the other solutions, servers can access their individual portions of the storage system. Typical requirements are the SAS HBA on the servers and a storage system with a SAS interface, for example, IBM System Storage DS3200 and others.
368
IBM Midrange System Storage Implementation and Best Practices Guide, SG24-6363 This book represents a compilation of best practices for deploying and configuring IBM Midrange System Storage servers, which include the DS4000 and DS5000 family of products. http://www.redbooks.ibm.com/abstracts/sg246363.html IBM System Storage DS3000: Introduction and Implementation Guide, SG24-7065 This book introduces the IBM System Storage DS3000, providing an overview of its design and specifications, and describing in detail how to set up, configure, and administer it. http://www.redbooks.ibm.com/abstracts/sg247065.html Implementing an IBM/Cisco SAN, SG24-7545 This book consolidates critical information while discussing procedures and tasks that are likely to be encountered on a daily basis when implementing and IBM/Cisco SAN. http://www.redbooks.ibm.com/abstracts/sg247545.html IBM XIV Storage System: Architecture, Implementation, and Usage, SG24-7659 This book describes the concepts, architecture, and implementation of the IBM XIV Storage System. http://www.redbooks.ibm.com/abstracts/sg247659.html IBM Midrange System Storage Hardware Guide, SG24-7676 This book consolidates, in one document, detailed descriptions of the hardware configurations and options that are offered as part of the IBM Midrange System Storage servers, which include the IBM System Storage DS4000 and DS5000 families of products. http://www.redbooks.ibm.com/abstracts/sg247676.html
369
System event logs System event logs in UEFI and IMM maintain histories of detected error events. IBM Dynamic System Analysis (DSA) IBM DSA is a tool to collect and analyze complete system information to aid in diagnosing system problems. Among other things, DSA result files contain information about hardware inventory, firmware level, and system event logs that assist in system problem determination. You can access the DSA application and user guide at the following website: http://www.ibm.com/support/entry/portal/docdisplay?lndocid=SERV-DSA Tips: Use DSA Portable to run the application without having to install it. Ensure that the Intelligent Platform Management Interface (IPMI) driver and RAID Manager agent are installed in the system prior to running DSA in order to get a complete result file. The DSA result file is stored in C:\IBM_Support\ for Windows and in /var/log/IBM_Support/ for Linux. IBM Electronic Service Agent (ESA) IBM ESA performs system monitoring and automatic problem reporting to IBM when a hardware error is detected. This no-charge software can run as stand-alone software or as an extension of IBM Systems Director. See 9.6, IBM Electronic Services on page 493 for more information about IBM ESA. Problem Determination and Service Guide This guide describes the diagnostic tests, troubleshooting procedures, and an explanation of error messages and error codes to solve many problems without needing outside assistance.
370
You can download the Problem Determination and Service Guide for the IBM System x3690 X5 from the following website: http://www.ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5085205
After the MAX5 expansion is functioning properly, schedule downtime and reattach it to the x3690 X5 using the QPI scalability cable. Follow the guidelines for the MAX5 expansion attachment in 7.4, MAX5 considerations on page 311. Important: The attachment and detachment of MAX5 require that the system is down and that the power is removed from the System x3690 X5.
371
372
Chapter 8.
HX5 2-socket
HX5 4-socket
373
8.1 Before you apply power for the first time after shipping
Before you begin populating your server with processors, memory, and expansion adapters, and before you install an operating system, perform the recommendations that are described in the following sections: 8.1.1, Verifying that the components are securely installed on page 374 8.1.2, Clearing CMOS memory on page 375 8.1.3, Verifying the server boots before adding options on page 376
IBM MAX5
Upper ridge
Figure 8-2 If installing an HX5, remove the power sharing cap, but do not discard it
374
Jumper method
Use the following procedure to clear the CMOS using the jumper method: 1. Power-cycle the real-time clock (RTC). Locate the switch block SW2 in the middle of the system board at the extreme rear of the server, as shown in Figure 8-3. You must remove all of the expansion cards to reach the jumper block.
SW2
0 9 8 7 6 5 4 3 2 1
ON
375
2. The numbers on the switch block represent the OFF side of the switch. Locate switch 5, which is the RTC switch. To clear the CMOS, remove the plastic sheet from the jump block, as shown in Figure 8-3 on page 375. Set switch 5 on switch block SW2 to the ON position for 5 seconds and then return it back to the OFF position.
3. Use your finger to pry up the battery on the side toward the second processor. It is easy to lift the battery out of the socket. Light path diagnostics: The light path diagnostics (LPD) lights are powered from a separate power source (a capacitor) than the CMOS memory. LPD lights remain lit for a period of time after the ac power and the CMOS memory battery have been removed. 4. Wait 30 seconds and then insert one edge of the battery, with the positive side up, back into the holder. 5. Push the battery back into the socket with your finger and it will clip back into place.
installed option rather than having to look at the entire server to try to find a good starting point during problem determination.
The HX5+MAX5 configurations support all of the processors that are listed in 5.9, Processor options on page 192.
377
Table 8-1 shows the recommended installation guide to install the memory in a nonuniform memory access (NUMA)-compliant DIMM installation for the HX5 2-node configuration.
Table 8-1 NUMA-compliant DIMM installation for the HX5 2-node Hemisphere Modea Processor 1 Buffer Buffer Buffer Buffer Buffer DIMM 10 Processor 2 Buffer DIMM 11 DIMM 12 Buffer DIMM 13 DIMM 14 Buffer DIMM 15 x DIMM 21 DIMM 22 DIMM 16 x x x x Buffer DIMM 20 DIMM 23 x x x x DIMM 24 x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
Number of processors
Number of DIMMs
DIMM 1
DIMM 2
DIMM 3
DIMM 4
DIMM 5
DIMM 6
DIMM 7
DIMM 8
4 4 4 4
8 16 24 32
N Y N Y
x x x x x x x x
x x x x x x x x x x x x
DIMM 9 x x x x
x x x x x x x x x x x
a. For more information about Hemisphere Mode and its importance, see 2.3.5, Hemisphere Mode on page 26.
For an HX5 with MAX5 configuration, when installing memory in a MAX5, you do not need to follow a specific order. However, we recommend that you install the memory in the blade before populating the memory in the HX5. After you install the memory in the blade, install the memory in the method that is shown in Table 8-2.
Table 8-2 DIMM installation for the MAX5 for IBM BladeCenter Number of DIMMs Power domain A Buffer DIMM 1 DIMM 2 DIMM 3 DIMM 4 DIMM 5 Buffer DIMM 6 DIMM 7 DIMM 8 DIMM 9 Domain C () Buffer DIMM 10 DIMM 11 DIMM 12 DIMM 13 Power domain B Buffer DIMM 14 DIMM 15 DIMM 16 DIMM 17 Buffer DIMM 18 DIMM 19 Domain C ()
2 4 6 8 10 12 14 16 18 20 22 24
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
378
Before joining blades together or adding a MAX5, check the firmware versions of both the AMM and the individual blades. MAX5: MAX5 only has Field Programmable Gate Array (FPGA) and is not a factor. To locate the current blade and AMM code, you need to log in to the AMM and choose Monitors Firmware VPD. Figure 8-7 shows the current firmware versions of each blade.
Tip: In Figure 8-7, FW/BIOS refers to the UEFI and Blade Sys Mgmt Processor refers to the Integrated Management Module (IMM). Further down the same Firmware VPD page, the AMM firmware version is also displayed, as shown in Figure 8-8.
379
Table 8-3 shows the minimum firmware levels that the HX5 must be running for a 2-node scaled configuration. It also shows the necessary firmware levels to scale with a MAX5.
Table 8-3 Minimum code levels required before scaling HX5 with either MAX5 or a 2-node scale AMM code HX5 single-node HX5 2-node HX5 with MAX5 BPET54L BPET54G BPET54P UEFI 1.00 1.25 1.25 IMM 1.16 1.20 1.21 FPGA 1.00 1.01 1.01
Minimum requirements: The blades and the AMM are required to be, at a minimum, at these code levels in order to scale. However, IBM however recommends that you update to the latest versions, which are available at the following website: http://www.ibm.com/support/us/en/
It is important, in a 2-node scaled configuration, that both HX5s have exactly the same firmware versions for BIOS (UEFI), the blade system management processor (IMM), and FPGA. If they are not at the same version when the QuickPath Interconnect (QPI) wrap card is added, neither blade will boot until the QPI wrap card is removed. See Figure 8-9 on page 380 for an example of the error that you will see.
380
Another characteristic of mismatched firmware is the following message when attempting to create the scale in the scalable complex, as shown in Figure 8-10: The complex state data cannot be read at this time. Check the event log to verify the completion of prior complex operations and try again.
Figure 8-10 Error when trying to create a partition with mismatched firmware on the blade
You must perform the updates on blades individually without the QPI wrap card installed. The AMM might give you the options to flash both blades at one time, as shown in Figure 8-11 on page 381, when the QPI wrap card is installed, but this approach can cause issues and must not be used for the initial setup.
Figure 8-11 Update blade option in blade task: Update Blade Firmware
When the blades are scaled together and operational, you can use any form of updating: Bootable Media Creator, Update Express, AMM, and so on. Both nodes will be updated automatically. Figure 8-12 shows a 2-node scale being updated together from AMM.
For details about these update methods, see the detailed information about firmware update methods that is available in 9.10, Firmware update tools and methods on page 509.
381
Important tip: When updating a 2-node configuration from the AMM, it is important that you do not choose Ignore Complex Groupings. This option treats each blade as an individual blade, and both blades might not be updated, which might then lead to the complex not being able to boot. The blades will not boot until the 2-node scalability card is removed and the blades are reflashed individually.
8.3 Recommendations
In this section, we describe the considerations when creating a 2-node scale or adding a MAX5 to an HX5 blade. These settings help you to make decisions for performance and maximum uptime.
cap.
382
IBM MAX5
Upper ridge
383
Figure 8-15 shows a breakdown of each domain and the current amount of power that is currently being used for each domain.
Click the headings to see the detailed power consumption of each domain.
Figure 8-15 Main view when looking at the AMM Power Management option on a BladeCenter H
You can see more details about the power consumption of each domain by clicking each power domain heading, as shown in Figure 8-15. Figure 8-16 on page 385 shows an example of the details that are displayed.
384
A further consideration is when a 2-node or HX5+MAX5 configuration straddles the two power domains. If you install the primary blade server of a scalable blade complex in blade server bay 7 of a BladeCenter H chassis, the second node or the MAX5 is installed in bay 8. Therefore, the scale is split between two separate power domains. The following situations can occur if there is a power loss to either power domain, depending on how the scalable blade complex is implemented: A loss of power to power domain 1 results in both blade servers in the scalable blade complex going down. If this configuration is a 2-node configuration, you also might get errors on the reboot and the light path diagnostics reporting that a non-maskable interrupt (NMI) error has occurred. This result is normal due to losing processors abruptly. The servers will not come back up until the power is restored or the scalability card is removed. If the scalable blade complex is implemented in stand-alone mode, a loss of power to any power domain will still result in both blade servers going down. The blades will not come back up until either the QPI card is removed or the power is restored.
385
The HX5 offers two methods to create RAID arrays: either via the LSI Setup Utility in UEFI or through the use of IBM ServerGuide.
386
4. Select LSI Logic SAS Controller and press Enter, as shown in Figure 8-18.
You can now create an array. We describe creating a RAID-1 array in 8.4.2, Creating a RAID-1 mirror using the LSI Setup Utility on page 389.
387
3. On the Boot Manager menu, select Add Boot Option, as shown in Figure 4 on page 389.
388
4. Select Legacy Only, as shown in Figure 8-21, and press Enter. The selection disappears when selected. If this option is not available, the Legacy Only option might already be part of the boot list.
5. When the system reboots, the LSI boot information displays. Press Ctrl-C when prompted, as shown in Figure 8-22.
You can now create a RAID array. We describe creating a RAID-1 array in the next section.
389
You can use one of the following options to create arrays: Create IM Volume: Creates an integrated mirror (IM) RAID-1 array RAID 1 drives are mirrored on a 1 to 1 ratio. If one drive fails, the other drive takes over automatically and keeps the system running. However, in this configuration, you lose 50% of your disk space because one of the drives is a mirrored image. The stripe size is 64 Kb and cannot be altered. This option also affects the performance of the drives, because all data has to be written twice (one time per drive). See the performance chart in Figure 5-23 on page 207 for details. Create IME Volume: Creates an integrated mirrored enhanced (IME) RAID-1E array This option requires three drives. This option is not available in the HX5, because the HX5 only has two drives on each node. Create IS Volume: Creates an integrated striping (IS) RAID-0 array The RAID-0 or IS volume, as shown in LSI, is one of the faster performing disk arrays, because the read and write sectors of data are interleaved between multiple drives. The downside to this configuration is immediate failure if one drive fails. This option has no redundancy. In RAID-0, you also keep the full size of both drives. We recommend that you use drives of identical sizes for performance and data storage efficiency. The stripe size is 64 Kb and cannot be altered. In our example, we create a RAID-1 array with two drives. Follow these steps: 1. Select Create IM Volume and press Enter. Figure 8-24 on page 391 appears.
390
v6.04.07.00 (2005.11.03)
Array Type: Array Size (MB) Slot Num 1 0 Device Identifier IBM-ESXSMAY2036RC IBM-ESXSMAY2036RC T106 T106
IM -----RAID Disk [No] [No] Hot Spr [No] [No] Drive Status --------------Pred Fail No No Size (MB) 34715 34715
Esc = Exit Menu F1/Shift+1 = Help Enter=Select Item Alt+N=Next Array C=Create an array
2. In Figure 8-24, click No under RAID Disk to form the array. For each disk, you are prompted to confirm by pressing D, as shown in Figure 8-25. Pressing D ensures that you understand that all data on this disk will be deleted.
LSI Logic MPT Setup Utility Create New Array SAS1064 M v6.04.07.00 (2005.11.03)
Keep existing data, migrate to an IM array. Synchronization of disk will occur. Overwrite existing data, create a new IM array ALL DATA on ALL disks in the array will be DELETED!! No Synchronization performed
Esc = Exit Menu F1/Shift+1 = Help Enter=Select Item Alt+N=Next Array C=Create an array
3. Repeat this deletion for the other drive. After you finish, you see Figure 8-26 on page 392. The RAID Disk column now lists each drive as Yes.
391
v6.04.07.00 (2005.11.03)
Array Type: Array Size (MB) Slot Num 1 0 Device Identifier IBM-ESXSMAY2036RC IBM-ESXSMAY2036RC T106 T106
IM -----RAID Disk [Yes] [Yes] Hot Spr [No] [No] Drive Status --------------Pred Fail No No Size (MB) 34715 34715
Esc = Exit Menu F1/Shift+1 = Help Enter=Select Item Alt+N=Next Array C=Create an array
4. After you have set both disks to Yes, press C to create the array. This step completes creating and initializing a RAID-1 array.
392
If desired, you can install an operating system on each node. Then, you can use this configuration to boot each blade individually if you need increased flexibility. You can use the temporary boot in stand-alone mode in the complex scaling section, as shown in Figure 8-28 on page 394, which allows you to temporarily boot each node separately without having to delete the 2-node configuration.
393
394
Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer Memory buffer
SMI links
QPI links
Figure 8-30 shows where the Speed Burst Card is installed on the HX5.
395
Figure 8-32 on page 397 shows the main UEFI System Configuration and Boot Management panel.
396
You can obtain more general information about UEFI at the following website: http://www.uefi.org/home/ For an explanation of UEFI settings, see 2.7, UEFI system settings on page 36. Generally, the HX5 UEFI requires little configuration. The factory default UEFI settings allow the HX5 to work correctly in a 2-socket, 4-socket, and MAX5-attached configuration. There are, however, a number of UEFI settings that can be adjusted to meet specific needs. You can adjust the UEFI settings, for example, for better performance. This section contains the following topics: 8.5.1, UEFI performance tuning on page 397 8.5.2, Start-up parameters on page 398 8.5.3, HX5 single-node UEFI settings on page 400 8.5.4, HX5 2-node UEFI settings on page 401 8.5.5, HX5 with MAX5 attached on page 401 8.5.6, Operating system-specific settings in UEFI on page 402
397
A server is only as good as its weakest link, so you can have a well-tuned server, but a poorly configured SAN to which it connects. Generally, you need to look at an entire solution, including the SAN and network infrastructure, to fully achieve the best performance from the server. Table 8-4 provides an overview of UEFI settings for the HX5 that address a number of performance scenarios.
Table 8-4 Overview of settings Setting TurboMode TurboBoost Processor Performance states C states C1E state Prefetcher HyperThreading Execute Disable Virtualization Extensions QPI Link Speed IMM Thermal Mode CKE Policy DDR Speed Page Policy Mapper Policy Patrol Scrub Demand Scrub MaxImum performance Enabled Traditional Enabled Virtualization Enabled Power Optimized Enabled Low latency Disabled Automatically disabled Disabled Performance per watt Disabled Automatically disabled Enabled HPC Disabled Automatically disabled Disabled
Disabled Disabled Enabled Enabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled Enabled
Enabled Enabled Enabled Enabled Enabled Enabled Max Performance Performance Disabled Max Performance Closed Closed Disabled Enabled
Disabled Disabled Enabled Disabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled Disabled
Enabled Enabled Enabled Enabled Enabled Enabled Power Efficiency Normal Disabled Max Performance Closed Closed Disabled Enabled
Enabled Enabled Enabled Enabled Disabled Disabled Max Performance Performance Disabled Max Performance Closed Closed Disabled Disabled
398
A simple step to reduce the overall system start-up time is to remove any unnecessary boot devices from the boot order list. In particular, ensure that PXE network boot is removed from the boot order list if it is not used. PXE network boot can increase start-up time substantially, especially if it is also placed in the incorrect order in the boot order list. Use the following steps to remove entries from the Boot Order menu and configure the correct boot order: 1. Log in to the Advanced Management Module and select Blade Tasks Remote Control. Click Start Remote Control and select the blade that you want to control from the blade selection pull-down box. 1. Power on the system using the power control feature on the remote control, and press F1 when the UEFI splash panel displays, as shown in Figure 8-31 on page 396. 2. Navigate to the Boot Manager menu by selecting it from the System Configuration and Boot Management main menu. The Boot Manager displays, as shown in Figure 8-33.
3. Select Delete Boot Option from the Boot Manager menu. 4. Select all of the items from which you do not want to boot by using the Spacebar to select them. When you have selected all of the items, scroll down to the end of the page using the down arrow key and select Commit Changes, as shown in Figure 8-34 on page 400.
399
Note that, in our example in Figure 8-34, we also have selected Hard Disk 0 for removal. Our system is running Windows 2008 with GUID Partition Table (GPT) disk, and therefore, Windows Boot Manager is being used to boot the operating system. 5. Press Esc to return to the Boot Manager menu. 6. Select Change Boot Order from the Boot Manager menu. 7. Press Enter to make the device list active. 8. Use the up and down arrow keys to navigate to the device for which you want to change the order. After you highlight the device, use the - or Shift and + keys to move the device up or down the list. You can then perform the same actions to move other devices up or down the list. Press Enter when finished. 9. Use the down arrow key to highlight Commit Changes and press Enter to commit the changes that you have made. 10.Press Esc to return to the Boot Manager main menu. 11.Press Esc again to exit to the System Configuration and Boot Management main menu. 12.Press Esc again to exit the UEFI and press the Y key to exit and save any changes that you have made. The HX5 then proceeds to boot normally.
400
The MAX5 Memory Scaling setting provides two options to determine how the system will present the memory capacity in the MAX5 unit to the running operating system: Non-Pooled The default option divides and assigns the memory in the MAX5 between the two installed processors. Pooled This option presents the additional memory in the MAX5 as a pool of memory without being assigned to any particular processor.
401
Use these general settings for each operating system type: Pooled: Use this setting with Linux operating systems. Non-Pooled: VMware and Microsoft operating systems must be configured to use the Non-Pooled setting.
402
3. Ensure that both HX5s are powered off by looking at the Status column in the Unassigned Nodes section. If the systems are not in a Powered Off state, the following warning message (Figure 8-37) displays when you attempt to create a partition.
Figure 8-37 Partition creation failure message due to blades being in a powered on state
4. If the systems are powered on, you can shut them down from this page by clicking the check boxes next to them and by selecting the Power Off Node action from the Available actions pull-down, as shown in Figure 8-38 on page 404. The Status will change to Powered Off momentarily after the Power Off Node action has been applied.
Chapter 8. IBM BladeCenter HX5
403
Figure 8-38 Powering off an HX5 from the scalable complex configuration page
5. Select the two HX5 systems in the Unassigned Nodes area by clicking the check boxes next to them. The Create Partition action in the Available actions pull-down list box is preselected for you, as shown in Figure 8-39.
6. Click Perform action. The complex will form and the task of creating a partition is now complete. Figure 8-40 on page 405 shows the actions that you will be able to perform against the complex.
404
The following options are available: Power Off Partition This option powers off all blades in the partition (in this example, it powers off blade 1 and blade 2). Power On Partition This option powers on all blades in the partition. Remove Partition This option removes the partitioning of the selected partition. The nodes that formed the partition are then removed from the Assigned Nodes section and become available again in the Unassigned Nodes section. Toggle Stand-alone/Partition Mode This mode allows you to toggle between single partition mode and stand-alone mode without having to modify the physical setup of the blade servers, for example: You can toggle the scalable blade complex to stand-alone mode and install a separate operating system on each blade server and run separate applications on each blade server. You can then toggle the blade server complex back to a single partition and run applications that take advantage of the four processors and 32 DIMMs. The operating system that is in use in a single partition is the operating system that is installed on the primary blade server. Later, you can toggle the complex back to stand-alone mode again to gain access to the operating system on the secondary blade server.
Chapter 8. IBM BladeCenter HX5
405
406
407
By clicking the Remote Drive icon (as highlighted in Figure 8-43 on page 409), you can mount any local CD\DVD Rom drive from the management workstation from which you are running the remote control session. You can also mount a supported .ISO or .IMG type file, such as an operating system image file.
408
Remote drive
Use the following steps to mount an image as virtual CD/DVD media at the controlled system: 1. Click Select Image under the Available Resources area of the Remote Disk panel, as shown in Figure 8-43. 1. Click Add Image. 2. Browse to an image that is in a .iso or .img format and click Open when finished. 3. Click Mount All when finished to mount the image to the blade.
409
For Linux, look on the vendor websites. They contain information about installation with a USB flash drive. For example, the following websites provide details for using a USB key as an installation medium: Installing Red Hat Linux from a USB flash drive: http://ibm.com/support/techdocs/atsmastr.nsf/WebIndex/WP101131 How to create a bootable USB drive to install SLES: http://www.novell.com/support/php/search.do?cmd=displayKC&docType=kc&externalId =3499891 You can also use the IBM ServerGuide Scripting Toolkit to create a bootable USB flash drive, as explained in the next section.
Windows installation
This section describes the process to install the ServerGuide Scripting Toolkit, to create a deployment image for Windows 2008 R2 Enterprise Edition, and to copy this image to a USB key for deployment. To configure a USB key for deployment, you need the following devices: A system running Windows Vista, Windows Server 2008, Windows 7, Windows Server 2008 R2, Windows 2.1 PE, or a Windows 3.0 PE session A USB key with a storage capacity at least 64 MB larger than your Windows PE image, but not less than 4 GB Use this procedure: 1. Install the ServerGuide Scripting Toolkit You must install the English language version of the Windows Automated Installation Kit (AIK) for the Windows 7 family, Windows Server 2008 family, and Windows Server 2008 R2 family, which is available at the following website: http://www.microsoft.com/downloads/en/details.aspx?familyid=696DD665-9F76-4177A811-39C26D3B3B34&displaylang=en Follow these steps to install the ServerGuide Scripting Toolkit, Windows Edition: a. Download the latest version from the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-TOOLKIT b. Create a directory, for example, C:\sgshare. c. Decompress the ibm_utl_sgtkwin_X.XX_windows_32-64.zip file to the directory that you have created, for example, C:\sgshare\sgdeploy. 2. Create a deployment image Follow these steps to create a Windows installation image: a. Start the Toolkit Configuration Utility in the C:\sgshare\sgdeploy directory. 410
IBM eX5 Implementation Guide
c. Choose the operating system type that you want and click Next, as shown in Figure 8-45.
d. Insert the correct OS installation media or select the folder that contains the installation files for the source, as shown in Figure 8-46 on page 412. If necessary, modify the target and click Next.
411
e. When the file copy process completes, as shown in Figure 8-47, click Finish.
f. Open a command prompt and change to the C:\sgshare\sgdeploy\SGTKWinPE directory. Use the following command to create the Windows installation image: SGTKWinPE.cmd ScenarioINIs\Local\Win2008_R2_x64_EE.ini g. When the process is finished, as shown in Figure 8-48 on page 413, your media creation software is started to create a bootable media from the image. Cancel this task.
412
18:26:21 - Creating the WinPE x64 ISO... 18:27:07 - The WinPE x64 ISO was created successfully. *** WinPE x64 ISO: 4_EE\WinPE_x64.iso c:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x6
18:27:07 - Launching the registered software associated with ISO files... *** Using ISO File: 64_EE\WinPE_x64.iso c:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x
18:27:08 - The WinPE x64 build process finished successfully. SGTKWinPE complete. Figure 8-48 Build process is finished
3. Prepare the USB key Follow these steps to create a bootable USB key with the Windows installation image that was created in step 2 on page 410: a. Insert your USB key. b. Use diskpart to format the USB key using FAT32. All files on the USB key will be deleted. At the command prompt, type the commands that are listed in Figure 8-49 on page 414.
413
C:\>diskpart Microsoft DiskPart version 6.1.7600 Copyright (C) 1999-2008 Microsoft Corporation. On computer: DISKPART> list disk Disk ### -------Disk 0 Disk 1 Disk 2 Status ------------Online Online Online Size Free Dyn ------- ------- --271 GB 0 B 135 GB 0 B 7839 MB 0 B Gpt --*
DISKPART> select disk 2 Disk 2 is now the selected disk. DISKPART> clean DiskPart succeeded in cleaning the disk. DISKPART> create partition primary DiskPart succeeded in creating the specified partition. DISKPART> select partition 1 Partition 1 is now the selected partition. DISKPART> active DiskPart marked the current partition as active. DISKPART> format fs=fat32 100 percent completed DiskPart successfully formatted the volume. DISKPART> assign DiskPart successfully assigned the drive letter or mount point. DISKPART> exit Figure 8-49 Using diskpart to format the USB memory key
c. Copy the contents from C:\sgshare\sgdeploy\WinPE_ScenarioOutput\Local_Win2008_R2_x64_EE\ISO to the USB key. The USB key includes the folders and files that are shown in Figure 8-50.
d. Boot the target system from the USB key. The deployment executes automatically.
414
RAID controller: If the target system contains a RAID controller, RAID is configured as part of the installation.
415
d. Select Change Boot Order. Change the boot order so that Legacy Only appears at the top of the list and the Embedded Hypervisor option appears beneath it, as shown in Figure 8-52.
e. Select Commit Changes and press Enter to save the changes. 3. Configure UEFI for embedded ESXi 4.1 if the MAX5 is attached Systems running VMware ESXi Server must use Non-Pooled mode in the MAX5 Memory Scaling option with the UEFI. See 8.5.5, HX5 with MAX5 attached on page 401 for instructions to configure the MAX5 Memory Scaling option. 4. Boot a new embedded ESXi 4.1 with MAX5 attached Use the following steps to successfully boot a new ESXi 4.1 embedded hypervisor on the HX5 with MAX5 attached: a. Ensure that you have the latest FPGA code installed on the HX5 by updating the FPGA code if required.
416
b. Physically attach the MAX5 by using the instructions that are provided in the IBM BladeCenter HX5 Installation and Users Guide, which is available at this website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5084612 c. Reflash the FPGA code to ensure that both the HX5 FPGA firmware and the MAX5 FPGA firmware are at the same level. d. When the FPGA code has been flashed, enter UEFI by pressing F1 at the UEFI splash panel. e. Boot the host. f. In the Loading VMware Hypervisor panel, press Shift+O when the gray progress bar is displayed. g. Enter the following command at the prompt: esxcfg-advcfg -k TRUE allowInterleavedNUMAnodes h. After the system boots, connect to the system using the vSphere Client. i. Select the Configuration tab of the host and click Advanced Settings under the Software panel, as shown in Figure 8-53.
j. Click VMkernel in the left pane and select the check box next to VMkernel.Boot.allowInterleavedNUMAnodes, as shown in Figure 8-54 on page 418.
417
k. Click OK. At this point, you have completed the process of configuring the embedded ESXi 4.1 on the HX5 with MAX5.
418
3. Click VMkernel in the left pane and select the check box next to VMkernel.Boot.allowInterleavedNUMAnodes, as shown in Figure 8-56 on page 420.
419
4. Click OK. 5. Shut down the HX5. 6. Ensure that you have the latest FPGA code installed on the HX5 by updating it, if required. 7. Physically attach the MAX5 by using the instructions that are provided in the Installation Guide. 8. Reflash the FPGA code to ensure that both the HX5 FPGA code and MAX5 FPGA code are at the same level. 9. After the FPGA code has been flashed, log in to the UEFI by pressing F1 at the UEFI splash panel. 10.Select System Settings Memory from the System Configuration and Boot Management main menu. 11.Ensure that the MAX5 Memory Scaling option is set to Non-Pooled. 12.Continue to boot the server normally. Log in to the vSphere Client and check that the additional memory shows on the Systems Summary tab.
420
Description VMware ESX Server 3i v 3.5 Update 3 VMware ESX Server 3i v 3.5 Update 2 VMware ESX Server 3i version 3.5
To order a recovery CD, contact your local support center at the following website: http://www.ibm.com/planetwide/region.html
Updating ESXi
You can install the latest version of ESXi 4 on IBM Hypervisor keys, and it is supported by IBM. Use the following VMware upgrade mechanisms for the update: VMware Upgrade Manager Host Update Utility For more information, see the VMware Documentation website: http://www.vmware.com/support/pubs
When installing any supported version of the ESX server operating system onto the HX5, the memory must be balanced across all processors in the system. This rule applies to 2-socket and 4-socket HX5 configurations, as well as the HX5 with MAX5 attached. Failure to follow this rule prevents the operating system from installing correctly.
421
Common steps for both ESX 4.1 and ESXi 4.1 Installable editions
Perform the following steps for both operating system types: 1. Ensure that you have the latest FPGA code installed on the HX5 by updating it, if required. 2. Physically attach the MAX5 using the instructions that are provided in the IBM BladeCenter HX5 Installation and Users Guide, which is available at this website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5084612 3. Reflash the FPGA code to ensure that both the HX5 FPGA code and MAX5 FPGA code are at the same level. 4. When the FPGA code has been flashed, log in to UEFI by pressing F1 at the UEFI splash panel. 5. Select System Settings Memory from the System Configuration and Boot Management main menu. 6. Ensure that the MAX5 Memory Scaling option is set to Non-Pooled. Exit the UEFI when finished and proceed to the installation of the respective version of ESX server.
422
f. Select Commit Changes and press Enter to save the changes. 2. Boot the host from the ESX installation media. 3. Press F2 when you receive the installation options panel, as shown in Figure 8-58.
4. The Boot Options line appears on the panel. Type the following parameter at the end of the Boot Options line: allowInterleavedNUMAnodes=TRUE The edited result looks like Figure 8-59 on page 424. Press Enter to proceed.
423
5. Proceed through the installer until you reach the Setup Type page. Click Advanced setup and clear Configure boot loader automatically (leave checked if unsure), as shown in Figure 8-60 on page 425.
424
6. Proceed through the installer to the Set Boot loader Options page and type the following parameter in the Kernel Arguments text box, as shown in Figure 8-61 on page 426: allowInterleavedNUMAnodes=TRUE
425
7. Complete the remainder of the ESX installation and reboot the host.
426
c. Select CD/DVD Rom, Legacy Only, and Hard Disk 0, as shown in Figure 8-62, for example, if you are using internal drives that have been configured as an integrated mirror with the onboard RAID controller. If you cannot find these options, the options are already in the boot list. d. Press Esc when finished to go back one panel. e. Select Change Boot Order. Change the boot order to look like the boot order that is shown in Figure 8-62.
f. Select Commit Changes and press Enter to save the changes. 2. Boot from the ESXi Installable installation media. 3. Press the Tab key when the blue boot panel appears, as shown in Figure 8-63.
4. Add the following line after vmkboot.gz: allowInterleavedNUMAnodes=TRUE Ensure that you leave a space at the beginning and the end of the text that you enter, as shown in Figure 8-64 on page 428. Otherwise, the command will fail to execute at a later stage during the installation. Press Enter to proceed.
Chapter 8. IBM BladeCenter HX5
427
5. Complete the ESXi installation and reboot when prompted. Ensure that you remove the media or unmount the installation image before the system restarts. 6. In the Loading VMware Hypervisor panel, when the gray progress bar is displayed, press Shift+O. Note: If you do not press Shift+O during the Loading VMware Hypervisor panel, you receive this error: The system has found a problem on your machine and cannot continue. Interleave NUMA nodes are not supported. 7. Enter the following command at the prompt after you have pressed Shift+O: esxcfg-advcfg -k TRUE allowInterleavedNUMAnodes Your output looks like Figure 8-65 on page 429.
428
8. Press Enter after the command has been entered. Press Enter again to continue to boot. 9. After the system boots, connect to it using the vSphere Client. 10.Select the Configuration tab of the host and click Advanced Settings under Software, as shown in Figure 8-66 on page 430.
429
11.Click VMkernel in the left pane and select the check box next to VMkernel.Boot.allowInterleavedNUMAnodes, as shown in Figure 8-67 on page 431.
430
12.Click OK. At this point, you have completed the process of installing ESXi 4.1 Installable on the HX5 with MAX5.
Attaching MAX5 to HX5 with an existing installation of ESX 4.1 or ESXi 4.1 Installable edition
The easiest way to add the MAX5 to an existing installation of the ESX 4.1 or ESXi 4.1 Installable running on an HX5 is by using the following steps: 1. Connect to the relevant ESX 4.1 or ESXi 4.1 Installable server by logging in to it using the VMware vSphere Client. 2. Select the Configuration tab of the host and click Advanced Settings under Software. 3. Click VMkernel in the left pane and select the check box next to VMkernel.Boot.allowInterleavedNUMAnodes, as shown in Figure 8-68 on page 432. Click OK when finished.
431
4. Shut down the HX5. 5. Ensure that you have the latest FPGA code installed on the HX5 by updating it, if required. 6. Physically attach the MAX5 using the instructions that are provided in the Installation Guide. 7. Reflash the FPGA code to ensure that both the HX5FPGA code and the MAX5 FPGA code are at the same level. 8. After the FPGA code has been flashed, log in to UEFI by pressing F1 at the UEFI splash panel. 9. Select System Settings Memory from the System Configuration and Boot Management main menu. 10.Ensure that the MAX5 Memory Scaling option is set to Non-Pooled. 11.Continue to boot the server normally. Log in to the vSphere Client and check that the additional memory shows on the Systems Summary tab.
Adding MAX5 to HX5 with an existing installation of ESX 4.1 without first editing the VMkernel before attaching
Use the following steps if you have already shut down the existing ESX 4.1 server without having made the changes that are detailed in Attaching MAX5 to HX5 with an existing installation of ESX 4.1 or ESXi 4.1 Installable edition on page 431: 1. Power on the HX5. 2. At the VMware boot loader panel, as shown in Figure 8-69 on page 433, ensure that VMware ESX 4.1 is highlighted and press the a key to modify the kernel arguments.
432
3. Add the following line at the beginning of the boot load command: allowInterleavedNUMAnodes=TRUE Ensure that you leave a space at the beginning and at the end of the text that you enter, as shown in Figure 8-70. Otherwise, the command will fail to execute at a later stage during the boot process. Press Enter to proceed after you have correctly edited the line to continue the boot process.
4. When the ESX 4.1 operating system has completely loaded, proceed with step 1 on page 431 through step 3 on page 431 to complete the process.
Adding MAX5 to HX5 with an existing installation of ESXi 4.1 Installable without first editing the VMkernel before attaching
If you have already shut down the existing ESXi 4.1 server without having made the changes that are detailed in Attaching MAX5 to HX5 with an existing installation of ESX 4.1 or ESXi 4.1 Installable edition on page 431, complete step 6 on page 428 through step 11 on page 430.
433
We recommend that you review all of the RETAIN tips, which relate to this system and the operating system that you are installing, which are available at the following website: http://ibm.com/support/entry/portal/Troubleshooting/Hardware/Systems/BladeCente r/BladeCenter_HX5/7872 For further general VMware tuning information, see the paper VMware vCenter Server Performance and Best Practices for vSphere 4.1, which is available at this website: http://www.vmware.com/resources/techresources/10145
434
the HX5 has not been zoned to external SAN storage at the time of installation unless you are performing boot from SAN. Failure to follow this rule potentially will result in
Installing Microsoft Windows Server 2008 R2 on HX5 without the use of ServerGuide
The following website provides specific information about installing Microsoft Windows Server 2008 R2 without using the ServerGuide CD: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5085004 You can access more information about the Windows Server 2008 R2 installation and configuration at the following website: http://www.microsoft.com/windowsserver2008/en/us/product-documentation.aspx
435
436
Red Hat Linux 5 software, firmware, drivers, and fixes for HX5
See the following website for all the latest information relating to software, firmware, drivers, and fixes for HX5 when used with Red Hat Linux 5: http://www-933.ibm.com/support/fixcentral/systemx/quickorder?parent=ibm~BladeCente rHX5&product=ibm/systemx/7872&&platform=RHEL+5+x64&function=all&source=fc
437
SUSE Linux Enterprise Server software, firmware, drivers, and fixes for HX5
Go to the following websites for all the latest information relating to software, firmware, drivers, and fixes for HX5 when used with SUSE Linux Enterprise Server: For SUSE Linux Enterprise Server 11: http://www-933.ibm.com/support/fixcentral/systemx/quickorder?parent=ibm~BladeCe nterHX5&product=ibm/systemx/7872&&platform=SLES+11+x64&function=all&source=fc SUSE Linux Enterprise Server 10: http://www-933.ibm.com/support/fixcentral/systemx/quickorder?parent=ibm~BladeCe nterHX5&product=ibm/systemx/7872&&platform=SLES+10+x64&function=all&source=fc
Server firmware
Software that resides on flash memory and controls the lower-level function of server hardware is called the server firmware. An IBM System, such as the HX5, runs a number of firmware images that control various components of the blade. The following list shows the primary firmware for the HX5:
Unified Extensible Firmware Interface (UEFI) Integrated Management Module (IMM) Field-Programmable Gate Array (FPGA) Preboot Dynamic System Analysis (DSA)
Additional devices, such as network cards and RAID controllers, also contain their own firmware revisions. IBM provides firmware updates, including proven firmware from other manufacturers to be applied on IBM systems, that can be downloaded from the IBM website. We describe several methods of performing firmware updates on IBM eX5 servers in Chapter 9, Management on page 447. Important tip: Update all System x firmware to the latest levels prior to installing the operating system or application installation.
438
Tip: IBM Bootable Media Creator (BoMC) is a tool that simplifies the IBM System x firmware update process without an operating system running on the system. More information about this tool is available at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=TOOL-BOMC
Device drivers
Device drivers are software that controls the hardware server components on the operating system level. They are specific to the operating system version, and therefore, critical device drivers are included with the installation media.
IBM, operating system vendors, and component device vendors provide device driver updates. Often, you can download them from each companys support website. Whenever possible, we recommend acquiring tested and approved driver updates from IBM. Tip: The Windows Server installation process using IBM ServerGuide will perform IBM driver updates after the operating system installation has completed. You can obtain this tool from the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-GUIDE
Tip: IBM UpdateXpress is a tool that allows the IBM System x firmware and drivers to be updated through the operating system. More information about this tool is available at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-XPRESS
439
Available support Documentation, driver, and OS update Documentation, driver, and OS update Documentation, driver, and OS update
If you do not have internal drives installed, disable the onboard SAS RAID Controller in UEFI by navigating to System Settings Devices and IO ports Enable/Disable Onboard Devices and disabling the SAS Controller or Planar SAS. Set the HBA as the first device in the Option ROM Execution Order by selecting System Settings Devices and IO Ports Set Option ROM Execution Order. All operating systems, except Windows 2008 and SLES 11 SP1, now have Legacy Only set as their first boot device. Remove all devices that might not host an operating system from the boot order. The optimal minimum configuration is CD/DVD and Hard Disk 0. For existing operating systems only, set Legacy Only as the first boot device. You must set the BIOS on the HBA to Enabled. Make sure that the logical unit number (LUN) that will host the operating system installation is accessible through only one path on the SAN at the time of the installation. Verify that your HBA can see a LUN from your storage. After installation, do not forget to install the multipath driver before you set more than one path if you have more than one path to the LUN.
441
442
For the configuration of SNMP traps and email alerts, see the IBM BladeCenter Advanced Management Module Installation Guide, which is available at the following website: http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073392 For the configuration of Service Advisor, see to 9.3.2, Service Advisor on page 458. For information about managing and centralizing alerts for managed systems using the IBM Systems Director, see Implementing IBM Systems Director 6.1, SG24-7694.
443
444
You might see the following effects of split power domains for HX5 with MAX5 configurations: A loss of power to either power domain results in complete system shutdown. The HX5 will not be able to power on if the attached MAX5 in slot 8 has no power, because the HX5 treats the entire system as a single unit. You must remove the IBM HX5 MAX5 1-Node Scalability kit to allow the HX5 to resume normal operation if power cannot be restored to the second power domain that contains the MAX5. Where possible, avoid placing HX5 4-socket or HX5 and MAX5 configurations across blade bays 7 and 8. BladeCenter design: The loss of power to an entire power domain is extremely unlikely due to the redundant design of the BladeCenter H chassis. We provide this information merely for guidance in the case of this unlikely event.
445
446
Chapter 9.
Management
In the information technology sector, server systems management has received greater focus over recent years. The ability to maintain and manage systems efficiently is essential to business IT operations. We briefly describe the embedded hardware and external applications that are available to manage and maintain the eX5 range of systems. And, we demonstrate several of them. This chapter contains the following topics: 9.1, Introduction on page 448 9.2, Integrated Management Module (IMM) on page 449 9.3, Advanced Management Module (AMM) on page 454 9.4, Remote control on page 462 9.5, IBM Systems Director 6.2 on page 467 9.6, IBM Electronic Services on page 493 9.7, Advanced Settings Utility (ASU) on page 495 9.8, IBM ServerGuide on page 501 9.9, IBM ServerGuide Scripting Toolkit on page 507 9.10, Firmware update tools and methods on page 509 9.11, UpdateXpress System Pack Installer on page 511 9.12, Bootable Media Creator on page 514 9.13, MegaRAID Storage Manager on page 521 9.14, Serial over LAN on page 525
447
9.1 Introduction
IBM provides a number of tools to successfully deploy, manage, and maintain the eX5 range of systems. The collective name for these tools is IBM ToolsCenter. You can access these tools at the following link: http://ibm.com/support/entry/portal/docdisplay?lndocid=TOOL-CENTER We group these tools in the following high-level categories: System deployment System configuration System updates System diagnosis System management We provide a summary of these tools in Table 9-1. In this chapter, we walk you through the process of using several of these tools.
Table 9-1 High-level overview of the available tools to manage the eX5 range of systems IBM tools Integrated Management Module Advanced Management Module Dynamic System Analysis Bootable Media Creator Scripting Toolkit IBM ServerGuide Start Now Advisor Update Express System Pack Installer IBM Systems Director IBM Electronic Service Agent Storage Configuration Manager Remote Control Advanced Settings Utility MegaRAID Storage Manager Serial over LAN Covered here p. 449 p. 454 No p. 514 p. 507 p. 501 No p. 511 p. 467 p. 493 No p. 462 p. 495 p. 521 p. 525 Firmware deployment Yesa Yes No Yes Yes No Yes Yes Yes No Yes Noc Noc No No OS installation assistance No No No No Yes Yes No No Yesb No No Noc Noc No No System management Yes Yes No No No No No No Yes No Yes Yes Yes Yesd Yes Problem diagnosis Yes Yes Yes No No No No No Yes Yes Yes No No Yes No
a. You can only update Integrated Management Module (IMM) firmware using the IMM web browser interface. b. Only when the Tivoli Provision Manager for OS Deployment IBM Systems Director edition is installed. c. This tool provides the ability to mount the media containing the firmware or operating system. d. This tool provides the management of RAID controllers and hard disks only.
448
Chapter 9. Management
449
Enabling Ethernet port 1: The onboard Ethernet port 1 is not shared for use with the IMM, by default. You must enable this feature in the Unified Extensible Firmware Interface (UEFI) of the server in both the x3690 X5 and the x3850 X5. The following sections contain the configuration instructions. Figure 9-1 shows the location of the available ports to manage the IMM on the x3690 X5.
Ethernet port 1
Figure 9-1 Ports over which the IMM can be managed out-of-band on an x3690 X5
Figure 9-2 shows the location of the ports available to manage the IMM on the x3850 X5.
Ethernet port 1
Figure 9-2 Ports over which the IMM can be managed out-of-band on an x3850 x5
To enable the IMM to use Ethernet port 1, you must complete the following instructions: 1. Boot up the server and press F1 when prompted. 2. Select System Settings from the System Configuration and Boot Management menu. 3. Select Integrated Management Module from the System Settings menu. 4. Select Network Configuration from the Integrated Management Module. This menu allows you to configure the network settings for the IMM. Also, you can configure the IMM to share the use of Ethernet port 1 for out-of-band management from this menu, as shown in Figure 9-3 on page 451. 450
Figure 9-3 Network Configuration menu showing the Network Interface Port settings
5. Set the Network Interface Port setting to Shared to allow the IMM to use Ethernet port 1. 6. For DHCP Control, choose the Static IP option. 7. For IP Address, enter the relevant IP address. 8. For Subnet Mask, enter the required subnet mask. 9. For Default Gateway, enter the required default gateway address. 10.When you have completed the IP address configuration, press Esc three times to return to the System Configuration and Boot Management menu. 11.For Exit Setup, press the Y key when prompted to save and exit the Setup utility. At this point, the server reboots with the new settings. 12.Plug a network cable into either the dedicated system management port or Ethernet port 1, if you set the IMM to share its use as per instructions. Make sure that you can ping the IP address of the IMM on the connected network port. After the IMM is available on the network, you can log in into the IMM web interface by typing its IP address in a supported web browser, as shown in Figure 9-4 on page 452.
Chapter 9. Management
451
Enter the default user name, which is USERID. This user name is case-sensitive. Enter the default password, which is PASSW0RD. You must use a zero in place of the letter o in the password. See the Users Guide for Integrated Management Module - IBM BladeCenter and System x at the following website for the additional configuration settings of the IMM: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5079770
452
After you log in to the AMM, you can manage and retrieve information from the HX5, such as hardware vital product data (VPD), power usage, and so on. We describe more of this information in 9.3, Advanced Management Module (AMM) on page 454.
Chapter 9. Management
453
IMM firmware update: After the IMM firmware update, the IMM will be reset. The restart of the IMM after the update can take up to 9 minutes (13 minutes on a scalable complex system). After this restart, the IMM is ready for any further firmware updates. Use the IMM to update firmware: 1. Log in into the IMM web interface. 2. Select Tasks Firmware Update, as shown in Figure 9-6.
Tasks Power/Restart Remote Control PXE Network Boot Firmware Update
3. Click Browse and select the update file. 4. Click Update to start the update process. A progress indicator opens as the file is transferred to the temporary storage of the IMM. When the transfer is completed, click Continue to complete the update process. A progress indicator opens as the firmware is flashed. A confirmation page opens to verify that the update was successful.
454
OK
Serial connector
Video
LINK
TX/RX
All BladeCenter chassis come standard with at least one AMM. Each chassis also supports a second management module for redundancy. One of the management modules is active, and the second management module, if installed, remains on standby until the management functions are manually switched over to it or if the primary management module fails. The service processor in the management module communicates with the service processor in each blade server. The IMM is the service processor on the HX5. The service processor allows support features, such as blade server power-on requests, error and event reporting, KVM requests, and requests to use the BladeCenter shared media tray (removable-media drives and USB connector). You configure BladeCenter components by using the management module to set information, such as IP addresses. The management module communicates with all components in the BladeCenter unit, detecting their presence or absence, reporting their status, and sending alerts for error conditions when required. With the AMM, you can perform the following tasks: Define the login IDs and passwords Configure the security settings, such as data encryption and user account security Select recipients for the alert notification of specific events Monitor the status of the BladeCenter unit, blade servers, and other BladeCenter components: Event log LEDs Hardware and firmware VPD Fan speeds Temperatures Power usage
Chapter 9. Management
455
Discover other BladeCenter units on the network and enable access to them through their management-module Web interfaces Control the BladeCenter unit, blade servers, and other BladeCenter components: Power on/off Firmware update Configuration settings Serial over LAN
Configure power management for BladeCenter unit Access the I/O modules to configure them Change the start-up sequence in a blade server Set the date and time Use a remote console for the blade servers Mount remote virtual media for the blade servers Change ownership of the keyboard, video, and mouse Change ownership of the removable-media drives and USB ports (the removable-media drives in the BladeCenter unit are viewed as USB devices by the blade server operating system) Set the active color of the critical (CRT) and major (MJR) alarm LEDs (for BladeCenter T units only) Use BladeCenter Open Fabric Manager functions Scale the HX5 systems Use Service Advisor functions to autonomously inform IBM support about any critical events happening The AMM supports the following management methods: Web-based interface with Secure Sockets Layer (SSL) support Command-line interface (CLI) through Telnet/Secure Shell (SSH) Systems Management Architecture for Server Hardware (SMASH) Command-Line Protocol Simple Network Management Protocol (SNMP)
456
The System Status summary page displays. This page provides you with information about the overall health of the chassis and its installed components, as shown in Figure 9-9 on page 458.
Chapter 9. Management
457
All high-level tasks that can be performed against the various components within the BladeCenter chassis are shown in the left pane, as highlighted in Figure 9-9. Of particular importance to this IBM Redbooks publication is the Scalable Complex section. In the Scalable Complex section, you can configure the HX5 blades to form 2-node scalable systems. We describe the scaling configuration of the HX5s in 8.6, Creating an HX5 scalable complex on page 402. For more information about the AMM and its configuration, see the IBM BladeCenter Advanced Management Module Installation Guide at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5073392
event log and any configured alerts are sent. The information that is gathered by Service Advisor is the same information that is available if you save service data from the AMM web interface. After gathering the information, Service Advisor automatically initiates a call to IBM. Upon receipt of the information, IBM returns a service request ID, which is placed in the call home activity log. Tip: Although Service Advisor can send alerts 24 hours a day, 7 days a week, your service provider responds according to the arrangement that you have in place with the service provider. On the Event Log page of the AMM web interface, you can choose to select the Display Call Home Flag check box. If you select the check box, events are marked with a C for call home events and an N for events that are not called home. In addition, you can filter the event log based on this setting. Tip: None of the information in the Call Home report contains client data from the servers or the I/O modules. Before you configure the BladeCenter Service Advisor, you must ensure that ports 80 and 443 are open for the AMM to call home. Complete the steps that follow to successfully configure the Service Advisor: 1. Log in to the AMM on which you want to activate the Service Advisor. 2. In the left navigation pane, click Service Tools Service Advisor. If this is the first time that you select this option, or if the AMM firmware was reset to the default values, you need to view and accept the license agreement. 3. Click View terms and conditions to view the Service Advisor agreement. Click I accept the agreement on the terms and conditions page to close the page. 4. Click the Service Advisor Settings tab, as shown in Figure 9-10 on page 460, and complete all relevant details. Ensure that you select the correct IBM support center for your geographical location. Also, note that the FTP/TFTP Server of Service Data must only be configured if an approved service provider other than IBM provides your hardware warranty. 5. After all data has been entered, click Save IBM Support. At this point, the Service Advisor is not enabled. 6. To enable Service Advisor, click the Service Advisor Settings tab again and click Enable IBM Support. The Service Advisor Disabled status, as shown in Figure 9-10 on page 460, now shows Enabled.
Chapter 9. Management
459
7. It is advisable to generate a test call to IBM to ensure that the BladeCenter chassis can call home correctly. Select the Test Call Home tab and click Test Call Home. The Test Call Home tab only appears if the Service Advisor is set to Enabled. 8. You will be returned to the Service Advisor Activity Log tab after you click Test Call Home. 9. Click Refresh on the activity log until a success or failure is registered in the Send column of the activity log. If the call was successful, a ticket number appears in the Assign Num column. The ticket that is opened at IBM is identified as a test ticket. No action is required from IBM support for a test ticket, and the call will be closed. See the Connectivity security for Service Advisor section of the IBM BladeCenter Advanced Management Module
460
Installation Guide if you experience difficulties configuring the Service Advisor. You can obtain this guide at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5073392
Use these recommendations for updating through the AMM: After the IMM update is complete, wait at least 15 minutes before you initiate any planned UEFI or DSA Preboot firmware updates. Ensure that TFTP is enabled. Select MM Control Network Protocols Trivial File Transfer Protocol (TFTP), as shown in Figure 9-11.
If you have an AMM level 50G or earlier and you will update multiple blades in the chassis, restart the AMM once before beginning multiple updates. For the UEFI update, the server must be powered up and booted into the operating system or in the UEFI menu. Use the AMM to update firmware: 1. Log in to the AMM web interface. 2. Select Blade Tasks Firmware update. 3. Select the target blade, as shown in Figure 9-12 on page 462.
Chapter 9. Management
461
Figure 9-12 Overview of the blades that you select for update
4. Use Browse to locate the update file. 5. Click Update to start the update.
9.4.1 Accessing the Remote Control feature on the x3690 X5 and the x3850 X5
Follow these steps to use the Remote Control feature on the IMM for the x3690 X5 and the x3850 X5: 1. Log in to the IMM of the specific system that you want to control. 2. Select Tasks Remote Control, as shown in Figure 9-13 on page 463. 3. To protect sensitive disk and KVM data during your session, click the Encrypt disk and KVM data during transmission check box before starting Remote Control. For complete security, use Remote Control in conjunction with SSL. You can configure SSL at IMM Control Security. 462
IBM eX5 Implementation Guide
4. If you want exclusive remote access during your session, click Start Remote Control in Single User Mode. If you want to allow other users remote console (KVM) access during your session, click Start Remote Control in Multi-user Mode.
5. Two separate windows open: one window for the Video Viewer (as shown in Figure 9-14 on page 464) and the other window for the Virtual Media Session (as shown in Figure 9-15 on page 465). The various controls that you have to control the server are all contained in the Users Guide for Integrated Management Module - IBM BladeCenter and System x, which you can access at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5079770
Chapter 9. Management
463
Figure 9-14 IMM Remote Control Video Viewer showing power control options
464
Chapter 9. Management
465
3. To select a blade that you want to control, click Start Remote Control. 4. To view the video of a blade, select the blade from the pull-down list box, as shown in Figure 9-17 on page 467. You can obtain additional instructions for power control, mounting remote media, and soft key programming in the IBM BladeCenter Advanced Management Module Installation Guide at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=MIGR-5073392
466
Chapter 9. Management
467
IBM Systems Director provides the following functionality for eX5 systems, as well as many other system types: Discovery and inventory Monitoring and reporting Software updates Configuration management Virtual resource management Remote control, which includes power control Automation Tip: Your IBM Systems Director server must be at Version 6.2.1 to support all the eX5 systems, including their ability to scale into a complex. This version is also required to support the MAX5. In 9.2, Integrated Management Module (IMM) on page 449, we described the concepts of in-band and out-of-band management of the IMM. In this section, we demonstrate how to discover the eX5 systems via IBM Systems Director.
9.5.1 Discovering the IMM of a single-node x3690 X5 or x3850 X5 out-of-band via IBM Systems Director
After your IMM has been configured for out-of-band management, as described in Configuring an x3850 X5 or x3690 X5 for out-of-band management on page 449, you can discover the system from within IBM Systems Director. After the IMM has been added to the IBM Systems Director console, you can perform management tasks against it. Perform these steps to add the IMM to the IBM Systems Director console: 1. Log in to the IBM Systems Director web interface by navigating to the following website, where servername is the Domain Name System (DNS) registered name of your IBM Systems Director: http://servername:8421/ibm/console For example: http://director6.ibm.com:8421/ibm/console You can also connect to the IBM Systems Director web interface using its IP address: http://ipaddress:8421/ibm/console For example: http://182.168.1.10:8421/ibm/console Tip: We advise that you configure your IBM Systems Director server correctly to use DNS for name resolution. We also recommend that you register the IMM in DNS to simplify the management of the IMM out-of-band.
468
2. After you log in to the console, navigate to Inventory System Discovery in the left navigation pane. In the right pane, select a discovery option in the Select a discovery option pull-down list. IBM Systems Director defaults to discovering a single system using IPv4 address resolution. 3. Enter the IP address of the IMM in the space that is provided under IP address. 4. An IMM is considered a server object in IBM Systems Director. To specify the IMM as a Server object, click the Select the resource type to discover list box and select Server as the resource type, as shown in Figure 9-19 on page 470. 5. Click Discover Now when you are ready to discover the IMM. 6. An informational message indicating that the discovery job has commenced appears at the top of the right pane. Click Close Message to acknowledge the message. 7. The discovered IMM appears at the bottom of the System Discovery pane under Discovered Manageable Systems. At this point, the IMM has been discovered but has not been authenticated to. Notice the No Access status under the Access column. Before authenticating to the IMM, we advise that you rename it first. The renaming process can be performed at a later stage as well, but we advise that you perform this process at discovery time, because it is easier to identify individually discovered IMMs for renaming than when they are in a group.
Chapter 9. Management
469
8. To rename the IMM, right-click the IMM server object in the Discovered Manageable Systems area and click Rename, as shown in Figure 9-20.
9. A Rename display box opens, as shown in Figure 9-21. Provide a meaningful name for the IMM in the text box provided and click OK when finished.
470
10.To authenticate to the IMM, right-click No access under the Access column and click Request Access, as shown in Figure 9-22. Tip: If the access status appears as unknown, right-click the Unknown icon and select Verify Connection. The status changes to No access if IBM Systems Director can communicate correctly with the IMM.
Tip: You can also right-click the IMM and select Security Request Access as an alternate method of authenticating to the IMM or any other supported object. 11.The Request Access pane opens, as shown in Figure 9-23. Enter the user name and password of an account that has supervisor access on the IMM in the text boxes provided and click Request Access.
Chapter 9. Management
471
12.A successful authentication to an IMM yields an OK access status, as shown in Figure 9-24. Click Close when finished.
13.Click Navigate Resources in the left pane of IBM Systems Director. In the right pane, under Groups (View Members), click the All Systems group. Right-click the IMM that you have discovered and select Inventory View and Collect Inventory. 14.Your IMM will be displayed in the Target systems list box, as shown in Figure 9-25.
15.Click Collect Inventory. Click OK to run the task immediately. Tip: Always collect the inventory of a system as soon as it is added to IBM Systems Director to ensure that all system data is captured immediately. 16.Click Display Properties to view the status of the running inventory collection task. After the task is 100% complete, you will then be able to perform tasks against the IMM.
472
Discovery tip: To correctly discover and manage a complex via IBM Systems Director, you must discover the system both out-of-band via the IMM and in-band via the operating system. A V6.2.1 or greater common or platform agent must also be installed. Complete these steps to discover and manage an x3850 X5 complex running a supported operating system via IBM Systems Director: 1. Verify that your systems are configured correctly as a complex by logging in to the IMM web console and navigating to Scalable Partitioning Manage Partitions. Verify that the values that are shown under System Partition Mode match the values that are shown in Table 9-2.
Table 9-2 System Partition Mode status panel System Started Started Partition Valid Valid Mode Multinode Multinode
See 6.5, Forming a 2-node x3850 X5 complex on page 235 for instructions to configure a complex if you have not configured a complex already. 2. Ensure that you have installed the x3850 X5 complex with a supported operating system. 3. Discover and request access to the IMM to the primary node, as described in Discovering the IMM of a single-node x3690 X5 or x3850 X5 out-of-band via IBM Systems Director on page 468. 4. Notice that a Scalable Partition object and a Scalable System object appear in the Navigate Resources All Systems group view. Also, you can see these objects by navigating to Navigate Resources Groups by System Type System x Scalable Systems Scalable Systems and Members. We explain the Scalable System and the Scalable Partition objects: Scalable System Scalable Partition Refers to the system containing all the physical nodes that have been cabled together. Refers to a logical partition running on the scalable system. Older generation multinode IBM systems scaled together can contain multiple partitions. Each partition can then run its own operating system and function independently, even though it is physically cabled to other nodes.
5. The access statuses for these objects appear as Partial access, because you have not yet authenticated to the second IMM in the complex. 6. Discover and request access to the IMM of the secondary node, as described in Discovering the IMM of a single-node x3690 X5 or x3850 X5 out-of-band via IBM Systems Director on page 468. 7. Wait one or two minutes before checking the access statuses of the Scalable System and Scalable Partition objects. Their statuses change to OK, as shown in Figure 9-26 on page 474.
Chapter 9. Management
473
Figure 9-26 Scalable complex with correctly configured access in IBM Systems Director
8. Discover the IP address of the operating system running on the primary node by navigating to Inventory System Discovery. Importing a common or platform agent: To deploy a common or platform agent via IBM Systems Director, you must import the agent first. See the following website for this procedure, because this procedure is out of the scope of this IBM Redbooks publication: http://publib.boulder.ibm.com/infocenter/director/v6r2x/index.jsp?topic=/com.ib m.director.agent.helps.doc/fqm0_t_working_with_agent_packages.html 9. Specify the IP address of the operating system under Select a discovery option. 10.Specify Operating System as the resource type to discover under the Select the resource type to discover list box. You must specify Operating System to enable IBM Systems Director to use the correct discovery protocols for the respective resource type. Click Discover Now when finished. 11.Rename the system if required and request access to it, as described in step 7 on page 469 to step 12 on page 472. 12.After the operating system is discovered, you need to deploy an agent to it to ensure that IBM Systems Director can manage the complex correctly. Select Navigate Resources All Operating Systems under the Groups (View Members) pane. 13.Right-click the operating system object that you have discovered and select Release Management Install Agent, as shown in Figure 9-27 on page 475.
474
14.The agent installation wizard window opens. Click Next. 15.Click the Platform Agent Packages group in the left pane and select the platform agent package that is relevant to your operating system. For Linux and WMware operating systems, select the PlatformAgent 6.2.1 Linux package. In this example, a Windows-based server was used; therefore, we selected the PlatformAgent 6.2.1 Windows package by clicking the radio button next to the agent and clicking Add. The selection options look like Figure 9-28 on page 476.
Chapter 9. Management
475
16.Click Next. 17.Click Next again. 18.Click Finish. 19.You are prompted to either schedule the deployment of the agent or deploy it immediately. For our example, we deployed the agent immediately. IBM Systems Director defaults to Run Now for all tasks unless specified otherwise. Click OK when ready to deploy the agent. Click Display Properties to view the status of the agent deployment process. Progress indicator: The progress indicator might remain on 5% for a period of time before its value increases, which is normal. 20.After the agent deployment has completed, inventory the operating system object by navigating to Navigate Resources All Operating Systems. Right-click the respective operating system object and click Inventory View and Collect Inventory. 21.Click Collect Inventory and click OK to run the task immediately. 22.Click Display Properties again to view the progress of the inventory task. 23.After the task has completed, you are ready to view and manage the scalable systems.
476
Figure 9-29 Selecting the correct resource type to discover a BladeCenter chassis
5. Click Discover Now. The discovered chassis appears in the Discovered Manageable Systems area at the bottom of the System Discovery pane. 6. Right-click the No access icon under the Access column and click Request Access. Tip: If the access status appears as Unknown, right-click the Unknown icon and select Verify Connection. Click Verify Connection. The status changes to No access if IBM Systems Director can communicate correctly with the IMM.
Chapter 9. Management
477
7. Enter the user name and password credentials of an account that has supervisor access to the AMM and click Request Access when finished. Wait time: Requesting access to a BladeCenter chassis might take time to complete. IBM Systems Director has to discover all the components within the chassis, including blades, I/O modules, power supplies, and so on. 8. Click Close after the process completes. You return to the System Discovery pane. 9. Close this pane and select Navigate Resources. In the Groups (View Members) pane, click Groups by System Type BladeCenter Systems BladeCenter Chassis and Members. All blade service processors (the IMM in the case of the HX5) and I/O module switches are displayed here. Notice the scaled HX5 that has been discovered in this group view, as shown in Figure 9-30. Also, note that the MAX5 does not show up as a separate item. The blade IBM:7872-AC1-06EC578 has a MAX5 attached. Installing a platform agent onto the operating system of the HX5 allows you to view the memory that is installed in the HX5, as well as the memory that is installed in the MAX5.
Figure 9-30 BladeCenter Chassis and Members (View Members) group view
You have completed the exercise of discovering a single-node HX5 out-of-band via IBM Systems Director.
478
Discovery tip: To correctly discover and manage a complex via IBM Systems Director, you must discover the system both out-of-band via the IMM and in-band via the operating system. A V6.2.1 or greater common or platform agent must also be installed. For the purposes of this exercise, we demonstrate how to install a platform agent on an HX5 blade running Windows Server 2008. The method of deploying a platform agent to a Linux or WMware ESX server is the same. Complete these steps to discover and deploy a platform agent to an HX5 complex: 1. Install Windows Server 2008 on the HX5 complex. We recommend using the IBM ServerGuide, because it installs all necessary drivers as part of the process. You can download the ServerGuide CD at the following website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-GUIDE 2. Ensure that you have configured the complex correctly by accessing the AMM and navigating to Scalable Complex Configuration. 3. On the Scalable Complex Information window, click the Complex tab that contains the scalable systems. The numbers in parentheses on the tabs refer to the slots that contain the blades, as shown in Figure 9-31.
4. Ensure that the Mode column status for the respective complex indicates Partition, as shown in Figure 9-31. If it does not indicate the Partition mode, see 8.6, Creating an HX5 scalable complex on page 402. 5. Complete the instructions in Discovering a 2-node HX5 via IBM Systems Director 6.2.x if you have not done so already. 6. You are still logged in to the IBM Systems Director web console at this point. Discover the IP address of the operating system running on the primary HX5 node by navigating to Inventory System Discovery.
Chapter 9. Management
479
Importing a platform agent: To deploy a platform agent via IBM Systems Director, you must import the agent first. See the following website for this import procedure, because it is beyond the scope of this IBM Redbooks publication: http://publib.boulder.ibm.com/infocenter/director/v6r2x/index.jsp?topic=/com .ibm.director.agent.helps.doc/fqm0_t_working_with_agent_packages.html 7. Specify the IP address of the operating system under Select a discovery option. 8. Specify Operating System as the resource type to discover under the Select the resource type to discover list box. You must specify Operating System to enable IBM Systems Director to use the correct discovery protocols for the respective resource type. Click Discover Now when finished. 9. Rename the system if required and request access to it, as detailed in step 7 on page 469 to step 12 on page 472. 10.After the operating system is discovered, you need to deploy an agent to it to ensure that IBM Systems Director can manage the complex correctly. Select Navigate Resources All Operating Systems under the Groups (View Members) pane. 11.Right-click the operating system object that you have discovered and select Release Management Install Agent. 12.The Agent Installation wizard window opens. Click Next. 13.Click the Platform Agent Packages group in the left pane and select the PlatformAgent 6.2.1 Windows package by clicking the radio box next to the agent. When deploying to a Linux or WMware ESX server, you need to select the PlatformAgent 6.2.1 Linux agent package. Click Add. Your selection options look like Figure 9-32.
480
17.You are prompted to either schedule the deployment of the agent or deploy it immediately. For our example, we deploy the agent immediately. IBM Systems Director defaults to Run Now for all tasks unless specified otherwise. Click OK when ready to deploy the agent. Click Display Properties to view the status of the agent deployment process. Progress indicator: The progress indicator might remain on 5% for a period of time before its value increases, which is normal. 18.After the agent deployment completes, inventory the operating system object by going to Navigate Resources All Operating Systems. Right-click the respective operating system object and click Inventory View and Collect Inventory. 19.Click Collect Inventory and click OK to run the task immediately. 20.Click Display Properties again to view the progress of the inventory task. 21.After the task completes, you are ready to view and manage the scalable systems.
Chapter 9. Management
481
www6.software.ibm.com www.ecurep.ibm.com
443 443
Tip: IP addresses are subject to change, so ensure that you use host names whenever possible. Collect inventory for the systems before you start the installation. This task is not mandatory, but it is recommended.
482
5. Select the language again that you want to use from the list box and click OK. 6. The Introduction pane appears. Accept the license terms if prompted and click Next. 7. Click the Restart IBM Systems Director check box if you want Service and Support Manager to be enabled immediately after installation, as shown in Figure 9-34 on page 484. Leave the Restart IBM Systems Director check box unchecked if you will restart the IBM Systems Director server at a later stage. Click Next when ready.
Chapter 9. Management
483
8. Click Install. The Service and Support Manager installer stops the IBM Systems Director server service. The installation completes. Click Done when finished. The IBM Systems Director server service starts after the installation of the Service and Support Manager completes. 9. Log in to the IBM Systems Director server console after the IBM Systems Director service has started. 10.Click the Manage tab in the right pane and scroll down to the Service and Support Manager plug-in, as shown in Figure 9-35 on page 485. The Service and Support Manager icon is blue, which indicates that additional configuration is required for this plug-in to operate.
484
11.Click Getting Started with Electronic Service Agent, as shown in Figure 9-35. 12.The Service and Support Manager wizard Welcome pane opens. Click Next. 13.Enter your company contact information in the provided fields. The more information that you provide, the easier it will be for IBM support to assist you. Pay particular attention to ensuring that the Country or region field is completed correctly, as shown in Figure 9-36 on page 486. 14.Click Next.
Chapter 9. Management
485
15.Provide default details for the physical location of your systems, as shown in Figure 9-37. You can change individual system location details at a later stage using the details that are provided in Changing the location settings of a system on page 488.
486
16.On the Connection pane, leave the default settings if your IBM Systems Director server has a direct connection to the Internet, as shown in Figure 17. Enter the proxy server details if you are required to connect to the Internet via a proxy server. Always ensure that you test the Internet connection by clicking Test Connection when finished. Click Next after a successful Internet connection has been confirmed.
17.Provide the authorized IDs of the personnel who need access to the service information that is transmitted to IBM. This information is not a requirement to activate Service and Support Manager. If you have not already created the IDs, you can do so by clicking the link, as shown in Figure 9-39.
18.On the Automatic monitoring pane, leave the check box checked if you want all newly discovered systems to be monitored by Service and Support Manager. Click Next. 19.Click Finish when done. You return to the Manage tab and the status of the Service and Support Manager plug-in is now green. The service now actively monitors all eligible systems that are monitored by IBM Systems Director. You can click the Service and Support Manager plug-in on the Manage tab to view systems that might have a serviceable problem. See the IBM Systems Director 6.2.x
Chapter 9. Management
487
Information Center, which is available at this website, for further details regarding Service and Support Manager: http://publib.boulder.ibm.com/infocenter/director/v6r2x/index.jsp
488
2. Click OK when the Task Launch Dialog pane appears to run the task immediately. The system powers on normally. The power control menus within IBM Systems Director are adaptive, which means that they change based on the power state of the system. Figure 9-41 shows the available options when the same system is powered on.
Figure 9-41 Power off controls for the operating system object
Chapter 9. Management
489
3. To view the installed software, right-click the respective scalable partition object, select Related Resources, and then select Installed Software. All software, including system drivers, is displayed for that system. You can also view additional information about the complex by either viewing the properties of the Scalable System or Scalable Partition objects. Use the following procedure: 1. Right-click either the Scalable System or Scalable Partition objects and select Properties from the menu. Figure 9-42 shows the properties of the Scalable Partition object indicating the number of nodes participating in the partition, as well as various other information.
490
3. Right-click the IMM of the server or blade that you want to identify and select System Status and Health System Identification LED On, as shown in Figure 9-43.
Figure 9-43 Identifying a system within the IBM Systems Director console
Tip: You can only perform system identification against the IMM object for IMM-based systems. You cannot perform this operation against the operating system, Scalable System object, or Scalable Partition object.
Chapter 9. Management
491
Consider the following items when updating the system firmware on a complex or scalable partition: Before starting any system firmware update processes, ensure that the multinode systems are discovered with both in-band and out-of-band methods. For more information, see the following sections for in-band and out-of-band configuration: 9.5.2, Discovering a 2-node x3850 X5 via IBM Systems Director 6.2.x on page 472 9.5.4, Discovering a 2-node HX5 via IBM Systems Director 6.2.x on page 478 The system firmware updates are installed to all the physical server systems within a system partition. All of the systems within the system partition are then rebooted after the installation. To update the firmware of either an x3850 X5 or an HX5 complex via IBM Systems Director, you need to configure Update Manager to download updates from the Internet. See Chapter 10, Update Manager, in Implementing IBM Systems Director 6.1, SG24-7694, which is available at this website: http://www.redbooks.ibm.com/abstracts/sg247694.html Use the following procedure to update the firmware of either a x3850 X5 or an HX5 complex via IBM Systems Director: 1. Log in to the IBM Systems Director console. 2. Select Navigate Resourcesand click the All Systems group. 3. Right-click the operating system object of the complex that you want to update and select Release Management Show needed updates. Any updates that are applicable to be deployed to the system will be displayed, as shown in Figure 9-44. For our example, we imported the latest UpdateXpress System Pack bundle for the HX5 into IBM Systems Director.
4. Click the check box next to the applicable update and click Install. 5. The update wizard window appears. Click Next. 6. Leave the default setting of Automatically install missing required updates on the Options window and click Next. 7. Update Manager will need to restart the systems that form the complex during the firmware update process. Leave Automatically restart as needed during installation as the default. Click Next.
492
8. A summary of the updates to be installed will be displayed, as shown in Figure 9-45. Click Finish when ready to proceed.
9. Click OK to run the firmware update immediately. The server will need to be rebooted so the update process can be scheduled if required. Wait for the firmware update process to complete, which might take time depending on the number of required updates to be installed.
Chapter 9. Management
493
The ESA inventory does not include the following information: Collection or transmission of any of your companys financial, statistical, or personnel data Client information Your business plans The web component of IBM Electronic Services offers a single location for you to access many IBM Internet service and support capabilities. You can also view and use the ESA inventory information from any location around the world. The Electronic Services website offers the following information: A single portal for hardware and software information and reference materials My Systems to view and use ESA service information in customized reports, such as hardware and software inventory, fixes, and system parameters My Search facility that uses ESA information to provide customized results for your specific machines from the IBM reference databases A single place to submit a service request for either hardware or software, in any country My Messages to receive information for specific platforms or individual profile definition Access to web-delivered premium services, such as Performance Management or Enhanced Technical Support (ETS) contracted services My Links to customize the web view by your selections of IBM system platforms Tutorials or demonstrations provided for all major areas of the website For information about downloading and installing the relevant ESA for your systems and registering on the IBM electronic services website, see the following website: https://www-304.ibm.com/support/electronic/portal/navpage?category=5
494
Chapter 9. Management
495
Use the following procedure to download, install, and connect to the IMM using a Windows operating system: 1. Create a directory named ASU. 2. Download the ASU Tool for your operating system (32-bit or 64-bit) at the following website and save it in the ASU directory: http://ibm.com/support/entry/portal/docdisplay?lndocid=TOOL-ASU 3. Unpack the utility: For Windows, double-click ibm_utl_asu_asutXXX_windows_yyyy.exe, which decompresses the following files into the ASU folder, as shown in Figure 9-46. 32-bit version: 64-bit version:
For Linux, open a terminal session and run the following command from the ASU directory: tar zxf ibm_utl_asu_asutXXX_linux_yyyy.tgz Figure 9-47 on page 497 shows how the files are installed.
496
64 bit
32 bit
4. Run a command, such as the asu show command, either in-band or out-of-band, using the commands that are listed in Table 9-4. This command confirms that the connection and utility work.
Table 9-4 Steps to run ASU both in-band and out-of-band In-band 1. Run the following command: - For 32 bit: asu show - For 64 bit: asu64 show Out-of-band 1. Ping the IMM to ensure that you have a network connection to the IMM. The default IP address is 192.168.70.125. Note: You might need to change your IP address. 2. Run the following command: - For 32 bit: asu show --host IPADRESS - For 64 bit: asu64 show --host IPADRESS
Example 9-1 shows the output from running the asu show command in-band.
Example 9-1 Output from the asu show command in-band IBM Advanced Settings Utility version 3.61.70I Licensed Materials - Property of IBM (C) Copyright IBM Corp. 2007-2010 All Rights Reserved Successfully discovered the IMM via SLP. Discovered IMM at IP address 169.254.95.118 Connected to IMM at IP address 169.254.95.118 IMM.SSH_SERVER_KEY=Installed
Chapter 9. Management
497
Default IMM user and password parameters: When you change the default user and password of IMM, you must specify the --user and --password parameters, for example: asu show --user USER --password PASSWORD
Link error
Example 9-2 shows a connection link error that might occur.
Example 9-2 Connection link error IBM Advanced Settings Utility version 3.61.70I Licensed Materials - Property of IBM (C) Copyright IBM Corp. 2007-2010 All Rights Reserved Connection link error.
To resolve this error, try the following steps: Ensure that your firewall allows ASU. Check if you have entered the correct IMM IP address in the command. Check if you can ping the IMM IP address. Restart the IMM.
To resolve this error, ensure that you have entered the correct user ID and password for the IMM. Remember that both the user ID and password are case-sensitive.
Permission error
Example 9-4 shows a permission error that might occur in Windows.
Example 9-4 User rights in Windows error IBM Advanced Settings Utility version 3.61.70I Licensed Materials - Property of IBM (C) Copyright IBM Corp. 2007-2010 All Rights Reserved IPMI command error. Please check your IPMI driver and IBM mapping layer installation.
Example 9-5 on page 499 shows a permission error that might occur in Linux. 498
IBM eX5 Implementation Guide
Example 9-5 User rights in Linux error IBM Advanced Settings Utility version 3.61.70I Licensed Materials - Property of IBM (C) Copyright IBM Corp. 2007-2010 All Rights Reserved User authority level is not sufficient. You must invoke ASU over the network (--host --user --password)
To resolve this error, try the following steps: Ensure that the user has administrator or root equivalent operating system privileges. When you use ASU to connect remotely to the system (out-of-band), the user must have adequate IMM rights (Supervisor).
Firewall error
Example 9-6 shows a firewall error that might occur.
Example 9-6 Firewall does not allow ASU - Linux IBM Advanced Settings Utility version 3.61.70I Licensed Materials - Property of IBM (C) Copyright IBM Corp. 2007-2010 All Rights Reserved SLP request fail. Try to configure usb-lan using DHCP. Unable to connect to IMM via LAN : Could not configure in band network. .............................. Failed to connect to local IMM using LAN IP and KCS
Chapter 9. Management
499
500
Enable Hyper-Threading
On the command line, enter asu set uefi.hyperthreading enable. See Example 9-13.
Example 9-13 Output uEFI.HyperThreading=Enable Waiting for command completion status. Command completed successfully.
Tip: If possible, use the latest version of IBM ServerGuide. ServerGuide can accelerate and simplify the installation of eX5 servers in the following ways: Assists with installing Windows-based operating systems and provides updated device drivers that are based on the detected hardware Reduces rebooting requirements during hardware configuration and Windows operating system installation, allowing you to get your eX5 server up and running sooner Provides a consistent server installation using IBM best practices for installing and configuring an eX5 server Provides access to additional firmware and device drivers that might not be applied at installation time, such as adapter cards that are added to the system later ServerGuide deploys the OS image to the first device in the boot order sequence. Best practices dictate that you have one device that is available for the ServerGuide installation
Chapter 9. Management
501
process. If you boot from SAN, make sure that you have only one path to the device because ServerGuide has no multipath support. See Booting from SAN on page 295 for more details. After the ServerGuide installation procedure, you can attach external storage or activate additional paths to the disk. For installation instructions about how to attach external storage or multipath drivers, see the respective User Guide. The following procedure describes how to install Windows Server 2008 R2 with ServerGuide. The method to install LInux is similar. Follow these steps: 1. Download the latest version of ServerGuide from the following website. http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-GUIDE 2. Burn a CD with the ISO image or mount the image through the IMM. 3. Boot the server from ServerGuide. 4. The progress bar appears, as shown in Figure 9-48.
5. After the files load, the Start window opens. Choose a language to continue. 6. Select your preferred keyboard layout and click Next. 7. Accept the license agreement and click Next. 8. The Welcome window opens and provides information about ServerGuide, which systems are supported, a readme file, and copyright and trademark information. Click Next. 9. Select the operating system that you want to install and click Next, as shown in Figure 9-49.
10.Enter the current date and time and click Next, as shown in Figure 9-50 on page 503.
502
11.Create a RAID configuration. If you previously created a RAID adapter, ServerGuide detects this configuration and displays it under Keep Current Adapter Configuration. Select a RAID configuration and click Next, as shown in Figure 9-51.
12.A Confirmation window opens indicating that RAID configuration is complete (Figure 9-52 on page 504). Click Next.
Chapter 9. Management
503
13.Restart the server. After the server restarts, the server will restart again to complete the ServerGuide setup, as shown in Figure 9-53. If you previously created a RAID configuration and kept the configuration, restarting the server is not necessary. Click Next.
14.After the server has restarted, you must create and format a partition. Choose your selection and click Next to start the process, as shown in Figure 9-54 on page 505.
504
15.When the process has completed, click Next, as shown in Figure 9-55.
Figure 9-55 Confirmation about completing the creation and formatting of the partition
17.ServerGuide copies the necessary files to the disk in preparation for the operating system installation, as shown in Figure 9-57 on page 506.
Chapter 9. Management
505
19.Insert the operating system installation DVD and click Next, as shown in Figure 9-59 on page 507. ServerGuide searches for the disc.
506
20.When ServerGuide successfully finds the installation medium, click Next, as shown in Figure 9-60.
21.The Windows setup installation procedure starts. Follow the installation procedure to complete the installation.
Chapter 9. Management
507
If used with IBM ServerGuide and IBM UpdateXpress, the ServerGuide Scripting Toolkit provides a total solution for deploying IBM System x and BladeCenter x86-based hardware in an unattended mode. The ServerGuide Scripting Toolkit enables you to create a bootable CD, DVD, or USB key that supports the following tasks and components: Network and mass storage devices Policy-based RAID configuration Configuration of system settings using Advanced Settings Utility (ASU) Configuration of fibre host bus adapters (HBAs) Local self-contained DVD deployment scenarios Local CD/DVD and network share-based deployment scenarios Remote Supervisor Adapter (RSA) II, Integrated Management Module (IMM), and BladeCenter Management Module (MM)/Advanced Management Module (AMM) remote disk scenarios UpdateXpress System Packs installation integrated with scripted network operating system (NOS) deployment IBM Director Agent installation, integrated with scripted NOS deployment The ServerGuide Scripting Toolkit, Windows Edition supports the following versions of IBM Systems Director Agent: Common Agent 6.1 or later Core Services 5.20.31 or later Director Agent 5.1 or later The Windows version of the ServerGuide Scripting Toolkit enables automated operating system support for the following Windows operating systems: Windows Server 2003, Standard, Enterprise, and Web Editions Windows Server 2003 R2, Standard and Enterprise Editions Windows Server 2003, Standard and Enterprise x64 Editions Windows Server 2003 R2, Standard and Enterprise x64 Editions Windows Server 2008, Standard, Enterprise, Datacenter, and Web Editions Windows Server 2008 x64, Standard, Enterprise, Datacenter, and Web Editions Windows Server 2008, Standard, Enterprise, and Datacenter Editions without Hyper-V Windows Server 2008 x64, Standard, Enterprise, and Datacenter without Hyper-V Windows Server 2008 R2 x64, Standard, Enterprise, Datacenter, and Web Editions The Linux version of the ServerGuide Scripting Toolkit enables automated operating system support for the following operating systems: SUSE Linux Enterprise Server 9 32 bit SP4 SUSE Linux Enterprise Server 9 x64 SP4 SUSE Linux Enterprise Server 10 32 bit SP1/SP2/SP3 SUSE Linux Enterprise Server 10 x64 SP1/SP2/SP3 SUSE Linux Enterprise Server 11 32 bit Base/SP1 SUSE Linux Enterprise Server 11 x64 Base/SP1 Red Hat Enterprise Linux 4 AS/ES 32 bit U6/U7/U8 Red Hat Enterprise Linux 4 AS/ES x64 U6/U7/U8 Red Hat Enterprise Linux 5 32 bit U1/U2/U3/U4/U5 Red Hat Enterprise Linux 5 x64 U1/U2/U3/U4/U5 VMware ESX Server 3.5 U4/U5 VMware ESX Server 4.0/4.0u1/4.0u2/4.1 508
IBM eX5 Implementation Guide
To download the Scripting Toolkit or the IBM ServerGuide Scripting Toolkit Users Reference, select the IBM ServerGuide Scripting Toolkit website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-TOOLKIT
Chapter 9. Management
509
For the HX5, from the AMM web interface, click in the left panel Blade Tasks Configuration Advanced Blade Policy Settings and ensure that the servers Ethernet over USB interface is enabled. See Figure 9-62.
510
partition from which they are executed. You must update the firmware for each blade independently before rebooting either system. With IMM firmware Version 1.15, a Dynamic Host Configuration Protocol (DHCP) server is included in the IMM. This DHCP assigns an IP address for the internal LAN over USB interface. Both LAN over USB interfaces in a scalable complex will get an IP address. You must enable DHCP on these interfaces in the operating system to enable firmware updating.
Chapter 9. Management
511
4. Accept the default Update the local machine, as shown in Figure 9-64. Click Next.
5. Accept the default Check the IBM web site, as shown in Figure 9-65. Click Next.
6. Select Latest available individual updates, as shown in Figure 9-66 on page 513. Click Next.
512
7. Select the directory in which you want to store the downloaded files, as shown in Figure 9-67. Click Next.
8. Enter the settings for an HTTP proxy server, if necessary, or leave the check box unchecked, as shown in Figure 9-68. Click Next.
9. A message displays showing that the UXSPI acquired the possible updates for the machine, as shown in Figure 9-69. Click Next.
10.A message appears showing that the download has completed, as shown in Figure 9-70. Click Next.
11.A component overview shows the components that need updating. UXSPI selects, by default, the components to update. Accept these settings and click Next.
Chapter 9. Management
513
12.When the update is finished, a message displays confirming the updates. Click Next.
13.Click Finish to close the UXSPI. 14.Restart the system to complete the update process.
4. Accept the license agreement. 5. The Welcome window opens, as shown in Figure 9-73 on page 516. Click Next.
Chapter 9. Management
515
7. Select Latest available individual updates and click Next, as shown in Figure 9-75 on page 517.
516
8. Enter the settings for an HTTP proxy server, if necessary, or check Do not use proxy, as shown in Figure 9-76. Click Next.
9. Select one or more machine types that are contained on the bootable media and click Next, as shown in Figure 9-77 on page 518.
Chapter 9. Management
517
10.Select the directory where you want to store the downloaded files, as shown in Figure 9-78. Click Next.
11.By default, BoMC creates an ISO file, as shown in Figure 9-79. You can choose another medium. Click Next.
12.Select Do not use unattended mode (Figure 9-80 on page 519) and click Next.
518
13.Review the selections and confirm that they are correct. Figure 9-81 shows an example. You can click Save to save this configuration information to a file. Click Next.
14.BoMC acquires the files. In the progress bar, you can see the progress of the updates, as shown in Figure 9-82 on page 520.
Chapter 9. Management
519
16.Click Finish.
17.You now have a bootable image with the updates. You can mount or burn the image to CD and then boot the system with the medium.
520
9.13.1 Installation
To download the latest MegaRAID Storage Manager software and obtain the Installation and Users Guide, select the ServeRAID software matrix website: http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-RAID The MSM supports the following operating systems: Microsoft Windows 2000, Microsoft Windows Server 2003, and Microsoft Windows Server 2008 Red Hat Enterprise Linux Version 4.0 and Version 5.0 SUSE SLES Version 9, Version 10, and Version 11, with the latest updates and service packs VMware 4.0, 4.0U1, and 4.1
Chapter 9. Management
521
To install the MSM software, read the Installation and Users Guide for MSM at the following website. http://ibm.com/support/entry/portal/docdisplay?lndocid=SERV-RAID You must have administrator or root equivalent operating system privileges to install and fully access the MegaRAID Storage software. There are four setup options: Complete: This option installs all program features. Client: This option installs the required components to remotely view and configure servers. Server: This option only installs the required components for remote server management. Stand Alone: This option only installs the required components for local server management.
522
Follow these steps to create a virtual drive with CLI: 1. Use the following command to locate the Enclosure Device ID and the Slot Number of both hard drives: MegaCli -PDList -aAll Example 9-15 shows the resulting output.
Example 9-15 Output of the MegaCli -PDList -aAll command Enclosure Device ID: 252 Slot Number: 1 Device Id: 8
Chapter 9. Management
523
Sequence Number: 4 Media Error Count: 0 Other Error Count: 0 Predictive Failure Count: 0 Last Predictive Failure Event Seq Number: 0 PD Type: SAS Raw Size: 136.731 GB [0x11176d60 Sectors] Non Coerced Size: 136.231 GB [0x11076d60 Sectors] Coerced Size: 135.972 GB [0x10ff2000 Sectors] Firmware state: Unconfigured(good), Spun Up SAS Address(0): 0x5000c50016f81501 SAS Address(1): 0x0 Connected Port Number: 1(path0) Inquiry Data: IBM-ESXSST9146803SS B53B3SD1GZCW0825B53B IBM FRU/CRU: 42D0422 FDE Capable: Not Capable FDE Enable: Disable Secured: Unsecured Locked: Unlocked Foreign State: None Device Speed: 6.0Gb/s Link Speed: 6.0Gb/s Media Type: Hard Disk Device Drive: Not Certified Enclosure Device ID: 252 Slot Number: 2 Device Id: 16 Sequence Number: 3 Media Error Count: 0 Other Error Count: 0 Predictive Failure Count: 0 Last Predictive Failure Event Seq Number: 0 PD Type: SAS Raw Size: 136.731 GB [0x11176d60 Sectors] Non Coerced Size: 136.231 GB [0x11076d60 Sectors] Coerced Size: 135.972 GB [0x10ff2000 Sectors] Firmware state: Unconfigured(good), Spun Up SAS Address(0): 0x5000c5001cf9c455 SAS Address(1): 0x0 Connected Port Number: 0(path0) Inquiry Data: IBM-ESXSST9146803SS B53B3SD2FM090825B53B IBM FRU/CRU: 42D0422 FDE Capable: Not Capable FDE Enable: Disable Secured: Unsecured Locked: Unlocked Foreign State: None Device Speed: 6.0Gb/s Link Speed: 6.0Gb/s Media Type: Hard Disk Device Drive: Not Certified
2. Now we can create the virtual drive. In our example, we issue the following command to create a RAID-1: MegaCli -CfgLDAdd -R1[252:1,252:2] -a0 Example 9-16 on page 525 shows the resulting output.
524
Example 9-16 Output from command MegaCli -CfgLDAdd -R1[252:1,252:2] -a0 Adapter 0: Created VD 1 Adapter 0: Configured the Adapter!! Exit Code: 0x00
Chapter 9. Management
525
In this section, we describe these topics: 9.14.1, Enabling SoL in UEFI on page 526 9.14.2, BladeCenter requirements on page 527 9.14.3, Enabling SoL in the operating system on page 529 9.14.4, How to start a SoL connection on page 533
Settings in UEFI for SoL: COM Port 1 has to be enabled only if the HX5 is used within a BC-H or BC-S where the Serial Pass-thru Module is implemented. Terminal Emulation can be set either to VT100 or ANSI; however, when configuring Linux operating systems, make sure that the OS settings match the terminal emulation that is selected in the hardware.
526
2. If the HX5 works properly with the management VLAN ID 4095, select Blade Tasks Power/Restart to verify that the Management Network icon is green, as shown in Figure 9-87.
Chapter 9. Management
527
3. Ensure that the SoL is enabled for the chassis. To verify, from the AMM, select Blade Tasks Serial Over LAN Serial Over LAN Configuration. Verify that Serial over LAN is Enabled, as shown in Figure 9-88.
4. Finally, check if the SoL for the HX5 is enabled. Select Blade Tasks Serial Over LAN Serial Over LAN Configuration and scroll down to the Serial Over LAN Status section. Ensure that the SOL Status is green (Figure 9-89).
528
Red circle icon: If you see a red circle icon next to the blade server, a requirement might not be satisfied. Or, you might have an issue with the Broadcom NIC drivers.
{bootmgr} partition=\Device\HarddiskVolume1 \EFI\Microsoft\Boot\bootmgfw.efi Windows Boot Manager en-US {globalsettings} {current} {87209f03-3477-11e0-a416-a69aee999ac5} {current} {memdiag} 30
{current} partition=C: \Windows\system32\winload.efi Windows Server 2008 R2 en-US {bootloadersettings} {87209f05-3477-11e0-a416-a69aee999ac5} Yes partition=C: \Windows {87209f03-3477-11e0-a416-a69aee999ac5} OptOut
Chapter 9. Management
529
C:\Users\Administrator>
3. Enter the command bcdedit /ems on. Example 9-18 shows the output from this command.
Example 9-18 Output of bcdedit /ems on C:\Users\Administrator>bcdedit /ems on The operation completed successfully.
4. Modify the EMS settings to match the parameters that were configured at the hardware level with the following command: bcdedit /emssettings emsport:2 emsbaudrate:115200
Example 9-19 Output of the bcedit /emssettings emsport:2 emsbaudrate:115200 command C:\Users\Administrator>bcdedit /emssettings emsport:2 emsbaudrate:115200 The operation completed successfully.
{bootmgr} partition=\Device\HarddiskVolume1 \EFI\Microsoft\Boot\bootmgfw.efi Windows Boot Manager en-US {globalsettings} {current} {87209f03-3477-11e0-a416-a69aee999ac5} {current} {memdiag} 30
{current} partition=C: \Windows\system32\winload.efi Windows Server 2008 R2 en-US {bootloadersettings} {87209f05-3477-11e0-a416-a69aee999ac5} Yes partition=C: \Windows {87209f03-3477-11e0-a416-a69aee999ac5} OptOut Yes
530
IMM setting
You can change the CLI mode for the COM port for EMS. Use the following procedure: 1. Log in to the web interface of IMM. 2. Navigate to IMM Control Serial Port. 3. Change the CLI mode to CLI with EMS compatible keystroke sequences (Figure 9-90).
4. Click Save to save the changes. For more information about Microsoft Emergency Messaging Service and the Special Administration Console, see the following documents: Boot Parameters to Enable EMS Redirection http://msdn.microsoft.com/en-us/library/ff542282.aspx Special Administration Console (SAC) and SAC commands http://msdn.microsoft.com/en-us/library/cc785873
3. Examine the output. If more than one boot entry exist, you must determine the default entry. 4. Enable EMS with the bootcfg /ems on /port com2 /baud 115200 /id 1 command. In our example, the default boot entry has the ID 1 (Example 9-22 on page 532).
Chapter 9. Management
531
Example 9-22 Output of the bootcfg /ems on /port com2 /baud 115200 /id 1 command C:\>bootcfg /ems SUCCESS: Changed SUCCESS: Changed SUCCESS: Changed on /port com2 /baud 115200 /id 1 the redirection port in boot loader section. the redirection baudrate in boot loader section. the OS entry switches for line "1" in the BOOT.INI file.
IMM Setting
To change the CLI mode for the COM port for use with EMS, use the following procedure: 1. Log in to the web interface of IMM. 2. Navigate to IMM Control Serial Port. 3. Change the CLI mode to CLI with EMS compatible keystroke sequences.
4. Click Save to save the changes. For more information about Microsoft Emergency Messaging Service and the Special Administration Console, see the following documents: Boot Parameters to Enable EMS Redirection http://msdn.microsoft.com/en-us/library/ff542282.aspx Special Administration Console (SAC) and SAC commands http://msdn.microsoft.com/en-us/library/cc785873
532
Linux
You must edit two files in Linux to ensure that the console redirection still works after the operating system has loaded. The same files are changed for Red Hat Linux (RHEL) and SUSE Linux. Edit these files: /boot/grub/menu.lst /etc/inittab RHEL 6: If you have installed RHEL 6 in UEFI mode, you must edit the /boot/efi/EFI/redhat/grub.conf file instead of the /boot/grub/menu.lst file.
Menu.lst or grub.conf
Add the parameter that is highlighted in bold in the file that is shown in Example 9-24.
Example 9-24 Example of the grub.conf # grub.conf generated by anaconda # # Note that you do not have to rerun grub after making changes to this file # NOTICE: You have a /boot partition. This means that # all kernel and initrd paths are relative to /boot/, eg. # root (hd0,1) # kernel /vmlinuz-version ro root=/dev/mapper/VolGroup-lv_root # initrd /initrd-[generic-]version.img #boot=/dev/sda1 device (hd0) HD(1,800,64000,699900f5-c584-4061-a99f-d84c796d5c72) default=0 timeout=5 splashimage=(hd0,1)/grub/splash.xpm.gz hiddenmenu title Red Hat Enterprise Linux (2.6.32-71.el6.x86_64) root (hd0,1) kernel /vmlinuz-2.6.32-71.el6.x86_64 ro root=/dev/mapper/VolGroup-lv_roo t rd_LVM_LV=VolGroup/lv_root rd_LVM_LV=VolGroup/lv_swap rd_NO_LUKS rd_NO_MD rd_N O_DM LANG=en_US.UTF-8 SYSFONT=latarcyrheb-sun16 KEYBOARDTYPE=pc KEYTABLE=de cras hkernel=auto console=ttyS1,115200n8 rhgb quiet initrd /initramfs-2.6.32-71.el6.x86_64.img [root@localhost redhat]#
/etc/inittab
Add the parameter that is highlighted in bold at the end of the /etc/inittab file, as shown in Example 9-25.
Example 9-25 The /etc/inittab file id:3:initdefault: co:2345:respawn:/sbin/agetty ttyS1 115200 vt100-nav [root@localhost etc]#
Chapter 9. Management
533
4. Start SoL with the console 1 command. The SoL console starts and you see whatever is transmitted over the SoL connection (for example, the UEFI setup windows). See Sample output for examples of what you see over an SoL connection when the system boots.
4. Start SoL with the console -T system:blade[5] command. Determining the slot number: In this example, we started a SoL session with the blade in slot 5. To determine the slot number of your blade, select Blade Tasks Power/Restart. The SoL console starts and you see whatever is transmitted over the SoL connection (for example, the UEFI setup windows). See Sample output for examples of what you see over an SoL connection when the system boots.
534
Sample output
After you connect to the system using SoL and power on the server, you see the output in the telnet/SSH window. Figure 9-94 shows the x3850 X5 booting.
Platform Initization Complete System x3850 X5 UEFI Builde Ver: 1.40 IMM Build Ver: 1.24 Diagnostics Build Ver: 3.30
2 CPU Packages Available at 2.00 GHz Link Speed 16 GB Memory Available at 1067 MHz in Lockstep Mode
Connecting Boot Devices and Adapters... Figure 9-94 Booting the x3850 X5
Figure 9-95 shows the logo window via SoL. The IBM logo is not displayed.
<F1> Setup
<F2> Diagnostics
Starting Windows... For troubleshooting and advanced startup options for Windows, press F8. Figure 9-96 Windows booting as seen via SoL
Chapter 9. Management
535
536
537
IP IPMB IPMI IS ISO IT ITSO JBOD KB KCS KVM LAN LDAP LED LGA LPD LUN MAC MB MCA MDIX MESI MIPS MM MSM NAS NHS NIC NMI NOS NTLM NUMA OGF OS PCI PD PDSG PE PFA
Internet Protocol Intelligent Platform Management Bus Intelligent Platform Management Interface information store International Organization for Standards information technology International Technical Support Organization just a bunch of disks kilobyte keyboard console style keyboard video mouse local area network Lightweight Directory Access Protocol light emitting diode land grid array light path diagnostic logical unit number media access control megabyte Machine Check Architecture medium-dependent interface crossover modified exclusive shared invalid millions of instructions per second Management Module MegaRAID Storage Manager network-attached storage non-hot-swap network interface card non-maskable interrupt network operating system NT LAN Manager Nonuniform memory access Open Grid Forum operating system Peripheral Component Interconnect problem determination Problem Determination and Service Guide Preinstallation Environment Predictive Failure Analysis
Project Management Institute power-on self test Personal System Preboot eXecution Environment QuickPath Interconnect redundant array of independent disks random access memory remote access services; row address strobe; reliability, availability, and serviceability Remote Direct Memory Access Remote Electronic Technical Assistance Information Network Red Hat Enterprise Linux reduced instruction set computer RAID-on-card read-only memory revolutions per minute Remote Supervisor Adapter Receive-side scaling real-time clock request to send Special Administration Console storage area network Serial Attached SCSI Serial Advanced Technology Attachment Small Computer System Interface static dynamic RAM self-encrypting drive small form-factor pluggable Single Level Cell SUSE Linux Enterprise Server Service Location Protocol Systems Management Architecture for Server Hardware scalable memory interconnect symmetric multiprocessing Simple Network Management Protocol service-oriented architecture Serial over LAN ServerProven Opportunity Request for Evaluation short range
RDMA RETAIN RHEL RISC ROC ROM RPM RSA RSS RTC RTS SAC SAN SAS SATA SCSI SDRAM SED SFP SLC SLES SLP SMASH SMI SMP SNMP SOA SOL SPORE SR
538
SSCT SSD SSH SSIC SSL STG TB TCG TCO TCP TCP/IP TDP TFTP TOE TPM UDP UE UEFI UPS URL USB UXSP UXSPI VFA VLAN VLP VMFS VPD VRM VSMP VT
Standalone Solution Configuration Tool solid-state drive Secure Shell System Storage Interoperation Center Secure Sockets Layer Server and Technology Group terabyte Trusted Computing Group total cost of ownership Transmission Control Protocol Transmission Control Protocol/Internet Protocol thermal design power Trivial File Transfer Protocol TCP offload engine Trusted Platform Module user datagram protocol Unrecoverable Error Unified Extensible Firmware Interface uninterruptible power supply Uniform Resource Locator universal serial bus UpdateXpress System Packs UpdateXpress System Packs Installer Virtual Fabric Adapter virtual LAN very low profile virtual machine file system vital product data voltage regulator module Virtual SMP Virtualization Technology
539
540
Related publications
The publications listed in this section are considered particularly suitable for a more detailed discussion of the topics covered in this book.
541
Vista Deployment Using Tivoli Provisioning Manager for OS Deployment, REDP-4295 You can search for, view, or download IBM Redbooks, Redpapers, Webdocs, draft publications and additional materials, as well as order hardcopy IBM Redbooks publications, at this website: ibm.com/redbooks
Other publications
Publications listed in this section are also relevant as further information sources.
Online resources
The following websites are also relevant as further information sources: IBM eX5 home page http://ibm.com/systems/ex5 IBM System x3850 X5 home page http://ibm.com/systems/x/hardware/enterprise/x3850x5 IBM BladeCenter HX5 home page http://ibm.com/systems/bladecenter/hardware/servers/hx5 IBM System x3690 X5 home page http://ibm.com/systems/x/hardware/enterprise/x3690x5 IBM System x Configuration and Options Guide http://ibm.com/support/docview.wss?uid=psg1SCOD-3ZVQ5W IBM ServerProven http://ibm.com/systems/info/x86servers/serverproven/compat/us/
Related publications
543
544
Index
Numerics
10Gb Ethernet HX5 183 x3690 X5 166 x3850 X5 104 39R6529 101, 162 39R6531 101, 162 39Y9306 211 41Y8278 175, 214 41Y8287 175, 214 42C1830 211 42D0632 92, 147 42D0637 92, 147 42D0672 92, 147 42D0677 92, 147 42D0709 147 42D0747 92 42D0752 92 43W4068 210 43W7714 92, 147 43W7734 95, 205 43W7735 150 44E8689 91, 149 44E8837 156 44T1486 194 44T1592 7879, 133 44T1596 194 44T1599 7879, 133 44W2266 147 44W2296 147 44W4465 211 44W4466 211 44W4475 210 44W4479 211 44X1940 211 44X1945 210 46C3447 107, 168 46C7448 78, 133 46C7482 7879, 133 46C7483 7879, 133 46C7499 194 46M0071 77 46M0072 72 46M0829 91, 149 46M0830 101, 162 46M0831 91, 149 46M0832 149 46M0901 102 46M0902 103 46M0914 151 46M0916 92, 149 46M0930 92, 101, 149, 162 46M0969 92, 149 46M6001 211 46m6003 46M6065 46M6140 46M6164 46M6168 46M6730 46M6863 46M6873 46M6908 46M6955 46M6960 46M6973 46M6975 46M6995 49Y4216 49Y4218 49Y4235 49Y4250 49Y4300 49Y4301 49Y4302 49Y4303 49Y4304 49Y4305 49Y4306 49Y4379 59Y5859 59Y5877 59Y5889 59Y5899 59Y5904 59Y5909 59Y6103 59Y6135 59Y6213 59Y6265 59Y6267 59Y6269 60Y0309 60Y0311 60Y0312 60Y0313 60Y0314 60Y0315 60Y0316 60Y0317 60Y0318 60Y0319 60Y0320 60Y0321 60Y0323 60Y0327 60Y0329 60Y0331 60Y0332 211 210 210 211 211 208 192 192 203 192 192 187 189 192 107, 168 107, 168 211 105, 167 74 74 74 74 74 74 74 68 192 187, 191 185 192 192 192 74 91 93 69, 127 69, 71 127, 129 157 130 130 130 130 130 130 130 130 130 130 130 132 173 165 165 69, 113, 127, 173174
545
60Y0333 148 60Y0337 165 60Y0339 146 60Y0360 150 60Y0366 165 60Y0381 146 675W Redundant Power Supply Kit 173 69Y2322 146 69Y2323 146 69Y2344 176 69Y2345 176 69Y2347 176 69Y4389 176 69Y4390 176 69Y4403 176 6Gb SSD Host Bus Adapter x3850 X5 99, 155 81Y4426 92, 149
benchmarks 24 BIOS See UEFI HX5 396 x3690 X5 337 x3850 X5 259 BladeCenter H HX5 considerations 383 BladeCenter HX5 See HX5 blue handle 100 Bootable Media Creator 514 Boxboro 62 BPE4 208
C
cable management arms x3690 X5 176 cabling x3690 X5 to MAX5 129, 313 x3850 X5 to MAX5 71, 232 x3850 X5 two-node 71, 237 cache coherency filter 19 Call Home 459 Chipkill 29 chipset 16 CMOS memory HX5 375 x3690 X5 302 x3850 X5 220 comparison 7 complex state data 381 consolidation 123 cooling x3690 X5 171 x3850 X5 112 CPU See Intel See processors
A
Active Energy Manager 42 Advanced Management Module See AMM Advanced Settings Utility 495 allowInterleavedNUMAnodes HX5 417 x3690 X5 355, 358 x3850 X5 273, 277, 279 AMM 454462 See also IMM firmware updates 461 functions 454 login 456 partitioning 34 redundancy 455 remote control 465 Serial over LAN enable 527 Service Advisor 458 System Status 457 tasks 455 ASU 495 credentials 498 examples of use 499 firewall 499 IMM 495 install 496 link error 498 password 498 permissions 498 tasks 495
D
database models 3 database systems 181 defaults HX5 375 DIMM sparing 29 DIMMs per channel HX5 179, 181 MAX5 186 x3690 X5 132 x3850 X5 84 disks HX5 203 x3690 X5 145 x3850 X5 90 downloads ASU 495 Bootable Media Creator 514 MegaRAID Storage Manager 521
B
backplanes x3690 X5 158 x3690 X5 1.8-inch x3690 X5 2.5-inch x3850 X5 1.8-inch x3850 X5 2.5-inch battery backup 97 150 145 93 91
546
ServerGuide 502 UpdateXpress System Pack Installer 511 DVD drive x3690 X5 163 x3850 X5 102 Dynamic Infrastructure 1 Dynamic System Analysis 448
E
Electronic Service Agent 481, 493 Emergency Messaging Service 529 Emulex 10Gb Ethernet Adapter HX5 183 x3690 X5 166 x3850 X5 104 energy efficiency 10, 21 ESXi 50 Ethernet HX5 212 x3690 X5 170 x3850 X5 109 eX5 chipset 16 EXA 16 eXFlash 4748, 149 x3690 X5 149 x3850 X5 93 expansion blade 208 external storage HX5 207 x3690 X5 162 x3850 X5 101
F
factory defaults HX5 375 x3690 X5 302 x3850 X5 220 FAMM 28 fans MAX5 113 x3690 X5 171 x3850 X5 112 features HX5 178 x3690 X5 118 x3850 X5 56 firmware updates 509 full array memory mirroring 28
G
Giga transfers 22 GPU adapters 165 green 21
H
hemisphere mode 26 HX5 197 x3690 X5 134
x3850 X5 78 Hot Spare 522 HS22 9 HX5 177216, 373395 10Gb Ethernet 183 2-node Scalability Card 189 across two power domains 444 allowInterleavedNUMAnodes 417 AMM 455 AMM troubleshooting 443 architecture 184 before you begin 374 BIOS 212, 396 BladeCenter H 383 block diagram 185, 191 boot from SAN 440 boot manager 399 BPE4 208 Broadcom 212 CFFh 210 chassis support 182 CIOv 209 CMOS memory, clearing 375 compare with HS22 9 complex state data 381 configurations 178 configuring RAID 205 CPUs 192 create a partition 403 crossing power domains 385 database 181 defaults 375 device drivers 439 DIMM placement 196 discovery 477 disks 203 double-wide 188 downloads 438 energy settings 398 Ethernet controller 212 event log 444 expansion blade 208 expansion cards 209 factory defaults 375 features 178 firmware 379 firmware updates 438 HS22 and HS22V comparison 180 Hyper-Threading 184 I/O cards 209 IM volume 390 IME volume 390 IMM 452 Integrated Management Module 213 Intel processors 192 internal view 179 introduction 178 IPv6 212 IS volume 390 jumpers 375
Index
547
latency settings 398 layout 179 light path diagnostics 443 local storage 385 log 444 LSI configuration utility 205, 386 LSI controller 207 MAX5 180, 186, 401 block diagram 187 population order 198 MAX5 and ESXi 418 media to install 407 memory 194 balance 199 block diagram 195 buffers 194 DIMM locations 195 DIMMs 194 MAX5 186 minimums 377 mirroring 200 population order 196 power domains 199 redundant bit steering 202 sparing 202 speed 196 models 6, 183 MR10ie 204 non-pooled mode 401 operating system installation 407 operating systems 215 out-of-band management 452 partition 403 partitioning 214 PCI Express Gen 2 Expansion Blade 208 PCIe full-length slots 208 PDSG 443 performance 196 pooled mode 401 POST event log 444 power consumption 383 power domains 199, 384, 444 power jumper cap 382 power loss 444 power sharing cap 374 power to a partition 405 power-on 374 processor scaling 377 processors 192 PXE boot 212, 407, 415 QPI scaling 189 RAID configuration 205 Red Hat install 436 remote control 465 remote media 407 SAN 440 SAS configuration utility 205, 387 SAS Connectivity Card (CIOv) 207 SAS Controller 203 scalability 188
scalability card 406 Scalability Kit 189 scalable complex 402 scale prerequisites 377 Serial over LAN 525 ServeRAID MR10ie 204 ServerGuide 501 ServerGuide Scripting Toolkit 410 single-wide 188 SLES install 437 SMI link speed 193 solid state drives 204 specifications 178 Speed Burst Card 185, 394, 444 split power domains 444 SSD Expansion Card 203 SSDs 204 standalone mode 405 storage 203 straddling power domains 385 system board 179 system event log 444 TCP Offload Engine 212 troubleshooting 406 Trusted Platform Module 213 Turbo Boost 184 UEFI 212, 396 boot manager 399 latency 398 MAX5 401 non-pooled mode 401 performance 397 pooled mode 401 settings 398 startup 398 USB 214, 409 USB flash drive option 214 video 213 virtualization 214 virtualization model 7 VMware ESXi 214, 415, 421 Wake on LAN 212 Windows install 410, 434 workload-optimized model 7 workloads 181 Hyper-Threading 17
I
I/O hub 30 IBM Systems Director 467 collect inventory 472 deploy packages 475 discovery 468, 472 Electronic Service Agent 481 HX5 discovery 477 HX5 two-node 478 IBM Electronic Service Agent 493 identify systems 490 install Director agent 475 inventory 472, 489
548
login 469 no access 471 partitions 473 power control 488 renaming systems 470 Request Access 471 scalable partitions 473 Service and Support Manager 481 IBM ToolsCenter 448 IMM 449454 See also AMM ASU 495 ASU commands 500 discovery 468 features 449 firmware updates 454 HX5 213 in-band management 453 login 452 out-of-band management 449 remote control 462 Serial over LAN 531 shared port 450 x3690 X5 169 x3850 X5 110 in-band management 453 integrated management module See IMM integrated virualization 50 Intel See also processors 16 cache coherency filter 19 hemisphere mode 26 Hyper-Threading Technology 17 Machine Check Architecture 30 MESI 20 QPI 16, 18 scalable memory buffers 30 SMI link 22 snoop filter 19 snoop protocol 16 Turbo Boost 18 Virtualization Technology 17 Xeon 5500/5600 comparison 16 Xeon 6500 processors 16 Xeon 7500 processors 16 IPv6 HX5 212 x3690 X5 167 x3850 X5 106
L
lab services 11 light path diagnostics HX5 213, 443 x3690 X5 170, 369 x3850 X5 111, 298 Linux HX5 436437 Serial over LAN 533 x3690 X5 340, 358 x3850 X5 291 local memory 26 lock step mode 30 LSI MegaRAID Storage Manager 521 SAS configuration utility 205
M
Machine Check Architecture 30 management 447 AMM 454 ASU 495 Bootable Media Creator 514 Call Home 459 Electronic Service Agent 493 firmware updates 454, 509 IBM Systems Director 467 IMM 449 in-band management 453 MegaCLI 523 MegaRAID Storage Manager 521 port 451 remote control 462 Serial over LAN 525 ServerGuide 501 ServerGuide Scripting Toolkit 507 shared port 451 UpdateXpress System Pack Installer 511 Matrox video 62 MAX5 31 bandwidth 33 cables 231 DIMMs 233 features of the rack version 60 firmware levels 230 HX5 186187 installing in a rack 231 non-pooled mode 139 performance 33 pooled mode 139 power 174 power supplies 113 See also x3850 X5 x3690 X5 121 x3850 X5 59, 230 MCA 30 MegaCLI 523 MegaRAID Web BIOS 94
J
jumpers HX5 375
K
KVM hypervisor 50
Index
549
MegaRAID Storage Manager 521 drive states 522 installation 521 MegaCLI 523 virtual drive states 523 MemCached 52 memory 22 Chipkill 29 DIMM placement 23 Giga transfers 22 hemisphere mode 26 HX5 194 lock step mode 30 mirroring 28 performance 23 QPI link 22 ranks 24 redundant bit steering 29 scalable memory buffers 30 SMI link 22 sparing 29 speed 22 UEFI settings 45 x3690 X5 131 x3850 X5 76 memory sparing x3690 X5 143 x3850 X5 89 memory speed 1333 MHz 23 MESI 20 mirroring 28 HX5 200 x3690 X5 141 x3850 X5 87 models 3 HX5 183 x3690 X5 124 x3850 X5 64 multiburner x3690 X5 163 x3850 X5 102
PCI Express Gen 2 Expansion Blade 208 PCIe slots HX5 209 expansion blade 208 x3690 X5 164 x3850 X5 103 PciROOT 248 performance 24 memory 23 x3690 X5 140 x3850 X5 84 positioning 7 power modes 38 power supplies x3690 X5 173 x3850 X5 112, 249 processors 16 energy efficiency 21 hemisphere mode 26 Hyper-Threading 17 Intel VT 17 Machine Check Architecture 30 NUMA 26 scalable memory buffers 30 SMI link 22 UEFI settings 43 x3690 X5 130 x3850 X5 74
Q
QPI Wrap Card 66 QuickPath Interconnect 18 See also Intel
R
RAID controllers HX5 203 x3690 X5 149 x3850 X5 91 rank sparing 29 ranks 24 RAS features 28 RBS See redundant bit steering Red Hat HX5 436 KVM hypervisor 50 Serial over LAN 533 x3690 X5 340, 358 x3850 X5 291 Redbooks Web site 542 Contact us xvii redundant bit steering 2930 HX5 202 x3690 X5 137 x3850 X5 79 remote control 462 remote memory 26 riser cards, x3690 X5 164
N
node controllers 16 NUMA 26
O
operating system install HX5 407 x3850 X5 263 optical drive x3690 X5 163 x3850 X5 102 out-of-band management 449
P
partitioning HX5 214
550
S
SAS drives HX5 203 x3690 X5 145 x3850 X5 90 SAS Expansion Adapter 146 scalability 33 HX5 188 scalable memory buffers 30 self-encrypting drives x3690 X5 148 Serial over LAN 525 connecting 533 enabling 526 telnet 534 ServeRAID MegaRAID Storage Manager 521 ServeRAID B5015 x3690 X5 155 x3850 X5 100 ServeRAID Expansion Adapter 146 ServeRAID M1015 x3690 X5 154 x3850 X5 99 ServeRAID M5014 x3690 X5 153 x3850 X5 97 ServeRAID M5015 x3690 X5 153 x3850 X5 97 ServeRAID M5025 x3690 X5 162 x3850 X5 98, 101 ServeRAID support HX5 204 x3690 X5 149 x3850 X5 91 ServeRAID-BR10i 96 x3690 X5 153 ServerGuide 501 download 502 minimum versions 501 using to install an OS 502 ServerGuide Scripting Toolkit 507 HX5 410 ServerProven 108, 211 Service Advisor 458 Service and Support Manager 481 services offerings 11 shared management port 451 SLES See SUSE Linux Smarter Planet 1 snoop filter 19 snoop protocol 16 SoL 525 solid state drives See also eXFlash HX5 204 SSD Expansion Card 203
use of cache 52 x3690 X5 149 x3850 X5 93 sparing 29 HX5 202 x3690 X5 143 x3850 X5 89 Special Administration Console 529 Speed Burst Card 185, 394 SSD See solid state drives SSD Expansion Card 203 SSH 534 Start Now Advisor 448 storage HX5 203 x3690 X5 145 x3850 X5 90 Storage Configuration Manager 448 SUSE Linux HX5 437 Serial over LAN 533 x3690 X5 340 x3850 X5 292 swap file 51 System x3690 X5 See x3690 X5 System x3850 X5 See x3850 X5 Systems Director 467
T
technology 15 telnet 534 tools 448 ToolsCenter 448 transceivers 107, 168 Trusted Platform Module HX5 213 x3690 X5 170 x3850 X5 111 Turbo Boost 18
U
UEFI 36 Acoustic Mode 39 acoustic mode 21 Active Energy Manager 42 ASU commands 500 C1 Enhanced Mode 41, 44 CKE Low Power 41, 47 Cores in CPU 45 CPU C-States 41, 44 Custom Mode 41 efficiency mode 21 energy efficiency 21 Execute Disable Bit 45 firmware updates 509 Halt On Severe Error 42
Index
551
HX5 212, 396 Hyper-Threading 45 Intel Virtualization Technology 45 main window 37 Mapper Policy 47 MAX5 Memory Scaling Affinity 47 Memory Data Scrambling 47 Memory Mirror Mode 46 Memory Spare Mode 46 Memory Speed 41, 46 operating modes 21 Page Policy 47 Patrol Scrub 47 Performance Mode 40 performance mode 21 power modes 38 Power Restore Policy 42 Proc Performance States 41 Processor Data Prefetch 45 Processor Performance States 44 processors 43 QPI Link Frequency 42, 45 Quiet Boot 42 Scheduler Policy 47 Spread Spectrum 47 system settings 37 Turbo Boost Power Optimization 42, 44 Turbo Mode 42, 44 updates 509 x3690 X5 337 x3850 X5 110, 259 Unconfigured Good 522 UpdateXpress System Pack Installer 511 USB HX5 214 x3690 X5 170, 174 UXSPI 511
W
Windows x3850 X5 266 Windows Server 2003 Serial over LAN 531 Windows Server 2008 HX5 434 Serial over LAN 529 x3690 X5 356 x3850 X5 289 workload-optimized models HX5 7 x3690 X5 5, 125 x3950 X5 3
X
x3650 M3 9 x3690 X5 117176, 301336 10Gb Ethernet 166 24 drives 162 675W Redundant Power Supply Kit 173 6Gb SSD Host Bus Adapter 155 acoustic mode 341 adapter ROM 324 adapter throughput 316 architecture 126 backplane cables 146 backplane for 1.8" drives 149 backplanes 145 battery cable 156 battery placement 155 battery trays 156 before you power on 302 block diagram 126 boot from SAN 368 Boot Manager 319 boot options 343 boot order 320, 322 cable management arms 176 cables, scalability 129 cabling 314 CMOS memory, clearing 302 combinations of drives 158 compare with x3650 M3 9 consolidation 123 controllers 149 cooling 171 CPUs 130 custom mode 341 database model 5 database workloads 123 dedicated SM port 329 depth 119 device drivers 366 diagnostics 307 dimensions 119 disk controllers 149 disk drives 147 drive combinations 158
V
Virtual Fabric Adatper 105 virtualization energy management 21 Virtualization Technology (VT) 17 VMware ESX HX5 install 421 x3690 X5 install 362 x3850 X5 290 x3850 X5 install 275 VMware ESXi 50 factory load 420 HX5 214, 415, 421 HX5 install 415, 421 recovery CD 420 updating the memory key 274 x3690 X5 install 358 x3850 X5 271, 290 x3850 X5 install 271, 275
552
DVD drive 163 Dynamic System Analysis 370 efficiency mode 341 Electronic Service Agent 370 embeded hypervisor 355 Emulex 10Gb Ethernet Adapter 166 energy efficiency 341 Ethernet 170 event log 334, 370 eXFlash 149 external storage 162 fans 171 features 118 firmware 365 Friction Slide kit 176 front view 119 full length slots 165 GPU adapters 165 height 119 High Efficiency 675W Power Supply 173 HPC settings 342 Hyper-Threading 130 I/O adapters 168 IMM 169, 327, 369 dedicated port 329 event log 334 light path diagnostics 333 log 334 network access 328 password 328 remote media 348 system status 331 troubleshooting 330 userid 328 integrated hypervisor 355 internal view 120 Legacy Thunk support 318 light path diagnostics 170, 333, 369 line cords 173 Linux 340, 355, 358 log 334 logs 370 low latency settings 342 management 169 management module 327 management port 450 MAX5 121, 127, 311 accessing memory 314 block diagram 127 cables 129, 312 firmware 311 internal view 123 population order 136 ports 122 power 174 VMware 139 media 346 memory 131 balance 139 by processor 132
considerations 306 DIMMs 133 fault tolerance 310 hemisphere mode 307 matched pairs 133 MAX5 121, 136 mezzanine 132 mirroring 141 non-pooled 339 performance 306 pooled 339 pooled mode 139 population order 133 redundant bit steering 137 sparing 143 speed 133 testing 307 VMware 135 minimum processors 304 models 4, 124 multiburner 163 non-UEFI adapters 318 operating systems 175 optical drive 163 OS install 346 out-of-band management 449 PCIe adapters 168, 316 PCIe slots 164, 321 performance 140, 144, 305 performance mode 341 power supplies 173, 326 power/performance 340 powering on 302 processor considerations 304 processors 130 product features 118 PXE boot 323 PXE network boot 346 QPI cables 312 QPI Link Frequency 316 QPI ports 126 rack mounting 176 rails 176 rear view 120 recovery 371 Red Hat 340, 358 Red Hat support 175 redundant power 326 remote media 348 resetting the server 302 riser cards 164, 316, 321 SAN 367 SAS backplanes 145 SAS controllers 149 SAS drives 145 SAS Expansion Adapter 146, 157 SATA drive 148 scalability 128 self-encrypting drives 148 Serial over LAN 525
Index
553
ServeRAID B5015 155 ServeRAID controllers 149 ServeRAID Expansion Adapter 146, 157 ServeRAID M1015 154 ServeRAID M5014 153 ServeRAID M5015 153 ServeRAID M5025 162 ServeRAID Performance Key 152 ServeRAID-BR10i 153 ServerGuide 501 ServerGuide Scripting Toolkit 349 service processor 327 single SATA drive 148 slide kit 176 slot selection 321 storage 145 external 162 storage configurations 158 SUSE Linux 340 SUSE Linux support 175 system event logs 370 system status 331 target workload 123 thunk support 318 transceivers 168 Trusted Platform Module 170 Turbo Boost 130 UEFI 337 memory 310 memory scaling 338 modes 341 non-pooled memory 339 pooled memory 339 USB 170, 174 USB port 349 virtual media 348 virtualization 123, 174 virtualization model 5 VMware 355 VMware ESX 340 VMware ESXi 174, 340, 355, 358 VMware support 175 weight 119 width 119 Windows Server 2008 339 Windows Server 2008 R2 350, 356 Windows support 175 workload-optimized models 5, 125 Xeon processors 130 x3850 X5 55115, 219258 6Gb SSD Host Bus Adapter 99 adapter BIOS 248 adapters 107 architecture 66 ASU 259 attaching the MAX5 230 battery backup 97 bays 90 before you power on 220 block diagram 66
blue plastic carrier 100 boot from SAN 295 Boot Manager 242 boot sequence 245 booting 242 Broadcom controller 109 cable connections 220 cables 71, 233 cards 107 Change Boot Order 243 clearing CMOS memory 220 CMOS memory 220 compare with x3850 M2 8 comparison to M2 61 components 58 cooling zones 112 CPU considerations 221 current 109 database model 3 dedicated PCIe slot 100 default settings 220 depth 115 device drivers 293 diagnostics 227 dimensions 115 DIMM placement 23 disk 90 disk controllers 96 downloads 293 drive bays 90 DSA 227 DVD 102 Emulex 10Gb Ethernet Adapter 104 ESXi 114 Ethernet 109 event log 256 eXFlash 93 external storage 101 factory defaults 220 fans 112 fault tolerance 229 features 56 firmware levels 230 fixes 293 front view 57 Full Memory Test 227 heat 109 heat sinks 220 height 115 hemisphere mode 78 Hyper-Threading 74 I/O adapters 238 IMM 110, 250 configuring 251 dedicated port 251 defaults 250 event log 256 light path diagnostics 255 network 251 password 251
554
remote control 257 settings 250 shared port 252 system status 253 troubleshooting 253 userid 251 using 250 virtual light path diagnostics 255 virtual media 265 installation order processors 224 integrated virtualization 114 internal disks 90 internal view 58 legacy option ROMs 246 Legacy Thunk Support 241 light path diagnostics 111, 255 Linux 271, 291 log 256 management port 450 management processor 250 Matrox video 62 MAX5 68 block diagram 69 cables 71, 231 connecting to 71 connections 61 ESXi configuration 272 features 59, 69 internal view 61 LEDs 60 memory 60 memory scaling affinity 261 non-pooled mode 261 pooled mode 262 power supplies 60, 113 scalability 71 UEFI 261 MAX5 memory 79 MAX5, attaching 230 MegaRAID Web BIOS 94 memory 76 cards 76 diagnostics 227 DIMMs 78 DIMMs per channel 84 fault tolerance 229 hemisphere mode 78 installation order 79 matched pairs 78 MAX5 79, 82, 84 MAX5, attaching 230 mirroring 87 mirroring performance 89 mixed DIMMs 86 performance 8485 ranks 78 recommendations 226 redundant bit steering 79 sequence 79 sparing 89
speed 78 testing 226 memory performance 23 minimum processors 222 models 3, 64 models that cant scale 72 multi-burner 102 non-pooled mode 261 non-UEFI adapters 241 operating system support 114 optical drive 102 optical drive options 263 option ROMs 245 orange handles 57 OS install 263 OS instructions 288 out-of-band management 449 PCIe 1.1 compatibility 240 PCIe adapters 107, 238 PCIe slots 103 dedicated 100 selection 244 PciROOT 248 performance 84 pooled mode 262 power supplies 112, 249 processor considerations 221 processor installation 224 processor rules 75 processors 74 PXE 271 PXE Configuration 246 QPI cables 232 QPI Link Frequency 239 QPI Wrap Card 66 quick diagnostics 228 rack 115 RAID controllers 91 Red Hat install 291 Red Hat support 114 redundant bit steering 79 redundant power 249 remote control 257 removing a processor 225 resetting the server 220 riser cards 238 ROM execution order 248 SAN storage 294 SAS controllers 96 SAS support 91 SATA drives 92 scalabie complex 235 scalability 71 Serial over LAN 525 ServeRAID B5015 100 ServeRAID M1015 99 ServeRAID M5014 97 ServeRAID M5015 97, 100 ServeRAID M5025 98, 101 ServeRAID Perforamance Key 95
Index
555
ServeRAID support 91 ServeRAID-BR10i x3850 X5 96 ServerGuide 501 ServerGuide Scripting Toolkit 266 ServerProven 108 service processor 250 shared port 252 slot differences 244 slots 238 SSD controllers 96 SSD drives 93 SSD support 91 standard processors 74 storage 90 SUSE Linux 292 SUSE Linux support 114 system status 253 target audience 56 target workload 63 thunk support 241 TPM 111 transceivers 107 Turbo Boost 74 two-node 235 cabling 236 firmware 235 hemisphere mode 236 memory 236 powering on 238 processors 236 QPI cables 236 scaling 71 UEFI 261 UEFI 110, 259 ASU 259 MAX5 261 MAX5 Memory Scaling Affinity 261 non-pooled mode 261 performance tuning 262 pooled mode 262 settings 259 tuning 262 two-node 261 updates 293 USB ports 265 video 62 virtual media 265 virtualization key 114 virtualization model 3 VMware ESX install 271 VMware ESXi 114, 271 VMware support 114 voltage 109 VRMs 62 weight 115 width 115 Windows install 266 Windows Server 2008 289 Windows Server 2008 R2 289
Windows support 114 workload-optimized models 3, 64 write through 94 x3950 X5 See also x3850 X5 comapre with x3950 M2 8 database model 3 features 59 models 3, 64 virtualization model 3 Xeon See Intel
556
Back cover
Covers the IBM System x3950 X5, x3850 X5, x3690 X5, and the IBM BladeCenter HX5 Details technical information about each server and option Describes how to implement two-node configurations
High-end workloads drive ever-increasing and ever-changing constraints. In addition to requiring greater memory capacity, these workloads challenge you to do more with less and to find new ways to simplify deployment and ownership. And although higher system availability and comprehensive systems management have always been critical, they have become even more important in recent years. Difficult challenges, such as these, create new opportunities for innovation. The IBM eX5 portfolio delivers this innovation. This family of high-end computing introduces the fifth generation of IBM X-Architecture technology. The family includes the IBM System x3850 X5, x3690 X5, and the IBM BladeCenter HX5. These servers are the culmination of more than a decade of x86 innovation and firsts that have changed the expectations of the industry. With this latest generation, eX5 is again leading the way as the shift toward virtualization, platform management, and energy efficiency accelerates. This book is divided into two parts. In the first part, we provide detailed technical information about the servers in the eX5 portfolio. This information is most useful in designing, configuring, and planning to order a server solution. In the second part of the book, we provide detailed configuration and setup information to get your servers operational. We focus particularly on setting up MAX5 configurations of all three eX5 servers as well as 2-node configurations of the x3850 X5 and HX5. This book is aimed at clients, IBM Business Partners, and IBM employees that want to understand the features and capabilities of the IBM eX5 portfolio of servers and want to learn how to install and configure the servers for use in production.
SG24-7909-00
ISBN 0738435643