Planning, Installing, and Managing The IBM System x3950 M2
Planning, Installing, and Managing The IBM System x3950 M2
Planning, Installing, and Managing The IBM System x3950 M2
David Watts
Jens Reizel
Paul Tan
Kevin Galloway
ibm.com/redbooks
International Technical Support Organization
November 2008
SG24-7630-00
Note: Before using this information and the product it supports, read the information in
“Notices” on page ix.
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . x
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
The team that wrote this book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Become a published author . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi
Contents v
4.9.1 Problem with merging if prerequisites were met . . . . . . . . . . . . . . . 237
4.9.2 Problems with merging if prerequisites were not met . . . . . . . . . . . 239
4.9.3 Known problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 395
Contents vii
viii Planning, Installing, and Managing the IBM System x3950 M2
Notices
This information was developed for products and services offered in the U.S.A.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area.
Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM
product, program, or service may be used. Any functionally equivalent product, program, or service that
does not infringe any IBM intellectual property right may be used instead. However, it is the user's
responsibility to evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document.
The furnishing of this document does not give you any license to these patents. You can send license
inquiries, in writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A.
The following paragraph does not apply to the United Kingdom or any other country where such
provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION
PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR
IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT,
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer
of express or implied warranties in certain transactions, therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may
make improvements and/or changes in the product(s) and/or the program(s) described in this publication at
any time without notice.
Any references in this information to non-IBM Web sites are provided for convenience only and do not in any
manner serve as an endorsement of those Web sites. The materials at those Web sites are not part of the
materials for this IBM product and use of those Web sites is at your own risk.
IBM may use or distribute any of the information you supply in any way it believes appropriate without
incurring any obligation to you.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm
the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on
the capabilities of non-IBM products should be addressed to the suppliers of those products.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the
sample programs are written. These examples have not been thoroughly tested under all conditions. IBM,
therefore, cannot guarantee or imply reliability, serviceability, or function of these programs.
The following terms are trademarks of the International Business Machines Corporation in the United States,
other countries, or both:
Advanced Micro Devices, AMD, AMD-V, ATI, ES1000, Radeon, the AMD Arrow logo, and combinations
thereof, are trademarks of Advanced Micro Devices, Inc.
Cognos, and the Cognos logo are trademarks or registered trademarks of Cognos Incorporated, an IBM
Company, in the United States and/or other countries.
Snapshot, and the NetApp logo are trademarks or registered trademarks of NetApp, Inc. in the U.S. and
other countries.
Novell, SUSE, the Novell logo, and the N logo are registered trademarks of Novell, Inc. in the United States
and other countries.
Oracle, JD Edwards, PeopleSoft, Siebel, and TopLink are registered trademarks of Oracle Corporation
and/or its affiliates.
SAP, and SAP logos are trademarks or registered trademarks of SAP AG in Germany and in several other
countries.
Virtual SMP, VMotion, VMware, the VMware "boxes" logo and design are registered trademarks or
trademarks of VMware, Inc. in the United States and/or other jurisdictions.
Java, OpenSolaris, Solaris, Ultra, and all Java-based trademarks are trademarks of Sun Microsystems, Inc.
in the United States, other countries, or both.
BitLocker, Hyper-V, Microsoft, SQL Server, Windows NT, Windows Server, Windows, and the Windows logo
are trademarks of Microsoft Corporation in the United States, other countries, or both.
Intel Core, Intel SpeedStep, Intel Xeon, Intel, Itanium-based, Itanium, Intel logo, Intel Inside logo, and Intel
Centrino logo are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United
States, other countries, or both.
Linux is a trademark of Linus Torvalds in the United States, other countries, or both.
Other company, product, or service names may be trademarks or service marks of others.
Notices xi
xii Planning, Installing, and Managing the IBM System x3950 M2
Preface
The x3950 M2 server and x3850 M2 are the System x™ flagship servers and
implement the fourth generation of the IBM® X-Architecture®. They delivers
innovation with enhanced reliability and availability features to enable optimal
performance for databases, enterprise applications, and virtualized
environments.
The x3950 M2 and x3850 M2 features make the servers ideal for handling
complex, business-critical On Demand Business applications such as database
serving, business intelligence, transaction processing, enterprise resource
planning, collaboration applications, and server consolidation.
This IBM Redbooks® publication describes the technical details of the x3950 M2
scalable server and the x3850 M2 server. We explain what the configuration
options are, how 2-node, 3-node, and 4-node complexes are cabled and
implemented, how to install key server operating systems, and what
management tools are available to systems administrators.
Jens Reizel is a Support Specialist at IBM Germany and is responsible for the
post-sales technical support teams in the EMEA region. He has been working in
this function and with IBM for nine years. His areas of expertise include IBM
System x high end systems, management hardware, and Windows®, Linux®,
and VMware® operating systems.
Preface xv
Your efforts will help increase product acceptance and customer satisfaction. As
a bonus, you will develop a network of contacts in IBM development labs, and
increase your productivity and marketability.
Find out more about the residency program, browse the residency index, and
apply online at:
ibm.com/redbooks/residencies.html
Comments welcome
Your comments are important to us!
Although they have the same technical specifications and features, the x3850 M2
cannot be used to form a multinode unless you upgrade it to an IBM System
x3950 M2 by adding the ScaleXpander Option Kit, as described in section 1.2,
“Model numbers and scalable upgrade options” on page 9.
1 2
3 4
The front panel of the x3850 M2 and the x3950 M2, as shown in Figure 1-2,
provides easy access to a maximum of four hot-swap 2.5-inch SAS drives,
DVD-ROM, two USB 2.0 ports, an operator information panel with power on/off
button, and LEDs indicating information such as scalability, network activity, and
system errors and warnings.
The scalability LED on an x3950 M2 indicates whether the node (building block in
a scalable system) is participating in a multinode x3950 M2 complex. After each
node has successfully merged with the primary node in a partition, the scalability
LED is lit on all nodes in a partition of a multinode complex.
Power
supply 2
The rear panel of the x3850 M2 and x3950 M2, as shown in Figure 1-3, has:
PCI Express (PCIe) slots 1 to 7 (from left to right on the panel)
System serial port
Three scalability SMP expansion ports used for multinode x3950 M2
complexes
External SAS port
Three USB 2.0 ports
Integrated dual-port Broadcom Gigabit Ethernet RJ45 ports
Remote Supervisor Adapter II panel, which contains the servers video
connector port, 10/100 Mbps RJ45 out-of-band remote management port
(there is also a mini-USB port and a power adapter socket that is not used for
the x3850 M2/x3950 M2)
Two hot-swap redundant power supplies
Figure 1-4 On the hypervisor models of x3850 M2, a USB flash drive is pre-installed in
the internal USB socket and contains VMware ESXi 3.5 pre-loaded
Processors
Processor features include:
One 4U Rack-optimized sever with one of the following Intel processors:
– Xeon 7200 series (Tigerton) dual-core processors
– Xeon 7300 series (Tigerton) quad-core processors
– Xeon 7400 series (Dunnington) quad-core processors
– Xeon 7400 series (Dunnington) 6-core processors
Two processors standard, with support for up to four processors
One IBM eX4 “Hurricane 4” chipset with four 1066 MHz front-side buses
Support for Intel Virtualization Technology (Intel VT), Intel 64 technology
(EM64T), and Execute Disable Bit feature
Power
Two hot-swap redundant 1440 W power supplies are standard. At 220 V, one
power supply is redundant. At 110 V, the power supplies are non-redundant.
Figure 1-5 ScaleXpander chip (left); ScaleXpander chip installed on processor board
near the front panel of the x3950 M2 (right)
Unlike with the System x3950 and xSeries® 460, the x3950 M2 does not require
a special modular expansion enclosure. The multinode configuration is simply
formed by using another x3950 M2 or an x3850 M2 that has the ScaleXpander
Option Kit installed as described previously in 1.2.4, “Scalable upgrade option for
x3850 M2” on page 11
Multinode configurations
The x3950 M2 can form a multinode configuration by adding one or more
x3950 M2 servers. A number of configurations are possible as shown in
Figure 1-10.
Four nodes
8-way or 16-way
Three nodes (Each node is
2-way or 4-way)
6-way or 12-way Up to 1 TB RAM
(Each node is
Two nodes
2-way or 4-way)
4-way or 8-way Up to 768 GB RAM
x3950 M2
(Each node is
2-way or 4-way) x3950 M2
One node Up to 512 GB RAM x3950 M2
2-way or 4-way
Up to 256 GB RAM
x3950 M2 x3950 M2 x3950 M2
Note: At the time of writing, only Windows 2003 Enterprise and Datacenter
64-bit editions, RHEL 5 64-bit, and SLES 10 64-bit support this amount of
memory. See 2.6, “Operating system scalability” on page 66 for details.
Partitioning
Partitioning is the concept of logically splitting a multinode complex into separate
systems. You can then install an operating system on a partition and have it run
independently from all other partitions. The advantage of partitioning is that you
can create and delete partitions without having to recable the complex. The only
requirement is that partitions be formed on node boundaries.
The interface where you set up and maintain partitions is an extension of the
Remote Supervisor Adapter II Web interface. It is used to create, delete, control,
and view scalable partitions.
Table 1-2 on page 11 shows the system models for both 32-bit and 64-bit
versions of the operating system.
With the IBM Datacenter Unlimited Virtualization option, the x3950 M2 models
come with two processors, 8 GB of memory (eight 1 GB DIMMs), four memory
cards, and no disks. The system is shipped with the Datacenter installation CD,
OS documentation, recovery CD, and a 4-socket Certificate of Authenticity (COA)
to license the system. Windows Server® 2003 R2 Datacenter Edition is not
preloaded. This offering is available in both English and Japanese languages.
Note: IBM no longer offers a Software Update Subscription for this offering.
Customers should purchase a Microsoft Software Assurance contract for
operating system maintenance and upgrades. For information see:
http://www.microsoft.com/licensing/sa
This end-to-end offering provides a fully configured and certified solution for
customers who want to maintain a tightly controlled environment for maximum
IBM Datacenter UVHA solution offerings are tested and certified on specific
System x servers and with standard ServerProven options, storage systems, and
applications. All components must be both ServerProven and Microsoft
cluster-logo certified.
Table 1-2 on page 11 shows the models for both 32-bit and 64-bit versions of the
operating system.
With this option, the x3950 M2 models come with two processors, 8 GB memory
(eight 1 GB DIMMs), four memory cards, and no disks. Unlike previous
high-availability offerings, Windows Server 2003 R2 Datacenter Edition is not
preloaded. Also shipped with the system are a recovery CD, OS documentation,
and a 4-socket Certificate of Authenticity (COA) to license the system. This
offering is available in both English and Japanese languages.
Note: IBM no longer offers a Software Update Subscription for this offering.
Customers should purchase a Microsoft Software Assurance contract for
operating system maintenance and upgrades. For information see:
http://www.microsoft.com/licensing/sa
IBM Datacenter preload upgrades can be ordered only after receiving approval
from the IBM world-wide System x marketing team. IBM Sales Representatives
should notify their geography Marketing Product Manager and Sales Managers
of these opportunities, and the Product and Sales Managers should, in turn,
notify the World Wide Marketing Product Manager of the sales opportunity.
Business Partners should notify their IBM Sales Representative, who should
engage with geography Product Marketing and Sales Managers.
Table 1-3 Upgrade options for the IBM Datacenter Unlimited Virtualization offering
Upgrade kits Order number
Windows Server 2003 Datacenter Edition R2, 32-bit, 1-4 CPUs 4818-NCU
Windows Server 2003 Datacenter Edition R2, 32-bit, 1-4 CPUs 4818-NCJ
(Japanese)
Note: These upgrade order numbers can only be ordered from the IBM World
Wide System x Brand and might not appear in IBM standard configuration
tools.
Windows Server 2003 Datacenter Edition R2, 32-bit, 1-4 CPUs 4816-NCU
Windows Server 2003 Datacenter Edition R2, 32-bit, 1-4 CPUs 4816-NCJ
(Japanese)
Note: These upgrade order numbers can only be ordered from the IBM World
Wide System x Brand and might not appear in IBM standard configuration
tools.
Like the ESX Server, VMware ESXi supports the entire VMware Infrastructure 3
suite of products, including VMFS, Virtual SMP®, VirtualCenter, VMotion®,
VMware Distributed Resource Scheduler, VMware High Availability, VMware
Update Manager, and VMware Consolidated Backup.
Figure 1-11 on page 21 shows a components diagram of the overall ESXi 3.5
architecture.
1
Available from http://www.vmware.com/files/pdf/ESXServer3i_architecture.pdf. This section
contains material from VMware. Used with permission.
Third-party
hostd DCUI syslog VMX VMX VMX
CIM plug-ins
VMM VMM VMM
User world API
Figure 1-11 The architecture of VMware ESXi eliminates the need for a service console
With an IBM eX4 server running VMware ESXi (or VMware ESX), applications
and services can be deployed in highly reliable and secure virtual machines.
Virtual machines can be provisioned, consolidated, and managed centrally
without having to install an operating system, thus simplifying the IT
infrastructure and driving down total cost of ownership for businesses with
constrained IT budgets and resources.
2
See Tips and Techniques for Implementing Infrastructure Services on ESX Server, available at:
http://www.vmware.com/vmtn/resources/409. Reproduced by permission.
Table 1-5 shows that scheduling opportunities scale exponentially rather than
linearly when more cores are available.
Table 1-5 Scheduling opportunities scale exponentially when there are more cores
ESX Host Server Number of Cores Scheduling
opportunities
(VM = 2 vCPUs)
The System x3850 M2 and x3950 M2 servers are designed for balanced system
performance, and are therefore uniquely positioned to take advantage of the
larger workloads now available to be virtualized.
Table 1-7 shows the limitations of each VMware distribution that is supported on
the x3850 M2 and x3950 M2 (single node).
Note: The values in the table are correct at the time of writing and may change
as testing completes. The values do not reflect the theoretical values but set
the upper limit of support for either distribution.
For details of part numbers, refer to VMware Offerings in the IBM System x
Configuration and Options Guide:
http://www.ibm.com/systems/xbc/cog/vmwareesx.html
The exact description of the parts above might differ slightly from country to
country, or by the length (in years) of subscription and support. License
upgrades, subscription upgrades, and support must be purchased as a complete
set to upgrade ESXi Edition to Virtual Infrastructure Foundation, Standard, and
Enterprise Editions.
This chipset is designed for the Xeon MP processor family from Intel. The IBM
eX4 chipset provides enhanced functionality and capability with significant
improvements in scalability, decreased memory latency, increased memory
bandwidth and increased I/O bandwidth. The architecture consists of the
following components:
One to four Xeon dual-core, quad-core, or 6-core processors
Hurricane 4 Memory and I/O Controller (MIOC)
Eight high-speed memory buffers
Two PCI Express bridges
One Enterprise Southbridge Interface
Figure 1-12 on page 28 shows a block diagram of the x3850 M2 and x3950 M2.
DDR2 Buffer
PCI-E bridge 1 PCI-E bridge 2
2 GBps
DDR2 Buffer
South bridge
1 2 3 4 5 6 7
PCI +
Serial USB LSI
1078
RSA2 + SAS
DVD drive Seven PCI Express x8 slots
Video
B = bytes (slots 6 & 7 are hot-swap)
b = bits 6x USB 2.0
MR10k
PCI-E x4
External SAS port
Gb Ethernet
BCM5709C HDD backplane
Each memory port out of the memory controller has a peak read throughput of
4.26 GBps and a peak write throughput of 2.13 GBps. DIMMs are installed in
matched pairs, two-way interleaving, to ensure the memory port is fully utilized.
Peak throughput for each PC2-5300 DDR2 DIMM is 4.26 GBps.
There are eight memory ports; spreading the installed DIMMs across all ports
can improve performance. The eight independent memory ports provide
simultaneous access to memory. With four memory cards installed, and eight
DIMMs in each card, peak read memory bandwidth is 34.1 GBps and peak write
bandwidth is 17.1 GBps. The memory controller routes all traffic from the eight
memory ports, four microprocessor ports, and the three PCIe bridge ports.
The memory controller also has an embedded DRAM that, in the x3850 M2 and
x3950 M2, holds a snoop filter lookup table. This filter ensures that snoop
The three scalability ports are each connected to the memory controller through
individual scalability links with a maximum theoretical bidirectional data rate of
10.24 GBps per port.
IBM eX4 has two PCIe bridges and each are connected to a HSS-IB port of the
memory controller with a maximum theoretical bidirectional data rate of 6 GBps.
As shown in Figure 1-12 on page 28, PCIe bridge 1 supplies four of the seven
PCI Express x8 slots on four independent PCI Express buses. PCIe bridge 2
supplies the other three PCI Express x8 slots plus the onboard SAS devices,
including the optional ServeRAID-MR10k and a 4x external onboard SAS port.
1.6.1 Hurricane 4
Hurricane 4 is the north bridge component of the IBM eX4 chipset designed for
latest Intel Core™ Architecture-based processors which feature a new
architecture for the processor front-side bus. Hurricane 4 supports the
processors in the Xeon 7000 family of processors, including those with code
names of Tigerton and Dunnington.
Note: The amount of memory that BIOS reports is minus the portion used for
the XceL4v cache.
Each PCIe chip provides four separate PCIe x8 buses to support four PCIe x8
slots. PCIe Bridge 1 supports slots 1-4 of the PCIe x8 slots and PCIe Bridge 2
supports slots 5-7 of the PCIe x8 slots and a dedicated PCIe x8 slot for
ServeRAID MR10k SAS/SATA RAID controller.
1.6.5 Ranks
A rank is a set of DRAM chips on a DIMM that provides eight bytes (64 bits) of
data.
The capacity of each DRAM is a number of words where each word can be 4 bits
(x4), 8 bits (x8) and, starting to become prevalent, 16 bits in length (x16).
The word length is usually written as x4 for 4 bits, and so on. The number of
words in the DRAM is sometimes written on the label of the DIMM, such as a
DRAM chip on a DIMM.
The DRAM devices that make up a rank are often, but not always, mounted on
one side of the DIMM, so a single-rank DIMMs can also be referred to as a
single-sided DIMM. Likewise a double-ranked DIMM can be referred to as a
double-sided DIMM.
DDR2 SMI2
6 GBps 6 GBps
6 GBps
Calgary PCI-X bridge PCI-X bridge
RSA SL
1 2 3 4 5 6
USB 2.0 ServeRAID
IBM eX4 technology builds and improves upon its previous generation X3
technology. The key enhancements are:
Processor interface
– Quad 1066 MHz front-side bus (FSB), which has a total bandwidth of up to
34.1 GBps. In X3, the maximum bandwidth was 10.66 GBps.
– The front-side bus is increased to 1066 MHz from 667 MHz for 3.2x
bandwidth improvement.
– Snoop filter is for quad FSB coherency tracking compared to X3 with only
dual FSB coherency tracking.
Memory
– Increased (four-fold) memory capacity (2X from chipset, 2X from DRAM
technology) compared to X3.
1.7 Processors
As mentioned previously, the x3850 M2 and x3950 M2 models use one of the
following Intel Xeon Processor models:
Refer to 1.2, “Model numbers and scalable upgrade options” on page 9 for details
about the current models.
Tip: For the purposes of VMware VMotion, the Dunnington processors are
compatible with the Tigerton processors.
As described in 1.3, “Multinode capabilities” on page 14, you can also connect
multiple x3950 M2s to form larger single-image configurations.
The processors are accessible from the top of the server after opening the media
hood. The media hood is hinged at the middle of the system and contains the
SAS drives, optical media, USB ports and light path diagnostic panel.
Figure 1-14 shows the media hood half-way open.
Figure 1-14 The x3950 M2 with the media hood partially open
The processors are each packaged in the 604-pin Flip Chip Micro Pin Grid Array
(FC-mPGA) package. It is inserted into surface-mount mPGA604 socket. The
processors use a large heat-sink to meet thermal specifications.
Figure 1-15 compares the layout of the Tigerton dual-core and quad-core
processors.
L1 Instruct
L1 Cache
Instruct Processor
Core L1 Data
Cache Cache
Processor L2 L2
Core Cache L1 Instruct Cache
L1 Cache
Data Processor
Cache Core L1 Data
Cache
L1 Instruct
L1 Cache
Instruct Processor
Core L1 Data
Cache Cache
Processor L2 L2
Core Cache L1 Instruct Cache
L1 Cache
Data Processor
Core L1 Data
Cache Cache
The Xeon E7400 Series Dunnington processors, both 4-core and 6-core models,
have shared L2 cache between each pair of cores but also have a shared L3
cache across all cores of the processor. While technically all Dunnington Xeon
processors have 16MB of L3 cache, 4-core models only have 12MB of L3 cache
enabled and available. See Figure 1-16 on page 36.
Core
Core
Core
Core
Core
Core
L2 cache L2 cache L2 cache
L3 cache
Key features of the processors used in the x3850 M2 and x3950 M2 include:
Multi-core processors
The Tigerton dual-core processors are a concept similar to a two-way SMP
system except that the two processors, or cores, are integrated into one
silicon die. This brings the benefits of two-way SMP with lower software
licensing costs for application software that licensed per CPU socket plus the
additional benefit of less processor power consumption and faster data
throughput between the two cores. To keep power consumption down, the
resulting core frequency is lower, but the additional processing capacity
means an overall gain in performance.
The Tigerton quad-core processors add two more cores onto the same die,
and some Dunnington processors also add two more. Hyper-Threading
Technology is not supported.
Each core has separate L1 instruction and data caches, and separate
execution units (integer, floating point, and so on), registers, issue ports, and
pipelines for each core. A multi-core processor achieves more parallelism
than Hyper-Threading Technology because these resources are not shared
between the two cores.
With two times, four times, and even six times the number of cores for the
same number of sockets, it is even more important that the memory
subsystem is able to meet the demand for data throughput. The 34.1 GBps
peak throughput of the x3850 M2 and x3950 M2’s eX4 technology with four
memory cards is well suited to dual-core and quad-core processors.
A quad-pumped 266 MHz bus therefore results in a 1066 MHz front-side bus.
The bus is 8 bytes wide, which means it has an effective burst throughput of
8.53 GBps. This can have a substantial impact, especially on TCP/IP-based
LAN traffic.
Intel 64 Technology (formerly known as EM64T)
Intel 64 Technology is a 64-bit extension to the industry-standard IA32 32-bit
architecture. Intel 64 Technology adds:
– A set of new 64-bit general purpose registers (GPRs)
– 64-bit instruction pointers
– The ability to process data in 64-bit chunks
Although the names of these extensions suggest that the improvements are
simply in memory addressability, Intel 64 Technology is, in fact, a fully
functional 64-bit processor.
The processors in the x3850 M2 and x3950 M2 include the Intel 64
Technology extensions from Intel. This technology is compatible with IA-32
software while enabling new software to access a larger memory address
space.
To realize the full benefit of this technology, you must have a 64-bit operating
system and 64-bit applications that have been recompiled to take full
advantage of this architecture. Existing 32-bit applications running on a 64-bit
operating system can also benefit from EM64T.
The Tigerton processors limit memory addressability to 40 bits of addressing.
Intel 64 Technology provides three distinct operation modes:
– 32-bit legacy mode
The first and, in the near future, probably most widely used mode will be
the 32-bit legacy mode. In this mode, processors with Intel 64 Technology
Memory DIMMs are installed in the x3850 M2 and x3950 M2 using memory
cards, each card has eight DIMM sockets. The server supports up to four
memory cards, giving a total of up to 32 DIMM sockets.
Some models have two memory cards, others have all four cards as standard.
Using 8 GB DIMMs in every socket, the server can hold 256 GB of RAM. With
four nodes, the combined complex can hold up to 1 TB RAM.
The memory is two-way interleaved, meaning that memory DIMMs are installed
in pairs. Figure 1-12 on page 28 shows eight ports from the Hurricane 4 memory
controller to memory, with each supporting up to 4.26 GBps read data transfers
and 2.13 GBps write data transfers.
The DIMMs operate at 533 MHz, to be in sync with a front-side bus. However the
DIMMs are 667 MHz PC2-5300 spec parts because they have better timing
parameters than the 533 MHz equivalent. The memory throughput is 4.26 GBps,
or 533 MHz x 8 bytes per memory port for a total of 34.1 GBps with four memory
cards.
Memory scrubbing
Memory scrubbing is an automatic daily test of all the system memory that
detects and reports memory errors that might be developing before they
cause a server outage.
Memory scrubbing and Memory ProteXion work in conjunction and do not
require memory mirroring to be enabled to work properly.
When a bit error is detected, memory scrubbing determines whether the error
is recoverable:
– If the error is recoverable, Memory ProteXion is enabled, and the data that
was stored in the damaged locations is rewritten to a new location. The
error is then reported so that preventative maintenance can be performed.
If the number of good locations is sufficient to allow the proper operation of
the server, no further action is taken other than recording the error in the
error logs.
– If the error is not recoverable, memory scrubbing sends an error message
to the light path diagnostics, which then turns on the proper lights and
LEDs to guide you to the damaged DIMM. If memory mirroring is enabled,
Chipkill memory
Chipkill is integrated into the XA-64e chipset, so it does not require special
Chipkill DIMMs and is transparent to the operating system. When combining
Chipkill with Memory ProteXion and Active Memory, the x3850 M2 and x3950
M2 provides very high reliability in the memory subsystem.
When a memory chip failure occurs, Memory ProteXion transparently handles
the rerouting of data around the failed component as previously described.
However, if a further failure occurs, the Chipkill component in the memory
controller reroutes data. The memory controller provides memory protection
similar in concept to disk array striping with parity, writing the memory bits
across multiple memory chips on the DIMM. The controller is able to
reconstruct the missing bit from the failed chip and continue working as usual.
One of these additional failures can be handled for each memory port for a
total of eight Chipkill recoveries.
Hot-add and hot-swap memory
The x3850 M2 and x3950 M2 support the replacing of failed DIMMs while the
server is still running. This hot-swap support works in conjunction with
memory mirroring. The server also supports adding additional memory while
the server is running. Adding memory requires operating system support.
These two features are mutually exclusive. Hot-add requires that memory
mirroring be disabled, and hot-swap requires that memory mirroring be
enabled. For more information, see 3.2, “Memory subsystem” on page 111.
The system takes the following sequence of steps regarding memory failure
detection and recovery:
1. The first failure detected by the Chipkill algorithm on each port does not
generate a light path diagnostics error because Memory ProteXion recovers
from the problem automatically.
2. Each memory port can sustain a second chip failure without shutting down.
3. Provided that memory mirroring is enabled, the third chip failure on that port
sends the alert and takes the DIMM offline, but keeps the system running out
of the redundant memory bank.
SAS is the logical evolution of SCSI. SAS uses much smaller interconnects than
SCSI, while offering SCSI compatibility, reliability, performance, and
manageability. In addition, SAS offers longer cabling distances, smaller form
factors, and greater addressability.
The x3850 M2 and x3950 M2 has an external SAS x4 port used in conjunction
with the optional ServeRAID MR10k. This external port supports SAS non-RAID
disk enclosures such as the EXP3000. This port has an SFF-8088 connector.
For more information about the onboard SAS controller and the ServeRAID
MR10k daughter card, see Figure 3-22 on page 131 in section 3.3.3,
“ServeRAID-MR10k RAID controller” on page 128.
Figure 1-18 System planar layout showing the seven PCI Express slots
The optional ServeRAID MR10k adapter does not use a PCIe x8 slot because it
has a dedicated PCIe x8 customized 240-pin slot on the I/O board.
1.11 Networking
The IBM x3950 M2 and x3850 M2 servers have an integrated dual 10/100/1000
Ethernet controller that uses the Broadcom NetXtreme II BCM5709C controller.
The controller contains two standard IEEE 802.3 Ethernet MACs which can
operate in either full-duplex or half-duplex mode.
Note: These onboard Ethernet controllers do not support iSCSI nor RDMA.
Ethernet 2
The layout of the x3850 M2 and x3950 M2, showing the location of the memory
cards, power supplies, and fans is displayed in Figure 1-20.
Two
hot-swap
redundant
1440 W
power
supplies
Light path diagnostics can monitor and report on the health of microprocessors,
main memory, hard disk drives, PCI adapters, fans, power supplies, VRMs, and
the internal system temperature. See 6.1, “BMC configuration options” on
page 300 for more information.
NMI button
1 2 (trained service
technician only)
REMIND
Power-control Locator button/ OVER SPEC LOG LINK PS PCI SP
button cover locator LED
FAN TEMP MEM NMI
Ethernet port
activity LEDs CNFG CPU VRM DASD RAID BRD
1 2
3 4
For more information about enabling and configuring these management utilities,
see the x3850 M2 and x3950 M2 User’s Guide:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073029
The video adapter on this RSA II card is an ATI Radeon RN50 (ES1000) SVGA
video controller. A DB-15 video connector (shown in Figure 1-3 on page 4) is
provided on the Remote Supervisor Adapter II. The RSA II provides up to
1024x768 resolution, with a color depth of maximum of 32 bits at 85 Hz
maximum refresh rate, with 16 MB of video memory.
System
management
connector
10/100 Mbps
Video Adapter Ethernet port
System management
daughter card Power LED Reset Button
The most useful functions and features of the Remote Supervisor Adapter II
include:
IBM ASIC with integrated PowerPC® 405 core executing at 200 MHz
Automatic notification and alerts
The RSA II automatically sends different types of alerts and notifications to
another server such as IBM Director and SNMP destination, or it sends e-mail
directly to a user by using SMTP.
Continuous health monitoring and control
The RSA II continuously monitors all important system parameters such as
temperature, voltage, and so on. If a fan fails, for example, the RSA II forces
the remaining fans to increase speed to compensate for the failing fan.
Event log
You can access the server event logs and the power-on-self-test (POST) log,
and export them while the server is running.
Operating system failure window capture
When the operating system hangs, for example, with a blue screen, you might
want to capture the window for support purposes. Additionally, the RSA II
stores the last failure window in memory so you can refer to it later.
See 6.2, “Remote Supervisor Adapter II” on page 316 for more information about
the service processor.
The application software enables you to track actual power consumption trends
and corresponding thermal loading of servers running in your environment with
your applications.
Active Energy Manager enables customers to monitor actual power draw and
thermal loading information. This helps you with:
More efficient planning of new datacenter construction or modification
Proper power input sizing based on physical systems
Justification of incremental hardware purchases based on available input
power capacity
Better utilization of existing resources
The TPM is TCG V1.2-compliant and is ready for use with software purchased
from the third-party list of the TPM Ecosystem partners who are also in
compliance with the TPM V1.2 specification.
The IBM System x3850 M2 and x3950 M2 servers deliver additional processing,
expandability, and high-availability features over those of their predecessors, the
IBM System x3850 and x3950 servers. They are ideal for handling complex,
business-critical On Demand Business applications that must be supported by
space-saving, rack-optimized servers.
The IBM eX4 servers are part of the IBM System x scale-up server offering,
which is designed to provide: the highest level of processor scalability with
support for up to 16 multi-core processors; up to 1 TB of addressable memory
with higher levels of memory availability; and flexible I/O expansion with support
for up to 28 PCIe adapters.
Figure 2-1 on page 57 provides an overview of the scale-up and scale-out IBM
System x product portfolio including x3850 M2 and x3950 M2.
DS4800
High
x3850 M2 density
Storage
Servers DS3200/3300/3400
S3200/3300/34
320 00 0000 x3455 BladeCenter E
Scale
BladeCenter H
x3650
BladeCenter S
x3500
0 x3550
3550
X3400
X34
X x3250
x3200 M2
x3100
Scale out / distributed computing
tended Design Architecture¥
Figure 2-1 The IBM x3950 M2 and x3850 M2 are part of the high-end scale-up portfolio
Server
iDataPlex Consolidation,
Virtualization
BladeCenter,
System x Rack
Infrastructure
Scale Up
Web 2.0,
Simplification,
HPC,
Application
Grid
Serving
Scale Out
Figure 2-2 IBM eX4 compared to IBM BladeCenter and System x Rack and iDataPlex
IBM iDataPlex
iDataPlex is massive scale-out solution that is deployed in customized rack units.
It is designed for applications where workloads can be divided and spread across
a very large pool of servers that are configured identically from the application
workload perspective. Web 2.0, High Performance Clusters, and Grid Computing
are some of the targeted applications for IBM iDataPlex in which the applications
are stateless and use software for workload allocation across all nodes.
For more information about iDataPlex, refer to the paper Building an Efficient
Data Center with IBM iDataPlex, REDP-4418 available from:
http://www.redbooks.ibm.com/abstracts/redp4418.html
IBM BladeCenter
IBM BladeCenter products are designed for complete infrastructure integration,
ease of management, energy efficient servers, hardware and software Reliability,
Availability, and Serviceability (RAS), and network virtualization through Open
Fabric Manager. Figure 2-3 on page 59 shows the evolution of BladeCenter.
For more information about IBM BladeCenter, refer to the IBM Redbooks
publication, IBM BladeCenter Products and Technology, SG24-7523:
http://www.redbooks.ibm.com/abstracts/sg247523.html
The Hurricane 4 chipset also adds improved bandwidth for its three scalability
ports and has increased memory throughput with eight high speed memory
buffer chips. Furthermore, the x3850 M2 supports more I/O slots from previously
having six PCI-X (Tulsa based x3850 has four PCIe and two PCI-X slots) to
seven PCIexpress slots.
The onboard LSI 1078 RAID controller and the optional ServeRAID MR10k
installed in a dedicated PCIe x8 slot have significantly improved storage
subsystem bandwidth compared to the x3850’s Adaptec ServeRAID 8i RAID
controller which shared a slower common PCI-X 66 MHz bus to the Southbridge
with the onboard Broadcom Gigabit Ethernet controllers. The Hurricane 4 has a
dedicated Enterprise Southbridge Interface (ESI) for the dual port onboard PCIe
x4 Broadcom 5709C controllers, RSA II, Video, USB 2.0 and Serial interfaces.
The x3850 M2 also has an onboard 4x SAS port which can be used in
conjunction with the ServeRAID MR10k for additional disk drive expansion (for
example, using one or more EXP3000 storage enclosures) not previously
possible with the x3850, without the use of one PCIe slot for a MegaRAID 8480
SAS RAID controller.
Table 2-1 compares major differences between x3850 M2 and the x3850.
Front-side bus Two 667 MHz (two Four 1066 MHz (one
processors on each bus) processor on each bus)
PCI Express slots Some models have four Seven PCI Express x8
PCI Express x8 full-length half-length slots
slots
The Hurricane 4 chipset also adds improved bandwidth for its three scalability
ports and has increased memory throughput with eight high speed memory
buffer chips. Furthermore, the x3950 M2 has support for more I/O slots from
previously having six PCI-X slots to seven PCIe slots.
The onboard LSI 1078 RAID controller and the optional ServeRAID MR10k
installed in a dedicated PCIe x8 slot have significantly improved storage
subsystem bandwidth compared to the x3950’s Adaptec ServeRAID 8i RAID
controller which shared a slower common PCI-X 66 MHz bus to the Southbridge
with the onboard Broadcom Gigabit Ethernet controllers. The Hurricane 4 has a
dedicated Enterprise Southbridge Interface (ESI) for the dual port onboard PCIe
x4 Broadcom 5709C controllers, RSA II, Video, USB 2.0 and Serial interfaces.
The x3950 M2 also has an onboard 4x SAS port which can be used in
conjunction with the ServeRAID MR10k for additional disk drive expansion (for
example, using one or more EXP3000 storage enclosures) not previously
possible with the x3950.
Table 2-2 on page 63 compares the major differences between the x3950 M2
and the x3950.
Front-side bus Two 667 MHz (two Four 1066 MHz (one
processors on each bus) processor on each bus)
PCI Express slots per node Some models have four Seven PCI Express x8
PCI Express x8 full-length half-length slots
slots
For example, not all applications can use the full power of four processors in one
server. File and print servers often only take advantage of one or two processors
and popular mail systems typically only scale well to four processors. Table 2-3
on page 65 shows the suitability of multi-processor systems to application types.
Processors are only one part of the scalability story. Typically, an important task
is to examine the following items for scalability: memory, disk, and networking
subsystems. Normally, the performance gains from adding processors are only
realized when memory is added in parallel. For disk-intensive applications, such
as OLTP-type applications, it is essential to have a large disk array to stream data
to the CPU and memory subsystems so that any disk-related delays are kept to a
minimum.
Table 2-4 shows how application types scale and what is required to achieve
peak performance. This table lists the server configurations used to produce
several recent benchmark results. As you can see, the amount of memory and
disks varies widely depending on the application.
16 16 4 256 GB 775
16 16 4 128 GB 384
SPEC CPU2006 16 16 4 64 GB 1
The different server configurations reflect the different workloads of the these
benchmarks. The workload that the benchmark generates causes the server to
bottleneck in a particular subsystem.
As the table indicates, the SPEC CPU2006 Benchmark also highlights the
component-focused nature of the SPEC benchmarks and the CPU-intensive
applications they serve. This 8-way dual-core server required only 64 GB of
memory and one disk. Clearly, the workload isolates the CPU with very little
dependency on other subsystems. This means that the benchmark might be very
good for comparing raw CPU performance, but it provides limited information
regarding the performance of the entire system. The CPUs in a system can be
very fast, but performance remains poor if the memory or I/O subsystems cannot
supply data to them quickly enough.
In the single chassis 4-way configuration, the IBM eX4 server acts as an industry
standard symmetric multiprocessor (SMP) system. Each processor has equal
access to all system resources. In most industry standard SMP systems, scaling
beyond 4-way configurations has inherent processor, memory, and I/O
subsystem contention issues. These issues can limit the ability of the system to
scale linearly with the increased number of processors, memory, and I/O
resources greater than 4-way SMP systems.
The term NUMA is not completely correct because memory and I/O resources
can be accessed in a non-uniform manner. PCIe and USB devices may be
associated with nodes. The exceptions to this situation are existing I/O devices,
such as DVD-ROM drives, which are disabled because the classic PC
architecture precludes multiple copies of these existing items.
The key to this type of memory configuration is to limit the number of processors
that directly access a piece of memory, thereby improving performance because
of the much shorter queue of requests. The objective of the operating system is
to ensure that memory requests be fulfilled by local memory when possible.
These modern operating systems attempt to allocate resources that are local to
the processors being used by each process. So, when a process and its threads
start on node 1, all execution and memory access will be local to node 1. As
more processes are added to the system, the operating system balances them
across the nodes. In this case, most memory accesses are evenly distributed
across the multiple memory controllers, reducing remote access, greatly
reducing queuing delays, and improving performance.
Home nodes
VMware ESX assigns each VM a home node when the VM begins running. A VM
only runs on processors within its home node. Newly-allocated memory comes
from the home node also. Thus, if a VM’s home node does not change, the VM
uses only local memory, avoiding the performance penalties associated with
remote memory accesses to other NUMA nodes. New VMs are assigned to
home nodes in a round-robin fashion. The first VM goes to the first node, the
second VM to the second node, and so on. This policy ensures that memory is
evenly used throughout all nodes of the system.
The rebalancer selects an appropriate VM and changes its home node to the
least-loaded node. When possible, the rebalancer attempts to move a VM that
already has some memory located on the destination node. From that point on,
the VM allocates memory on its new home node, unless it is moved again. It only
runs on processors within the new home node.
Transparent page sharing for VMware ESX systems has also been optimized for
use on NUMA systems like the IBM x3950 M2. With VMware ESX running on a
multinode IBM x3950 M2 partition, pages are shared per node, so each NUMA
node has its own local copy of heavily shared pages. When virtual machines use
shared pages, they do not have to access remote memory.
Figure 2-4 Setting CPU or memory affinity on VMware ESX 3.0.x / 3.5.x in VI Client or Virtualcenter
If both options are set before a virtual machine starts, the virtual machine runs
only on the selected node and all of its memory is allocated locally.
An administrator can also manually move a virtual machine to another node after
the virtual machine has started running. In this case, the page migration rate of
the virtual machine should also be set manually, so that memory from the virtual
machine’s previous node can be moved to its new node.
Note: VMware ESX 3.5 currently supports no more than 256 GB of RAM
installed.
Note: At the time of the writing, VMware ESX 3.5 Update 1 was not supported
by IBM for multinode x3950 M2 with up to 64 processor cores. Although not
currently supported, VMware ESX 3.5 Update 1 support from IBM is planned
for 2-node x3950 M2 (32-cores) in 2H/2008. VMware ESXi is not supported on
multinode x3950 M2 complexes.
Windows Server 2003 Enterprise and Datacenter Editions are NUMA-aware and
are able to assign application threads to use processor and memory resource
pools on local NUMA nodes. Scheduling application threads to run on local
resource pools can improve application performance because it minimizes
internode traffic on the x3950 M2 scalability ports and also reduces contention
for resources with other applications and potential resources bottlenecks by
assigning the different applications to run on different NUMA nodes.
Table 2-5 on page 74 lists features of the various Microsoft Server 2003 editions.
1
At the time of the writing, Windows Server 2003 was able to detect and use only up to 64 cores.
Edition availability
Scalability
The following documents have more details about other features available on
each edition of Windows Server 2003:
Comparison of Windows Server 2003 editions
http://technet2.microsoft.com/windowsserver/en/library/81999f39-41e9
-4388-8d7d-7430ec4cc4221033.mspx?mfr=true
Virtual memory address space limits for Windows editions
http://msdn.microsoft.com/en-us/library/aa366778(VS.85).aspx#memory_
limits
Table 2-6 lists features supported by the various Windows Server 2008 editions.
Edition availability
Scalability
Memory — 32-bit 4 GB 64 GB 64 GB 4 GB
The following Web pages and documents have more details about other features
available on each edition of Windows Server 2008:
Compare Technical Features and Specifications
http://www.microsoft.com/windowsserver2008/en/us/compare-specs.aspx
Virtual memory address space limits for Windows limits
http://msdn.microsoft.com/en-us/library/aa366778(VS.85).aspx#memory_
limits
Physical memory address space limits
http://msdn.microsoft.com/en-us/library/aa366778(VS.85).aspx#physica
l_memory_limits_windows_server_2008
Inside Windows Server 2008 Kernel Changes, by Mark Russinovich
http://technet.microsoft.com/en-us/magazine/cc194386.aspx
The documents describe features of the Linux 2.6 kernel and components such
as the Linux task scheduler and memory allocator, which affect the scaling of the
Linux operating system on the IBM x3850 M2 and x3950 M2.
The Linux operating system determines where processor cores and memory are
located in the multinode complex from the ACPI System Resource Affinity Table
(SRAT) and System Locality Information Table (SLIT) provided by firmware. The
SRAT table associates each core and each contiguous memory block with the
node they are installed in. The connections between the nodes and the number
of hops between them is described by the SLIT table.
In general, memory is allocated from the memory pool closest to the core on
which the process is running. Some system-wide data structures are allocated
evenly from all nodes in the complex to spread the load across the entire
complex and to ensure that node 0 does not run out of resources, because most
boot-time code is run from that node.
Server support limits as defined by Red Hat Enterprise Linux Product Subscription
Maximum Logical Processors 32 (up to 128 with bigsmp 32 (up to 128 on certified systems)
kernel on certified systems)
For more information, review the SUSE Linux Enterprise Server 10 Tech Specs &
System Requirements at:
http://www.novell.com/products/server/techspecs.html
Three-tier
Clients
Two-tier
Clients
Three-in-one
Clients
Several editions of SQL Server 2005 have varying support for x64; however, only
these versions are suitable for creating a consolidated SQL Server environment:
SQL Server 2005 Enterprise Edition (32-bit and 64-bit)
SQL Server 2005 Standard Edition (32-bit and 64-bit)
For medium and large-scale SQL Server consolidation projects, the Standard
Edition and Enterprise Edition versions both have native x64 versions; however,
many of the advanced scalability features are only found in the Enterprise
Edition. Developer Edition has all the features of Enterprise Edition, but is
licensed only for development and testing, not for production use.
It is important to note that a database created using SQL Server 2005 Express
Edition can be moved to an installation of SQL Server 2005 Enterprise Edition
without any modifications. This provides a clear growth strategy for all new
databases created with SQL Server 2005 and demonstrates the ease with which
databases can be scaled-up on this platform.
Hot-add memory
Additional physical memory can be installed in a running server, and SQL Server
2005 will recognize and use the additional memory immediately. This could prove
useful if you must increase available memory to service new business
requirements without affecting database availability. This feature also requires
hot-add memory support as provided in servers such as the IBM System
x3850 M2 and x3950 M2.
Feature comparisons
Most of the features that are mentioned in the following sections are found only in
the Enterprise Edition. For a detailed analysis of what is supported by Standard
Edition and Enterprise Edition, see the following documents:
Comparison Between SQL Server 2005 Standard and Enterprise Editions
http://www.microsoft.com/sql/editions/enterprise/comparison.mspx
SQL Server 2005 Features Comparison
http://www.microsoft.com/sql/prodinfo/features/compare-features.mspx
To group CPUs, you must edit the registry directly using a node configuration
affinity mask. After the soft-NUMA nodes have been created, you can assign
individual SQL Server instances to one or more soft-NUMA nodes.
You might create soft-NUMA nodes if your server hardware does not have
hardware NUMA capabilities or to sub-divide a NUMA node further. Each
soft-NUMA node gets its own I/O thread and lazy writer thread. If the SQL
instance has a high I/O requirement, it could be assigned two soft-NUMA nodes.
The SQL instance then has two I/O threads that can help it process I/O requests
better. Soft-NUMA provides the ability to fine-tune the use of the server
resources to ensure that critical databases get the resources that they require
within a consolidated environment.
CPU affinity and I/O affinity are SQL Server 2005 features for configuring each
database instance to use specific CPUs for database processing and I/O
requests. Assigning a set of CPUs only to handle I/O processing might provide
performance benefits with a database that relies heavily on I/O operations.
SQL Server 2005 is also able to allocate individual CPUs to a SQL database
instance using soft-NUMA and CPU affinity, so be sure no contention issues
arise during the configuration of WSRM.
Microsoft has positioned its new database platform as the answer to these
challenges and we have highlighted some of the new key features here. For a
comprehensive review of all the new features in SQL Server 2008, visit:
http://www.microsoft.com/sql/2008/default.mspx
2.8.1 Scale-up
Scaling-up is achieved by adding resources, such as memory, processors, and
storage, to an existing system that runs on a single server. It is also referred to as
vertical scaling. The benefit to scaling up is that it is relatively easy, because in
general it requires only hardware or software that is designed to take advantage
of additional memory, processor, and storage resources.
With the mass adoption of virtualization as a means for server consolidation and
driving up resource utilization on under utilized mulit-core processor based
systems, virtualization infrastructures are increasingly required to support larger
numbers of virtual and more diverse workloads with higher levels of software and
hardware redundancy. This has translated to virtualization trends seeking to
deploy virtualization platforms that deliver performance and availability, and also
the agility and flexibility to grow and shrink in line with business demands.
Scale-up systems, such as the IBM eX4 servers, are increasingly being exploited
by NUMA-aware virtualization hypervisors, such as VMware ESX, and the
demands of virtualized workloads.
For example, your database server might start out on a 2-way SMP system with
4 GB of memory and six hard drives. As the database grows in size or the
number of users increases, you can easily scale-up by adding more processors,
memory, and disk resources to maintain the same level of performance. You
might eventually have to replace the server with one that is capable of supporting
The x3950 M2 server is a good example of an SMP system based on eX4 and
NUMA technologies. The server starts with a base 2-way configuration and, as
your requirements grow, you can add incremental capacity to a maximum of 16
processor sockets (64 cores). Likewise, memory can be expanded from 4 GB to
1 TB. This modular server architecture delivers investment protection without the
up front costs of expensive switch-based alternatives.
2.8.2 Scale-out
Scale-out means adding discrete servers to your server farm to gain more
processing power. Although many options exist for implementing a farm
comprised of small low-end servers, we consider the use of the IBM
BladeCenter, 1U rack servers or iDataPlex for large scale-out implementations
such as the System x3550 as the most viable alternative when discussing this
requirement.
For example, your database server might start out on a 2-way system with 4 GB
of memory and six hard drives. As the database grows in size or the number of
users increase, you scale-out by adding another server with two processors,
4 GB of memory, and six disk drives to maintain the same level of performance.
Although you do not necessarily have to add another server with the exact
specifications, adding one does reduce the complexity of scaling out.
The benefit to scaling-out is that you can achieve near linear scalability. That is,
as you add each additional server to the system, you effectively increase your
system capacity proportionally. Thus, scaling-out provides much better returns in
terms of the additional costs associated with adding more servers to the system.
Another benefit inherent with scaling-out is that a cluster of smaller servers
generally costs less than a single large system.
Scale-up is preferred for smaller databases (150-200 GB). For larger databases,
large block I/O, data warehousing and decision support applications, use a
scale-out deployment.
Each server can be upgraded to a maximum of four processors. One, two, three,
or four processors are supported. Installed processors must be identical in
model, speed, and cache size.
Figure 3-1 on page 90 shows the locations of the four processors (CPUs),
locations of the required voltage regulator modules (VRMs), which you can
identify by the blue handles, and the memory cards that we describe later in 3.2,
“Memory subsystem” on page 111.
CPU/ VRM 3, 1, 2, 4
Air baffle
Memory card 1, 2, 3, 4
Figure 3-1 Top view of the x3850 M2 and x3950 M2 processor board; order of installation
Note: If you have a multinode complex and one or more nodes has only one
processor installed, and that node fails, then the complex will automatically
reboot without the memory resources and I/O resources that were in that
node. Therefore from a system availability perspective, we recommend you
have at least two processors in every node.
Table 3-1 Tigerton processor options for x3950 M2 and x3850 M2: machine type 7141/7144
Processor Cores Speed L2 / L3 TDPa Part Feature codeb
GHz cache number
Table 3-2 Dunnington processor options for x3950 M2 and x3850 M2: machine type 7233
Processor Cores Speed L2 / L3 cache TDPa Part number Feature codeb
Xeon L7445 4-core 2.13 GHz 2x3 /12 MB 50 W 44E4517 6976 / 4435
Xeon E7420 4-core 2.13 GHz 2x3 /12 MB 90 W 44E4469 3647 / 4419
Xeon E7430 4-core 2.13 GHz 2x3 /12 MB 90 W 44E4470 3648 / 4420
Xeon E7440 4-core 2.4 GHz 2x3 /12 MB 90 W 44E4471 3646 / 4418
Xeon L7455 6-core 2.13 GHz 3x3 /16 MB 65 W 44E4468 3645 / 4417
Xeon E7450 6-core 2.4 GHz 3x3 /16 MB 90 W 44E4472 3649 / 4421
Xeon X7460 6-core 2.67 GHz 3x3 /16 MB 130 W 44E4473 3644 / 4416
a. For a description of TDP, see footnote a in Table 3-1.
b. The first feature code is the base configuration. Additional processor orders should use the second
feature code.
Note: The x3950 M2 and x3850 M2 models with Xeon 7200 and 7300
Tigerton processors (machine types 7141/7144) do not support the
installation of Xeon 7400 Dunnington processors.
All processors must be installed with the VRMs and heat-sink, included in the
option packaging. If the VRM is missing, the following error message is displayed
and the system does not power up:
Processor configuration missmatch error
The blue locator LED is located on the rear side of the server. This LED
indicates whether AC power is connected to the system but the system is
powered off. The LED remains lit for up to 20 seconds after you remove the
power cords. Use this as a guide as to when you can start working in a
system. After the blue LED is off you can be sure that all components are
without power.
Captive
screws
Figure 3-3 x3850 M2 and x3950 M2 with fully opened media hood
Power Backplane
Fan 1 Fan 2 Fan 3
VRM 4
VRM 3
CPU
CPU
4
3
Memory Card 4
Memory Card 3
Memory Card 2
Memory Card 1
CPU CPU
1 2
VRM 2
VRM 1
8. If you are installing a processor in socket 2, remove the heat-sink blank and
store it for future use. Remove the protective cover, tape, or label from the
surface of the microprocessor socket, if any is present.
9. Note that sockets 3 and 4 are mounted on the processor board with the
processor’s release levers on opposite sides. These sockets are oriented
180° from each other on the processor board.
Verify the orientation of the socket before you install the processor in either of
these sockets. Figure 3-5 on page 97 shows the orientation of the sockets.
10.Note that the processor air baffle is always located between socket 1 and
socket 2, as shown in Figure 3-1 on page 90.
11.Lift the processor-release lever to the fully opened position, which is
approximately a 135° angle. See in Figure 3-5 on page 97.
12.Position the processor over the socket, as shown in Figure 3-6, and then
carefully press the processor into the socket. Close the processor release
lever.
Processor
Processor
orientation
indicator
Processor
socket
Processor
release
lever
13.Remove the heat-sink from its package and remove the cover from the bottom
of it.
Hint: The heat-sink in the processor option kit is already covered with the
correct amount of thermal grease.
Alignment posts
15.Ensure that the air baffle between processors is correctly installed, as shown
in Figure 3-1 on page 90.
16.Install a VRM in the connector next to the processor socket. The VRM and
handle are shown in Figure 3-8 on page 99.
Note: Make sure that the Front label on the VRM is facing the front of the
server.
17.Close the media hood, replace the cover and the bezel if you are not installing
other options.
The settings listed in the CPU Options menu are described in the following
sections:
“Active Energy Manager (power capping)” on page 100
“Processor Performance States” on page 101
“Clustering Technology” on page 105
“Processor Adjacent Sector Prefetch” on page 107
“Processor Hardware Prefetcher” on page 108
“Processor Execute Disable Bit” on page 108
“Intel Virtualization Technology” on page 109
“Processor IP Prefetcher” on page 109
“Processor DCU Prefetcher” on page 110
“C1E” on page 110
Only the onboard Base Management Controller (BMC) has Active Energy
Manager support. The BMC is on the system board and it shares the Ethernet
Processor power states (C-states) are low-power idle states. This means the
operating system puts the processor into different quality low-power states
(which vary in power and latency), depending on the idle time estimate, if the
operating system is idle. C0 is the state where the system is used. C1 to Cn are
states that are then set to reduce power consumption. After the idle loop is
finished the system goes back to the C0 state.
Lowering the processor performance state when processor demand is low can
significantly reduce CPU dynamic power consumption. These processor
performance states can be changed very quickly in response to processor
demand while software continues to execute. This technique, sometimes referred
to as demand-based switching (DBS), allows the operating system to provide
automatic scaling of the processor's power consumption in response to varying
workloads, with no required user intervention and no perceivable effect to system
performance.
The following sections show examples in Linux, Windows Server 2003, Windows
2008, and how you can identify the operating system power management
Linux
Various distributions of Linux operating systems, such as SUSE SLES10/SP2 or
RHEL5U2 based on kernel 2.6, integrate power management features by default.
The kernel processor frequency scaling subsystem can adjust the core
frequency as it goes. After the P-states are enabled in the BIOS, the subdevice is
found in the operating system at the following location:
/sys/devices/system/cpu/cpu0/cpufreq
To show the available core frequencies, which are adjustable, use the cat
command, shown in Example 3-1.
The command can also show the current value of the core frequency, as
indicated in Example 3-2.
For Intel processors, the Intelppm.sys file is used. Because the Intel Xeon is not
listed in the support matrix, the Microsoft generic driver Processr.sys is used
instead, until the support is available.
The power policies can be adjusted in the Power Options window. Each
processor power policy includes an upper and lower limit, referred to as the
Maximum processor state and the Minimum processor state. They determine the
range of currently available P-states that Windows may use. These values are
exposed in the Advanced settings panel of the Power Options window, shown in
Figure 3-11 on page 105.
You may set these values independently to define the bounds for any contiguous
range of performance states, or they may be set to the same value to force the
system to remain at a specific state. When you select a new target performance
state, Windows Server 2008 chooses the closest match between the current
power policy setting and the states available on the system, rounding up if
necessary.
For details about adjusting and optimizing your system to balance performance
to power consumption efficiency, see the following Microsoft Web page:
http://www.microsoft.com/whdc/system/pnppwr/powermgmt/ProcPowerMgmt.mspx
Clustering Technology
For certain operating systems, you must configure how the routing of processor
interrupts in a multi-processor system is handled. A low-level value sets the
multi-processor interrupt communication protocol (XAPIC). The settings are
functional only, and do not affect performance.
In the Clustering Technology menu, choose the appropriate mode for your
operating system, as advised in the operating system requirements, described in
5.4, “Installing the operating system” on page 264.
Although the IBM scalability chip can scale up to a total of eight nodes, IBM
supports only up to four nodes. The Hurricane controller supports four front-side
The 7300-series processors that we support with the x3850 M2 and x3950 M2
are dual-core or quad-core processors. Therefore, they have four processor
agents per processor package.
The IBM scalability chip can handle up to 128 agents in a maximum of eight
x3950 M2 scaled system complexes.
The following sections discuss the three available cluster technology modes:
Special mode
Logical mode
Physical mode
The sections also discuss the Linux and Windows operating systems regarding
clustering.
Special mode
This mode was created temporarily to allow 64-bit Linux operating systems to run
on the eX4 technology servers and the older X3 technology servers, if they do
not support clustering. Although flat mode is not really a clustering mode, it is a
mode by which most non-clustering operating systems can work simply, because
it abstracts them from the real physical clustering of the processors as a result of
the architecture. We do not recommend using this mode on X3 systems with
other operating systems, because it could cause situations resulting in failure to
boot. By definition, a maximum of eight agents are allowed, all are logical.
Because the special mode was developed for single-threaded processors only,
this mode is not used on x3850 M2 and x3950 M2.
Logical mode
This mode is applicable to XAPIC-based systems. A maximum of 60 logical
agents are supported, which means that 15 cluster IDs, each with four logical
agents, can be used.
When it is enabled the processor fetches both sectors of a cache line when it
requires data that is not currently in its cache.
For instance, only one 64-byte line from the 128-byte sector will be prefetched
with this setting disabled. This setting can affect performance, depending on the
application running on the server and memory bandwidth utilization. Typically, it
affects certain benchmarks by a few percent, although in most real applications it
will be negligible. This control is provided for benchmark users who want to
fine-tune configurations and settings.
If this option is enabled, and the operating system has marked the memory
segment as containing data, then the processor will not execute any code in the
segment. This parameter can be disabled in the BIOS, if the applications to run
on the server have problems with Execution Prevention. For added protection,
you might want to enable it, but you should first test your applications to ensure
they can continue to run as expected before you enable the option in a
production environment.
Processor IP Prefetcher
The purpose of the IP prefetcher, as with any prefetcher, is to predict what
memory addresses will be used by the program and deliver that data just in time.
To improve the accuracy of the prediction, the IP prefetcher tags the history of
each load using the Instruction Pointer (IP) of the load. For each load with an IP,
the IP prefetcher builds a history and keeps it in the IP history array. Based on
load history, the IP prefetcher tries to predict the address of the next load
according to a constant stride calculation (a fixed distance or stride between
subsequent accesses to the same memory area).
The IP prefetcher then generates a prefetch request with the predicted address
and brings the resulting data to the Level 1 data cache.
This setting is disabled by IBM (the opposite of the Intel default). The
observations in the IBM performance lab have shown that this prefetcher is
negligible in most real applications. Although Intel recommends that the
Processor Hardware Prefetcher option be enabled for some server workloads
similar to the Streams benchmark, the actual setting should be determined by
performance testing in your intended workload environment.
The hardware prefetch mechanisms are efficient, and in practice can increase
the success rate of the cache subsystem. However, the prefetch can also have
the opposite result. Frequent errors tend to pollute cache with useless data,
reducing its success rate. This is why you may deactivate most of the hardware
prefetch mechanisms. Intel recommends deactivating the DCU prefetch in
processors that are intended for servers, because it can reduce performances in
some applications.
Note: In the BIOS, all four prefetchers are disabled by default. Most
benchmarks in the performance lab indicate that this combination offers the
best performance because many benchmarks typically have very high CPU
and FSB utilization rates. Prefetching adds extra overhead, often slowly in a
very busy system.
C1E
This setting allows you to enable (by default) or disable the Enhanced Halt State
(C1E).
If the Enhanced Halt State is enabled, the operating system allows the processor
to alter the core frequency after sending an idle command such as HALT or
MWAIT. The processor core speed slows down and then transitions to the lower
voltage.
Enhanced Halt State is a low power state entered when all processor cores have
executed the HALT or MWAIT instructions and Extended HALT state has been
enabled. When one of the processor cores executes the HALT instruction, that
processor core is halted; however, the other processor cores continue normal
operation. The processor automatically transitions to a lower core frequency and
voltage operating point before entering the Extended HALT state.
While in the Extended HALT state, the processor will process bus snoops.
For more information about power saving modes, refer to the Intel Xeon 7200
and 7300 Processor Series Datasheet:
http://download.intel.com/design/xeon/datashts/318080.pdf
Important: All changes you make within the BIOS affects the particular local
node only. Any changes you make must also be made on all nodes within a
multinode configuration. Although you may have different settings on different
nodes, it is not recommended.
The number of installed DIMMs and memory cards as shipped is specified in the
server model description. You can obtain details in the latest version of
BladeCenter and System x Reference Sheet (xRef) at:
http://www.redbooks.ibm.com/xref
The DIMMs operate at 533 MHz, to be in sync with the front-side bus. However,
the DIMMs are 677 MHz PC2-5300 spec parts because these have better timing
parameters than the 533 MHz equivalent. The memory throughput is 4.26 GBps,
or 533 MHz x 8 bytes per memory port, for a total of 34.1 GBps with four memory
cards.
The server supports up to four memory cards. See Figure 3-12 on page 113.
Each memory card holds up to eight DIMMs to allow thirty-two DIMMs per
chassis.
Power Backplane
Fan 1 Fan 2 Fan 3
VRM 4
VRM 3
CPU
CPU
4
3
Memory Card 4
Memory Card 3
Memory Card 2
Memory Card 1
CPU CPU
1 2
VRM 2
VRM 1
You may order additional memory cards; use the part number in Table 3-4.
You must install at least one memory card with one pair of DIMMs because it is
two-way interleaved. These pairs of DIMMs must be the same size and type. The
server must have this to operate.
Note: When you install additional DIMMs on a memory card or a new memory
card, make sure they are installed in pairs.
You do not have to save new configuration information to the BIOS when you
install or remove DIMMs. The only exception is if you replace a DIMM that was
designated as Disabled in the Memory Settings menu. In this case, you must
re-enable the row in the Configuration/Setup Utility program or reload the
default memory settings.
If you replace the standard pair of DIMMs and install 32x 8 GB DIMMs and four
memory cards, both the x3850 M2 and x3950 M2 can be expanded to 256 GB.
The XceL4v Dynamic Server Cache consumes 256 MB in each chassis of the
main memory for use as L4 cache if a multinode system is formed, therefore
From a performance view of point, all nodes should have the same amount of
memory and the population within and over the memory cards should be
balanced. This reduction in memory is reflected in the power-on self-test (POST)
with the addition of a new line of text specifying the amount of available system
main memory after the L4 scalability cache for each node has been subtracted.
Example 3-6 shows what you see in a two-node complex.
To replace or add any DIMMs, remove one or more of the installed memory
cards, or add a new one.For an explanation of how this can even be done while
the system and the operating system are up and running, refer to sections:
3.2.3, “Memory mirroring” on page 118
3.2.4, “Hot-swap memory” on page 119
3.2.5, “Hot-add memory” on page 120
Figure 3-14 on page 116 shows the layout of a memory card with its available
status and failure LED indicators.
Light path
diagnostics button
DIMM 8 /
Error LED
– Performance-optimized configuration
Section 1.6, “IBM fourth generation XA-64e chipset” on page 27,
describes eight independent memory ports.
Therefore, to optimize performance, install four memory cards and then
spread the DIMMs, still installed in matched pairs, across all four memory
cards before filling each card with two more DIMMs, see Table 3-6.
1-4 1, 2, 3, 4 1/5
5-8 1, 2, 3, 4 2/6
9-12 1, 2, 3, 4 3/7
13-16 1, 2, 3, 4 4/8
– Reliability-increased configuration
To improve the reliability of your system fill your memory cards, as listed in
Table 3-7. Depending on the memory population, if a DIMM fails, it can be
removed and replaced by a new DIMM. The use of mirroring halves the
memory that is available for the operating system.
1, 2 1, 2, 3, 4 1/5
3, 4 1, 2, 3, 4 2/6
5, 6 1, 2, 3, 4 3/7
7, 8 1, 2, 3, 4 4/8
If you want to install the full 256 GB, remove the existing DIMMs and fully
populate the x3850 M2 and x3950 M2 with four memory cards, each with 8 GB
DIMMs.
The x3850 M2 and x3950 M2 have four separate memory power buses that each
power one of the four memory cards. Figure 3-12 on page 113 shows the
location of the memory cards, which are numbered 1 to 4, from left to right. The
DIMM sockets and Memory card LEDs are shown in Figure 3-14 on page 116.
Mirroring takes place across two memory cards, as follows:
The memory DIMMs in card 1 are mirrored to the memory DIMMs in card 2.
The memory DIMMs in card 3 are mirrored to the memory DIMMs in card 4.
Therefore, with memory mirroring enabled in the BIOS, you can hot-swap any
memory card if the hot-swap enabled LED is lit. For instructions to hot-swap a
memory card, see IBM: System x3850 M2 and System x3950 M2 User’s Guide:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073029
Table 3-8 shows the possible BIOS settings for the initialization of scrub control.
The setting is accessed by going to (from the system startup Main Menu)
Advanced Setup → Memory Settings → Initialization Scrub Control.
Scrub only after AC Cycle Performs scrub only after AC has been removed or applied
Note: A standard test is still performed across all memory and a run-time
scrub engine is always enabled regardless of these settings.
If memory mirroring is enabled, then the mirrored copy of the data from the
damaged DIMM is used until the DIMM replaced. After the damaged DIMM is
replaced, memory mirroring copies the mirrored data back to the new DIMM.
Note: Because of memory mirroring, you have only half of the total amount of
memory available. If 8 GB is installed, for example, then the operating system
sees 4 GB minus half the total XceL4v Dynamic Server Cache, if this is a
multinode system, after memory mirroring is enabled.
To easily identify whether hot-swap is enabled and the status of power to the
memory card, each memory card has a green memory hot-swap enabled LED,
and a green memory card power LED on the top panel of the memory card, as
shown in Figure 3-14 on page 116. The memory card has eject levers with
sensors, so that the system can recognize when a memory card is being
removed and power down that card’s slot accordingly.
For details about hot-swapping memory correctly and which sequence to follow,
see IBM: System x3850 M2 and System x3950 M2 User’s Guide:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073029
Requirements for enabling the hot-add memory feature on the server are:
The operating system must support the adding of usable system memory to a
running operating system. This is done with an ACPI sequence.
Currently, the only operating systems that have this capability and support on
the x3850 M2 and x3950 M2 are Windows Server 2003 and Windows Server
2008, both Enterprise Edition and Datacenter Edition.
Memory hot-add must be specifically enabled in the BIOS setup. When this is
done, the system allocates blank windows of memory space for future
memory additions. By enabling hot-add, memory mirroring will automatically
be disabled.
Memory cards 2 and 4 must not be installed because these are the only cards
that can be hot-added.
If only one memory card, memory card 1, is installed prior to the hot-add
operation, then only one more memory card may be added in slot 2.
If two memory cards are installed in slots 1 and 3, then two additional memory
cards can be added in slots 2 and 4.
For details about performing a hot-add operation, and restrictions, see IBM:
System x3850 M2 and System x3950 M2 User’s Guide:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073029
To access the settings and configure the memory subsystem in the server’s
BIOS setup:
1. Power on or reboot the system.
2. Press F1 during system startup, when prompted.
3. From the Main Menu window, choose Advanced Setup.
4. Select Memory Options. The Memory Settings selection menu opens, as
shown in Figure 3-15.
Memory Settings
Memory Card 1
Memory Card 2
Memory Card 3
Memory Card 4
Memory Array Setting [ HPMA (High Performance Memory Array ]
Initialization Scrub Control [ Scrub on every boot ]
Run Time Scrub Rate [ Default scrub rate ]
The memory configuration mode you select depends on the memory features
you want to use. Select one of the following modes:
HPMA if you are not using mirroring, hot-swap, or hot-add. This is now the
default or standard setting.
FAMM enables memory mirroring and hot-swap.
HAM enables hot-add in the future.
Note: The memory setting must be the same for all nodes in a multinode
complex before merging the scalable partition. This requires a KVM
connection to each node before the scalable partition is created.
Unlike with the x3850, the x3850 M2 and x3950 M2 support Memory ProteXion
with the HPMA setting, providing maximum performance, and they continue to
provide the reliability of Redundant Bit Steering (RBS).
In very large memory arrays, this particular memory scrub can take up to 10
minutes so you may choose to either disable this feature or only perform the
scrub when power has first been applied to the system.
Note that disabling the scrub or performing only the scrub after an AC cycle does
not eliminate the normal memory error detection capabilities of the system. Any
run-time correctable or uncorrectable memory error is still detected and the
failing memory device or devices are logged in the system event logs.
If you set the scrub rate to default, the chipset is configured to scrub the entire
memory array every 24 hours. The setting can help to ensure maximum system
performance by allowing the scrub engine to run only in this low speed mode.
You can use the fastest scrub setting if maximum system performance is not as
essential as maximum memory reliability. With this setting, the chipset scrubs the
entire memory array every 10 minutes.
Important: All changes you do within the BIOS affects the particular local
node only. Any changes you make must also be made on all nodes within a
multinode configuration. Although you can have different settings on several
nodes, it is not recommended.
Li-Ion
ServeRAID MR10k
Battery
iTBBU w/ custom
DDR2 connector
The LSI 1078 SAS controller operates in the Integrated RAID (IR) mode that
supports the internal port by communicating through a x4 SAS cable connection
to the SAS 4-port hot- swap backplane.
Figure 3-17 shows the internal SAS cable with the hot-swap backplane. The
cable is an industry standard MiniSAS 4i cable with SFF-8087 cable and board
connectors from the I/O board to the hot-swap backplane.
The hot-swap backplane is mounted in the media hood assembly and allows it to
attach up to four internal 2.5-inch SAS hot-swap disks drives. The internal SAS
connector on the I/O board located near the RSA II cable connector, as shown in
Figure 3-18 on page 126.
Figure 3-18 x3850 M2 and x3950 M2: internal SAS SFF-8087 board connector
SAS SFF-8088
external connector
on the rear of the
server
Figure 3-19 x3850 M2 and x3950 M2: external SAS SFF-8088 connector
2.5” hot-swap
SAS HDD
trays
Table 3-10 Supported internal disk options for x3850 M2 and x3950 M2
Part number Description
The ServeRAID-MR10k is installed into a dedicated socket on the I/O board near
the PCI Express slots as shown in Figure 3-21.
The ServeRAID-MR10k supports stripe sizes from 8 KB to 1024 KB. The default
stripe size is 128 KB. The SAS drive can be driven with up to 3 GBps for each
port in full-duplex mode.
RAID-10
A combination of RAID-0 and RAID-1, consists of striped data across
mirrored spans. It provides high data throughput and complete data
redundancy but uses a larger number of spans.
RAID-50
A combination of RAID-0 and RAID-5, uses distributed parity and disk
striping, and works best with data that requires high reliability, high request
rates, high data transfers, and medium-to-large capacity.
Note: Having RAID-0 and RAID-5 virtual disks in the same physical array
is not recommended. If a drive in the physical RAID-5 array has to be
rebuilt, the RAID-0 virtual disk can result in a rebuild failure.
RAID-60
A combination of RAID-0 and RAID-6, uses distributed parity, with two
independent parity blocks per stripe in each RAID set, and disk striping. A
RAID-60 virtual disk can survive the loss of two disks in each of the RAID-6
Note: RAID-50 and RAID-60, which are supported only externally, require at
least six hard disk drives.
Key features
The ServeRAID-MR10k controller has the following features:
Logical drive migration
You can increase the size of a virtual disk while the disk is online by using
RAID-level migration and by installing additional disks.
Global hot spare, dedicated hot spare
LSI defines any additional disks as global hot spares. These disk drives can
reconstruct your virtual disk in case of disk failures. Global hot spares are
defined to manage failing disks over all virtual drives.
You can assign a hot spare to a specific volume instead of to all available
volumes. Dedicated hot spares are used to recover from a failed drive in the
assigned virtual drive only.
Rebuild and rapid restore features
These reliability features help to secure your data.
Check consistency
This feature fixes media errors and inconsistencies.
Patrol read
This background operation, also known as data scrubbing, checks for media
errors in configured drives.
Selectable boot virtual disk
The first eight virtual disks can be chosen as a boot device.
The battery protects data in the cache for up to 72 hours, depending on operating
environment. IBM recommends that the battery be replaced annually.
Replacement part numbers are listed in Table 3-12 on page 132.
While the battery is charging for the first time, the controller cache is disabled
and set to the write-through (cache disabled) mode and is changed back to
the
write-back (cache enabled) mode automatically.
Installation guidelines
Consider the following guidelines when using RAID or installing the
ServeRAID MR10k:
No rewiring of the existing internal cabling is required when the
ServeRAID-MR10k is installed in an x3850 M2 or x3950 M2.
A RAID array created with the SAS LSI 1078 can be migrated for use with the
ServeRAID-MR10k, but the reverse is not possible.
This means that if you create RAID-0 (IS) and RAID-1 (IM) arrays using the
onboard LSI 1078 Integrated RAID controller, and later install a
ServeRAID-MR10k, you are given the option to convert those arrays to the
format used by the MR10k. However, if you want to later remove the MR10k,
you must first save all your data because the data in those arrays will be
inaccessible by the LSI 1078 Integrated RAID controller.
Warnings:
Prior to inserting the ServeRAID MR10k and converting your arrays you
must install the ServeRAID MR10 driver. Failure to do so prior to the
conversation will render all data on those drives inaccessible, permanently.
Existing arrays (created using the onboard RAID controller) will be
imported into MegaRAID arrays and they cannot be converted back again.
This is a permanent migration.
Installation instructions
To install the ServeRAID-MR10k controller and the battery package:
1. Turn off the server and peripheral devices, and disconnect the power cords
and all external cables as necessary to replace the device.
2. Remove the server cover.
3. From the server, remove the divider that contains the battery holder.
4. Open the retaining clip on each end of the connector.
5. Touch the static-protective package that contains the DIMM to any unpainted
metal surface on the outside of the server; then, remove the DIMM from the
package.
6. Turn the DIMM so that the keys align correctly with the slot.
7. Insert the DIMM into the connector by aligning the edges of the DIMM with the
slots at the ends of the connector.
8. Firmly press the DIMM straight down into the connector by applying pressure
on both ends simultaneously. The retaining clips snap into the locked position
when the DIMM is seated in the connector.
9. Install the iTBBU in the divider that contains the battery holder.
10.Install the divider that contains the iTBBU holder in the server.
Battery
Battery
cable
Cable Cable
guide guide
Battery
cable
connector
RAID
controller
For guidance with installing an SAS expansion enclosure, see 3.3.5, “SAS
expansion enclosure (unit)” on page 142. To configure your RAID controller, see
3.4, “Configuring RAID volumes” on page 154.
LSI 1078
SAS Li-Ion
SF8088
x8 PCI-Express Battery
x4
This adapter has 256 MB 667 MHz ECC SDRAM memory for cache on the card
as shown in Figure 3-25 on page 136. The cache has battery backup; the battery
is standard with the card.
The ServeRAID-MR10M has the same key features as the MR10k as described
in “Key features” on page 130.
iBBU mounts
directly to PCB
256MB 667MHz
onboard DDRII
Two Mini-SAS ECC SDRAM
SFF-8088 x4
external
connectors
This controller can be ordered with the part number listed in Table 3-13:
An iBBU can charge the battery pack automatically and communicate battery
status information such as voltage, temperature, and current, to the host
computer system.
Important: The battery in the iBBU must charge for at least six hours under
normal operating conditions. To protect your data, the firmware changes the
Write Policy to write-through until the battery unit is sufficiently charged. When
the battery unit is charged, the RAID controller firmware changes the Write
Policy to write-back to take advantage of the performance benefits of data
caching.
43W4343 Carrier
Installation guidelines
Consider the following guidelines when using RAID or installing the
ServeRAID-MR10M:
The ServeRAID-MR10M can be used in any slot of a x3850 M2 and x3950
M2 complex.
The ServeRAID-MR10M is not supported with the ServeRAID manager. Use
the MegaRAID Storage Manager (MSM) instead.
One or more arrays can be formed using up to nine cascaded EXP3000. An
x4 SAS SFF8088 cable is required to attach the external storage subsystem
to the ServeRAID-MR10M connectors.
To configure all 32 adapters you may remove the first sixteen adapters, configure
the remaining installed adapters, and then reinsert the adapters you removed
and configure them. Or, you may use the LSI MegaCli command line utility, which
does not have this limitation. This limitation should be fixed in future controller
firmware and MSM.
Installation instructions
To install the ServeRAID-MR10M SAS/SATA controller and the battery package:
1. Mount the iBBU package on the controller board.
2. Prepare the system: Unplug the power cords from the power supplies,
disconnect the computer from the network.
Important: Another blue locator LED is on the rear side of the server. This
LED indicates power in the system. It remains on (lit) for 20 seconds after
you remove the power cables.
The PCIe slots are beyond the adapter-retention bracket on the right at the
rear of the server shown in Figure 3-26 and Figure 3-27 on page 140.
5. Remove the screw for the non hot-plugable slot, or push the orange adapter
retention latch toward the rear of the server at the hot-plug, and open the tab.
That is where you add a PCIe adapter. The power LED for slot 6 and slot 7
turn off.
6. Insert the controller in a PCIe slot as shown in Figure 3-28 on page 141
Press Here
iBBU (Top View)
Press Here
85021-06
Edge of Motherboard
Edge of
Motherboard
7. Secure the controller to the computer chassis with the bracket screw. Close
the adapter-retention bracket, and use the retention pin as shown in
Figure 3-29 on page 142
Note: Secure all adapters with the retention pin. Adapters can loosen
during shipment or pushing the server out of the racks.
Adapter-retention
bracket
Removal button
10.Power off the system again. You can now install the disks in the external
enclosures and cable with the adapter.
You can attach the EXP3000 storage expansion unit to the SAS controllers
described in the following sections:
3.3.3, “ServeRAID-MR10k RAID controller” on page 128,
3.3.4, “ServeRAID-MR10M SAS/SATA II controller” on page 135
The EXP3000 machine type 1727-01X is available and shipped in the following
configurations:
Dual hot-swap redundant power supplies and two rack power cables
No disk; add up to 12 SAS, SATA or SATA II 3.5-inch hard disk drives
One ESM board (2 SAS ports); a second ESM is required for attachment to
dual controller models of DS3200/DS3400
Diamond icon
out port
Circle icon
in port
Key slot 2, 4, 6
Performance
Depending on your disk space requirement, type of application, and required
performance, the amount of disks in a number of chained EXP3000 can result in
higher performance, but does not increase linearly up to the maximum of nine
EXP3000 enclosures with up to 108 disk drives.
Installation guidelines
The EXP3000 is ready to use after minimal installation and configuration steps.
You must have the required rack space of 2U for each expansion unit.
Tip: You can reduce the weight for easier installation in your rack, while
you remove the power supplies. Press the orange release tab to the right,
just enough to release the handle as you rotate the handle downward.
Note: Install a minimum of four hard disk drives for each power supply to
operate in a redundant mode.
3. Start with the cabling of your expansion unit by connecting the power cables:
a. Attach one end of your 3 m SAS cable to the host port, shown in
Figure 3-32 on page 146.
b. Connect the other end to the closest EXP3000 IN-port.
c. Attach one end of the 1 m SAS cable to the EXP3000 OUT-port and the
other end to the IN-port of the next expansion. Repeat this step if you want
to attach multiple expansion units to this chain.
4. Turn on the EXP3000 expansion units before or at the same time as you turn
on the device that contains the RAID controller.
Figure 3-33 on page 147 shows the components of the ESM.
Note: The IBM Support for System x Web page contains useful technical
documentation, user guides, and so on. It is located at:
http://www.ibm.com/systems/support/x
Enter the system Type and Model 172701X in the Quick path field to link
to the information for this enclosure. We recommend you also check
regularly for new codes and tips. Use the Download and Troubleshoot
links.
Your enclosures are now prepared for configuration RAID volumes, which is
described in the next section.
3. Boot the server with disk 1 and insert disk 2 when prompted. The utility
attempts to flash the onboard LSI 1078 controllers in you system. You might
see warning messages stating that this flash is not compatible with all
controllers, as shown in Figure 3-35. These are not error messages.
****************************************
* *
* SAS Firmware & BIOS Flash Disk *
* *
****************************************
NOTE: This utility will scan all LSI 1064, 1068, and 1078 based
controllers in the system. Only the controllers which match
the update type will be flashed.
.
Do you wan to continue[Y,N]?_
C:\>_
1. Extract the contents of the package to floppy disks and then flash the
ServeRAID controller.
2. Download the most recent firmware:
– IBM ServeRAID-MR10k firmware update (Windows)
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073139
– IBM ServeRAID-MR10M firmware update (Windows)
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073389
We used the Microsoft Windows package and ran it on a Windows
workstation to create bootable diskettes. With these diskettes, you can boot
the server by an attached USB floppy drive or mount it in the Remote
Supervisor Adapter II remote drive in the Web interface as described in 6.2.4,
“Remote console and media” on page 324.
3. Start the executable package and create three diskettes (the first is bootable)
by selecting Extract to Floppy (in Figure 3-37 on page 151).
4. Boot the server with disks 1, 2, and 3 as prompted. It attempts to flash all
ServeRAID-MR10k or ServeRAID-MR10M controllers in the system. You
might see warning messages stating that this flash is not compatible with all
controllers, as shown in Figure 3-38. These are not error messages.
****************************************
* *
* SAS Firmware & BIOS Flash Disk *
* *
****************************************
This program will update the firmware on all IBM ServeRAID-MR10k
controllers in the system.
.
Do you wan to continue[Y,N]?_
5. Press Y to start the flash process. The update starts, as shown in Figure 3-39
on page 152.
C:\>_
To update the SAS hard disk drives internal in the media hood assembly and
external in the EXP3000 installed:
1. Download the bootable CD from the following Web address:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-62832
2. Burn a CD or mount the ISO image in the Remote Supervisor Adapter II
remote drive of the Web interface as described in 6.2.4, “Remote console and
media” on page 324.
3. Boot your server from this CD image.
4. Update the hard drive disks by clicking the Update button.
The update program window is shown in Figure 3-40 on page 153.
We recommend also updating the EXP3000. At the time of writing, the latest
update contained a critical fix for the ESM, which fixed the problem of an
unexpected reboot of the EXP3000.
Mouse [ Installed ]
Video
Decide which controller you want to boot the server from and select it in the PCI
Device Boot Priority field in Start Options menu, Figure 3-43.
Start Options
Startup Sequence Options
Power on your server. Watch the window in POST. Press Ctrl+C while the
messages in Figure 3-44 on page 156 are shown.
Initializing ..-
The LSI 1078 onboard SAS controller BIOS starts after the POST finishes.
Select the controller to start the LSI controller configuration utility. Figure 3-45
shows a two-node multinode system; the first controller is the primary node, the
second controller is the onboard SAS controller of the secondary node.
Creating volumes
To create a volume, select the LSI controller you want to configure by highlighting
it, as shown in Figure 3-45, and then pressing Enter. From the submenu that
You can create two volumes of the same type or a mixture of Integrated Mirroring
(IM) or Integrated Striping (IS). Select the RAID level IM or IS.
In the RAID creation menu, Figure 3-47, select the hard disk drives you want to
add, for example, to a RAID-0 (IS). You can add up to 10 hard disk drives to one
volume. Press C to create this array.
Figure 3-47 shows a previously created volume in RAID-1 (IM) and one assigned
hard disk drive as a hot-spare disk. see in Figure 3-48 on page 158.
You can assign a maximum of 12 hard disk drives to two volumes plus two hard
disk drives to hot spares.
After you create the volumes, select the RAID Properties menu again. You can
check the state of the created volumes. Figure 3-49 shows a missing disk in the
mirror, which will be synchronized after a new disk is inserted.
After you press Ctrl+H keys the WebBIOS starts after the POST finishes:
WebBIOS will be executed after POST completes
You can migrate RAID-0 and RAID-1 volumes that you created using the onboard
LSI1078 SAS controller to MegaRAID (MR) mode by installing the
ServeRAID-MR10k. The created volumes are automatically imported after the
first installation of this controller.
Note: You cannot reverse this step to go back to using the onboard SAS
controller if you remove the ServeRAID-MR10k again.
To choose the adapter you want to configure, use either the mouse or keyboard
(Tab key) to select the adapter, and then click Start or press Enter to begin the
process. The WebBIOS main window opens.
Note: If a mouse is not available, you can always operate with the Tab key to
switch between options and buttons and spacebar or Enter to select any.
The window also includes icons, as shown in Figure 3-53 on page 161.
Options in the Controller Function menu are explained in Figure 3-54. Details
about several of these options are provided in sections that follow.
The first page, in Figure 3-55, shows the physical specification of the adapter and
version for BIOS, firmware package, and summarized information for physical
and defined virtual drives.
Click Next to view the second page of the Adapter Properties window, shown in
Figure 3-56 on page 163. On the second page, click Next again to view the third
page, shown in Figure 3-57 on page 163. You may change any of the default
settings of the ServeRAID adapter properties.
Note: The second page allows you to change the Rebuild Rate, which can
negatively affect the performance of the SAS subsystem such as by
performance degradation within the operating system and its applications.
In the second and third pages, change any of the following default settings:
Battery Backup
This indicates whether you installed a battery backup unit (BBU) on the
selected controller. If a BBU is present, click Present to view details about it.
See details in Figure 3-58 on page 164.
Tip: We do not recommend setting the lowest nor the highest value.
A very low percentage can result in a loss of data, in case further hard
drives within the same RAID subset (VD) are going offline while the
Rebuild process is running.
We recommend that you enable Adapter BIOS only on the controller that
contains the boot device, and disable it on all others. This maximizes the
available PCI ROM space and reduces the chance of getting PCI ROM
Allocation errors.
This option indicates the rate for patrol reads for physical drives connected to
the selected adapter. The patrol rate is the percentage of system resources
dedicated to running a patrol read.
Tip: The corrective action depend on the virtual configuration and the type
of errors. It affects performance, the more iterations there are, the greater
the impact.
This view is available after creation of virtual disks only. Virtual disks, also known
as logical drives, are arrays or spanned arrays that are available to the operating
system. The storage space in a virtual disk is spread across all the physical
drives in the array.
Note: You can watch the status of any ongoing background process by
clicking the VD Progress Info button.
Operations panel
Many of the same operations listed in Figure 3-60 on page 168 can be performed
from the Operations panel shown in Figure 3-62.
Locate (Loc)
Causes the LED on the drives in the virtual disk to flash
Fast Initialize (Fast Init)
This initializes the selected virtual disk by quickly writing zeroes to the first
and last 10 MB regions of the new virtual disk. It then completes the
initialization in the background.
Slow Initialize (Slow Init)
This also initializes the selected virtual disk but it is not complete until the
entire virtual disk has been initialized with zeroes.
Use this window to specify the following policies for the virtual disk:
Access: RW is the default
Specify whether a virtual drive can be accessed by read/write (RW, default),
read-only (R), or no access (Blocked).
Read: Normal is the default
This is the read-ahead capability. The settings are:
– Normal: This is the default. It disables the read ahead capability.
– Ahead: This enables read-ahead capability, which allows the controller to
read sequentially ahead of requested data and to store the additional data
in cache memory, anticipating that the data will be needed soon. This
speeds up reads for sequential data, but there is little improvement when
accessing random data.
– Adaptive: This enables the controller to begin using read-ahead if the two
most recent disk accesses occurred in sequential sectors. If the read
requests are random, the controller reverts to Normal (no read-ahead).
Write: WBack is the default
This is the write policy. The settings are:
– WBack: This is the default. In Writeback mode, the controller sends a data
transfer completion signal to the host when the controller cache has
received all the data in a transaction. This setting is recommended in
Standard mode.
After a migration has been started, it runs in the background, in case you did
not change the default for the background initialization (BGI) setting. The
status of the virtual disk changes from Optimal to Reconstruction and
changes back after the task has completed successfully.
To remove physical drives:
a. Select Remove physical drive to remove a physical drive.
b. Select the physical drive you want to remove.
c. Click Go.
Note: When removing a drive, follow the steps, otherwise the drive might
not be stopped properly, which could damage the drive. The hard disk must
be allowed to spin down before you remove it from the slot.
On the panel that is displayed (shown in Figure 3-65 on page 175), you can
rebuild a drive either by selecting Rebuild, or by selecting Properties and then
working with the extended properties to do the rebuilding.
1
At the time of writing, the software User’s Guide does not list this migration option. However we
verified that the option is available in our tests in our lab.
Unconfigured Good
This is a disk that is accessible to the RAID controller but not configured as a part
of a virtual disk or as a hot spare. See Figure 3-67.
Unconfigured Bad
This state is a physical disk on which the firmware detects an unrecoverable
error, the physical disk was Unconfigured Good, or the physical disk could not be
initialized.
Online
You can access at a physical disk if it is in this state because it is member of a
virtual disk configured by you. Figure 3-68 shows the choices.
Important: If you set a drive offline, this can result in loss of data or
redundancy. We recommend to backup all important data, because the
selected physical disk is member of the virtual disks.
Note: We recommend using this option before you remove any disk from
the enclosure or the internal disk bays, to prevent damage to a disk that did
not spin down.
Offline
A drive you forced from Online to Offline can result in loss of data, redundancy, or
both. This depends of the RAID level you have configured. Figure 3-69 on
page 178 shows the choices.
Mark as Missing: This sets this command to mark this drive as missing.
Locate: This flashes the LED on the selected disk.
Failed
This shows you a physical disk that was configured as online or hotspare but
failed, but the controller firmware detected an unrecoverable error and marked
this disk as failed.
Tip: After a physical disk failed you should check the events at this disk. We
recommend to replace a failed disk. However after you have a valid backup for
any data you want to keep, you can try to rebuild this disk. If it failed or a
critical event is logged against this disk, replace it.
Missing
A missing drive is a physical drive that is removed or marked missing from an
enclosure and is a member of a configured virtual disk.
Tip: If the drive state is Missing but the drive is physically present, try
reseating it and seeing if that changes the state back to Unconfigured Good.
A physical disk previously forced to offline can be marked as missing also. See
“Offline” on page 177.
Note: When a disk is missing that brings a defined virtual disk to the degraded
state, a rebuild process (reconstruction) of the array starts immediately, if a
Global or Dedicated Hotspare is enabled.
Rebuild
This state indicates that the selected disk is in the progress of reconstruction of a
virtual disk (array). A rebuilding starts if you have a missing or failed disk and
another Unconfigured Good disk was enabled as Global or Dedicated Hotspare,
or if a physical disk that is in state Unconfigured Good is forced by you to
Rebuild. In both situations, reconstruction of the array starts immediately.
To start the Configuration Wizard, select it from the Function Controller panel
shown in Figure 3-54 on page 161. The Configuration Wizard window opens, as
shown in Figure 3-71 on page 180.
The three configuration types, shown in the figure, are discussed in the following
sections:
“Clear Configuration” on page 180
“New Configuration” on page 181
“Add Configuration” on page 187
Clear Configuration
Use this to clear existing configurations.
New Configuration
This option clears the existing configuration and guides you to add new virtual
disks.
Important: If you use this operation, all existing data will be lost. Be sure that
you do not require any existing virtual disks and their data. The following
reminder appears on the window:
This is a Destructive Operation! Original configuration and data
will be lost. Select YES, if desired so. <YES> / <NO>
After you select Yes, the following options are available for creating new disks, as
shown in Figure 3-72:
Custom configuration
Auto configuration with redundancy
Auto configuration without redundancy
Table 3-17 Algorithm to select the RAID level for automatic configuration
If you select this option... And you have ...then this RAID level
this many drives is automatically
installed... selected
Custom Configuration
When you want to set up different or more virtual disks, use the custom setup.
You can assign all available disks that have the status of Unconfigured good
to different disk groups as follow:
a. Select the disk in the panel view.
b. Click AddtoArray to assign to a disk group. Repeat this step to add more
disks.
TIp: You can assign multiple disks at the same time if you press the
Shift key and use the Up and Down arrow keys to select the disks, can
then click AddtoArray.
Tips: All disks that you define within a disk group will be assigned to the
virtual drives that you create later.
At the time of writing, the Reclaim button does not have any affect to
undo a disk addition.Instead, go back one window to the configuration
selection menu and start the process of creating a disk group again.
e. Click Next.
f. In the Array With Free Space panel, shown in Figure 3-75, select the disk
group for which you want to create a virtual disk.
Note: The WebBIOS suggests all available RAID levels for disk groups.
Note: The value of the stripe size affects system performance and
depends on the application you want to use. Learn what the best
value for your application is and fine tune later, if necessary.
• Access Policy: Select a policy for the virtual drive. Values are
read/write, read-only or not blocked.
• Read Policy: Specify the read policy for this virtual drive:
Normal: This disables the read ahead capability. This is the default.
Note: After you confirm to save the new configuration, the new created virtual
disks are shown in the Virtual Drives view at the Home Screen View.
Add Configuration
This option retains the old configuration and then adds new drives to the
configuration. You can add new virtual disks by using Auto Configuration after
inserting new disks, or by using Custom Configuration if you have unallocated
space (a hole) in the assigned disk groups.
Note: This is the safest operation because it does not result in any data loss.
Events option
The WebBIOS Event Information panel keeps any actions and errors that are
reported by the selected controller. The events can be filtered at the different
RAID components for:
Virtual Disk
Physical Device
Enclosure
BBU
SAS
Boot/Shutdown
Configuration
Cluster
Define the start sequence and the number of events you are looking for.
The slots are connected to one of two 4-port PCIe x8 controllers as follows:
PCIe controller 1: Slot 1-4
PCIe controller 2: Slot 5-7, onboard SAS LSI1078
The onboard SAS LSI1078 uses a dedicated PCIe port to the second PCIe
controller. The onboard SAS controller and the optional ServeRAID-MR10k are
discussed in 3.3, “Internal drive options and RAID controllers” on page 124.
The Remote Supervisor Adapter II and the new Broadcom NIC interface are
linked to the Intel ICH7 Southbridge, so that it is no longer shared with the
onboard SAS and RAID controller because it is at X3 chipset based systems.
You can use the Configuration Setup utility program to change the sequence and
have the server scan one of the first six PCIe slots before it scans the integrated
devices. Press F1 when prompted during system POST. In the Start Options
panel, set the PCI Device Boot Priority to the slot you want to boot from as
shown in Figure 3-78. The default is Planar SAS.
Disabling of specific features such as PXE boot at Ethernet adapters or the BIOS
at HBA adapters, that you do not have to boot from SAN devices, and disabling of
onboard devices are also options for solving any PCI ROM allocation errors. Your
in country technical support team may advice here.
IBM provides the relevant Active PCI slot software for Windows 2000 and
Windows Server 2003 from the following Web address:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-62127
Microsoft Windows 2008 supports several PCI Express features natively, such as
Active State Power Management (ASPM), Advanced Error Reporting as part of
Windows Hardware Error Handling Architecture (WHEA), Extended configuration
space access through the Memory-Mapped PCI Configuration Space (MCFG)
table, Hot-plug, Message Signaled Interrupt (MSI), and Power Management
Event (PME).
IBM provides Active PCI software for Windows Server 2008 from the following
Web address:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5074966
An asterisk (*) next to the slot number indicates that more than one device is in
this slot. Slot 0 is the planar system board, which contains the devices found at
all chassis in a multinode configuration in the POST after its merged.
Scroll to the
next/previous
PCI device
Slot/Device
Information
Vendor based
information
ROM usage
in POST and
Run time
The most important information that can be checked here is the initial ROM size
and whether all resources can be assigned correctly. If a system reports an
initialization error, by showing a message in POST or in the server error logs. The
initial ROM size of each installed and enabled PCI device must be determined.
This list is updated regularly and might contain limitations for using a specific
PCI Express option.
Choose your system and scroll to the required subcategory, based on the type of
adapter you are looking for. Select the link at the relevant option to get detailed
information.
Four nodes
8-way or 16-way
Three nodes (Each node is
2-way or 4-way)
6-way or 12-way Up to 1 TB RAM
(Each node is
Two nodes
2-way or 4-way)
4-way or 8-way Up to 768 GB RAM
x3950 M2
(Each node is
2-way or 4-way) x3950 M2
One node Up to 512 GB RAM x3950 M2
2-way or 4-way
Up to 256 GB RAM
x3950 M2 x3950 M2 x3950 M2
After multiple nodes are cabled together to form a multinode complex, the next
step is to define the single system image, a procedure called partitioning. As a
result, all hardware resources of each node in a configured partition are bound to
the one system and are shown to the installed operating system as a single
system.
Any RSA Adapter II in any x3950 M2 in this complex can be used to control the
partitions. The RSA II is necessary for configuration and management purposes,
and offers a very clear Web interface layout. Learn more about working with the
RSA2 Web interface by reading sections 4.7, “Configuring partitions” on
page 220 and 4.8, “Working with partitions” on page 228.
Because of the new scalability features that are included in the BMC firmware, a
disconnection of any RSA2 LAN interface does not affect the stability of your
production running multinode system. The BMC handles all instructions and
discovers the complex information through the scalability cables.
Local RSA II
n-node
Get Local Remote n-node
Complex
Chassis information BMC
Descriptor
Remote n-node
RSA II
Get n-node
Chassis information
The BMC gets the partition descriptor information of all other nodes in this
complex by exchanging information through the connected scalability cables on
the scalability ports at the rear side of your server. Figure 4-10 on page 211
shows the scalability ports.
Note: The Complex Descriptor is unique; it contains all system and partition
information.
Note: The BMC in each system discovers and keeps consistent complex and
partition information of all nodes in a complex. This allows the RSA2 in any
system in a complex to manage the all of the partitioning.
To isolate the failure, we recommend that you check the system event
information in the RSA II Web interface on all connected nodes.
See Figure 4-31 on page 238.
Note: Our recommendation is to flash all the system components after you
install the ScaleXpander chip. We recommend you reflash these
components, even if the latest versions are already installed.
Note: We recommend synchronizing the system time and the RSA II in the
BIOS to be sure they have the same time stamp information for events that
can occur. See Figure 4-3 on page 204. You may also set up and use a
time server for the operating system and Remote Supervisor Adapter II.
Boot to System Setup (press F1 when prompted) and select Date and Time.
Figure 4-3 on page 204 appears.
After you install the ScaleXpander chip, the Scalable Partitioning submenu
appears in the RSA II Web interface, shown in Figure 4-4.
3. Remove the cover and the front bezel of x3850 M2 server. See instructions in
the System x3850 M2 and x3950 M2 User’s Guide.
4. Loosen the captive screws and rotate the media hood (Figure 3-3 on
page 95) to the fully opened position. You now an open view to the processor
board.
5. Locate the ScaleXpander key connector at the front of the processor board.
See Figure 4-5 and Figure 4-7 on page 206.
6. Check that the ScaleXpander key is oriented correctly direction and push it
into the blue slides, which hold the ScaleXpander key in place, until it is firmly
seated. You hear a clicking sound. See Figure 4-7 on page 206.
7. Close the media hood assembly and tighten the captive screws.
The x3850 M2 is now enabled for scalability. To indicate that the server is now an
x3950 M2, you might want to replace the front bezel of the x3850 M2 with the
one provided in the ScaleXpander Option Kit. Follow the replacement
instructions included in the ScaleXpander Option Kit.
Note: The new bezel does not have a sticker where you can enter the original
serial number of your server. You should record this data separately.
You could also set up the RSA IP for each DHCP request. The following IP
settings are offered in the BIOS and RSA II Settings:
– Try DHCP then use Static IP (default)
– DHCP Enabled
– Use Static IP
Note: The RSA Adapter II should provide its LAN settings, at the latest
after 20 seconds, to the switch. The RSA II is pingable.
6. Use your workstation to ping and access the Remote Supervisor adapter’s
Web interface as described in 6.2.3, “Web interface” on page 321.
Ensure that all systems are at the same code levels. Failure to do so could cause
unpredictable results. As we described earlier, we recommend you force a
reflash of all firmware to prevent unexpected behavior.
The minimum required code levels for three-node and four-node system are
shown in Table 4-3.
Table 4-3 Three -node and four-node x3950 M2: system code level
Description Image release Build level
Important: After updating the BIOS, reset the BIOS to the default settings.
Our observation has shown that using system flash utilities that are based on the
following items, do not reflash any component if the version in the update source
is the same as that on the system:
UpdateXpress CD or UpdateXpress system packs
Operating system firmware update packages for Windows (wflash) or Linux
(lflash) systems
After applying the new codes, power off all systems, and then remove the power
source for 30 seconds.
This section assumes you have already installed the systems into your rack using
the instructions in Rack Installation Instructions for System x3850 M2 and x3950
M2, available from:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073030
Note: The scalability cables used to join the nodes together are a specific
length; no empty U-spaces should exist between the nodes.
Before you begin the multinode cabling, install the Enterprise Cable Management
Arm (CMA) that is shipped with your x3950 M2.
If your x3950 M2 has been upgraded from an x3850 M2, remove the standard
cable management arm that came with the x3850 M2 and replace it with the
Figure 4-10 on page 211 shows the ports of the x3950 M2 where the scalability
cables are installed. Depending on the number of nodes you plan to connect
together, you use either two or three of these ports in each node.
The ports are named port 1, port 2, and port 3, starting from the left. Each ports
has an indicator LED that shows an active connection on this port.
Figure 4-11 shows the scalability cable connector. The connectors are the same
on both ends of the cable.
Insert the cables on one server first, follow the installation guidance in the next
sections to route the scalability cables through the cable management arm.
Figure 4-12 on page 212 shows cables inserted.
Tip: After you insert a cable into the scalability port, ensure that the cable is
fully inserted and securely in place.
Management
Network
Node 2
3. Route the cables through the Enterprise Cable Management Arms (CMAs) as
follows (refer to Figure 4-14 on page 214), ensuring that in each case, you
have enough spare length of the cable between the rear side of the server
and the first hanger bracket, and on the front side of the cable management
arm:
a. Route the cable connected on port 1 of node 1 through the hanger
brackets on the cable management arms.
b. Route the cable connected on port 2 of node 1 through the hanger
brackets on the cable management arms.
4. Ensure each server can be pulled out fully at the front of the rack and the
scalability cables are clearly routed.
Node 2
Management
Network
Node 3
One 3.0 m cable is included with every the x3950 M2 or the ScaleXpander
Option Kit (for upgraded x3850 M2s). The fifth longer (3.26 m) cable is included
in the IBM Scalability Upgrade Option 2 Kit. These kits are described in Table 4-1
on page 202.
Node 2
Management
Network
Node 3
Node 4
The cabling of the scalability cables is now finished. Read the next section to
create a partition.
The Standalone Boot buttons can either force a partition out of stand-alone
mode, or put back in stand-alone mode, as follows:
– Stand-alone mode
The Force button resets a multinode system and boots a selected system
in stand-alone mode. First select the PID of the partition you want to
control, and then click Force.
Notes:
• The stand-alone mode allows the systems to boot separately for
debugging or diagnostic purposes.
• If the system is in this mode already, the command is ignored.
– Multinode mode
A system in a partition that is started in stand-alone mode can be
reassigned to boot in the multinode mode again. Select the SN (serial
number), then click Undo.
Notes:
• One or more systems that are members in a partition and booted in
stand-alone boot, swap back to the boot mode multinode.
• If the system is in multinode mode already, this command is ignored.
Partitions Configure buttons are shown in Figure 4-21.
Figure 4-21 RSA II Web interface: Partition Configure buttons (create partitions)
Figure 4-22 RSA II Web interface: Reset Defaults (reset the partition information)
The Reset button clears all the partition information in the complex. Use this
function when data is corrupted and a system becomes stuck in an invalid
partition where the firmware cannot detect it. All partition information from all
systems is cleared. Click the Reset button to delete the partition information
of all systems.
Partition Reorder functions redraws the boxes and serials numbers, as shown
in the before and after reordering in Figure 4-23.
before → after
System box
Partition ID
• Grey = cabled only, no
System serial partition
number • Blue = valid Complex
Descriptor; partition
Checkmark for the created
primary server • Red = Complex
Descriptor could not be
read.
The section on the right side in Figure 4-18 on page 220 (and also shown in
Figure 4-25 on page 225) indicates:
System power state
Partition state
Scalability mode
4. Check the system and partition states on the right side of the Scalable
Complex Management panel (shown in Figure 4-18 on page 220 and
Figure 4-25 on page 225). The states should indicate:
– Each system is Stopped.
– The partition state at each node is Invalid and is colored black.
Note: An Invalid system that is colored red shows you red colored
symbolic boxes in the middle section too.
Tip: We recommend that you reorder the systems in the configuration window
to match the physical installation in your rack. See Partition Reorder in 4.7.1,
“Understanding the Scalable Partitioning menu” on page 220 for details.
Partition ID: 1
SN: 23A0509 (Local)
Primary Stopped Valid Multinode
Port 1 Port 2 Port 3
Partition ID: 1
SN: 23A0143
Primary
Stopped Valid Multinode
Port 1 Port 2 Port 3
If you want to create a second partition in this complex, repeat the steps. You
may create partitions that are comprised of two, three, or four nodes.
Start a partition
All nodes in the partition power on immediately, after a partition is started.
To start a partition, select the Partition ID (PID) at the top of the blue symbolic
box at the primary node. Click Start (under Partition Control). The following
events occur, as shown in Figure 4-27 on page 229:
The selected partition starts immediately.
The system’s state changes from Stopped to Started.
Partition ID: 1
SN: 23A0509 (Local)
Primary Started Valid Multinode
Port 1 Port 2 Port 3
Partition ID: 1
SN: 23A0143
Primary
Started Valid Multinode
Port 1 Port 2 Port 3
The system starts the POST. In the POST you can watch the merge process on
the screens of all nodes in this partition. The process indicates:
The primary node searches for the secondary node, shown in Example 4-1.
The merging process can bypassed by pressing the blue REMIND button
located on the Light Path Diagnostics panel, shown in Figure 4-28 on
page 230.
Pressing the button breaks this node from the merging process. The
remaining nodes merge as normal, if partial merge is enabled.
Figure 4-28 x3950 M2: Light Path Diagnostics panel with blue REMIND button
After the boxes merged successful you can see the following information
about the window on the different boxes:
– Primary server display, shown in Example 4-3.
64 GB Memory: Installed
512 MB Memory: Consumed by Scalability
The white scalability LED on the front of all nodes, which merged successfully
to this partition, becomes solid on. This LED goes off again after a system in
this partition is swapped to the Standalone mode, followed by a reboot.
Figure 4-29 x3950 M2: Scalability LED and port link LED
The scalability port LED becomes solid on in green at the ports where a
scalability connection has been established.
Reset a partition:
To reset a partition, select the Partition ID you want to reset. Then, click Reset
(under Partition Control). The following events occur:
The Multinode resets.
A running operating system resets also.
You can also interrupt nodes after they are merged by pressing the Esc key on
your keyboard in POST. The nodes reset and bypass to merge on next boot. See
Example 4-5 on page 232.
Delete a partition
To change your partitions or add another system, clear this partition first and then
configure your new partition. Select the Partition ID, power off the systems in this
partition, and then click the Delete button.
Note: A partition or all partition settings cannot be deleted if any of the system
is powered on.
First, however, review the following checklist again if you experience difficulties
creating partitions:
Did you upgrade each x3850 M2 system to a x3950 M2 system to be capable
for scalability configuration as described in 4.5, “Upgrading an x3850 M2 to an
x3950 M2” on page 204?
Does each system match the prerequisites as described in 4.4, “Prerequisites
to create a multinode complex” on page 201?
Did you provide the correct cabling at all nodes depending on the number of
nodes in your complex as described in 4.6, “Cabling of multinode
configurations” on page 209?
Are you familiar with the Scalability Manager in the Remote Supervisor
Adapter II as described in 4.7, “Configuring partitions” on page 220?
To power off a running Multinode system, use any of the following methods:
Use the power button on the front of the Light Path Diagnostic panel at any of
the merged nodes in this partition. See Figure 4-28 on page 230
Use the reset options in the Power/Restart control of the RSA II Web
interface, which is installed in any of the x3950 M2 systems. See Figure 4-30
on page 233.
Use the reset option in the scalability management menu of the RSA II Web
interface. See “Power off a partition” on page 231.
Perform a shutdown in the operating system.
The systems cannot be bypassed for merging to boot in stand-alone mode. This
cannot be done remotely if the merge process is not completed.
If the RSA II scalability management menu does not enable you to manage any
partition, restart or power off the systems; the system can be recovered only by
an AC power cycle.
Note: You can bypass the boot to stand-alone remote if the scalability
management interface is unresponsive only by a support person who is
on-site, who must push the blue REMIND button.
Ensure that no hardware error occurred. See Figure 4-31. In the RSA II Web
interface, select Monitors → Event Log at each node in this partition.
Check the Severity column to determine if any events are colored with red or
yellow at this specific time frame (see Figure 4-32). You can filter the events at
Severity, Source, and Date.
64 GB Memory: Installed
512 MB Memory: Consumed by Scalability
After the scalability mode is changed to stand-alone mode, the serial numbers of
each system is selectable, as shown in Figure 4-33.
.
Partition ID: 1
SN: 23A0509 (Local)
Primary Started Valid Standalone
Port 1 Port 2 Port 3
Partition ID: 1
SN: 23A0143
Primary
Started Valid Standalone
Port 1 Port 2 Port 3
If you add a check mark in one of the serial number (SN) boxes, then click Reset,
the server immediately resets. The secondary server logs an error and reboots
unexpectedly several seconds later.
Do not reboot one system in stand-alone mode after the scalability mode was
changed to stand-alone mode and all other systems are still running in the
operating system.
Partition ID: 1
SN: 23A0509 (Local)
Primary Stopped Valid Multinode
Port 1 Port 2 Port 3
Partition ID: 1
SN: 23A0143
Primary
Started Valid Multinode
Port 1 Port 2 Port 3
As Figure 4-34 shows, the first system in the list is Stopped, however the second
system in the list is Started.
The power state is shown as Off in the Server Power/Restart Activity panel,
shown in Figure 4-35 on page 243.
You can work around this error by removing all systems in the partitions from the
AC power source. Wait 30 seconds and then replug the AC power.
We recommend that you regularly check for newer system firmware code and
updated product publications such as the Problem Determination and Support
Guide and new tips on the Support for IBM System Support Web pages at:
http://www.ibm.com/systems/support/x
As shown in Figure 4-36 on page 244, select your product family x3850 M2 or
x3950 M2, or the machine type to find the required information.
Chapter 5. Installation
This chapter describes the steps involved in installing and configuring supported
operating systems on the IBM x3850 M2 and x3950 M2. Firmware and BIOS
updates, BIOS settings and operating system support are also discussed as
important preliminary tasks prior to beginning the installation of operating
systems on the server.
Information about installing all hardware components for the x3850 M2 or the
x3950 M2, both as a single server and as a multinode complex, is in Chapter 3,
“Hardware configuration” on page 89 and Chapter 4, “Multinode hardware
configurations” on page 195.
If you have not already installed an operating system on your server, you can
download the non-OS specific firmware update utilities for the components listed
previously. Otherwise, there are also OS specific firmware update utilities which
you can use to update your server firmware from within the operating system.
To prepare the x3950 M2 servers in our labs for operating system installation, we
used the following firmware update sequence:
1. Remote Supervisor Adapter II firmware (RSA II requires a firmware update
specific to system x3850 M2 or x3950 M2.)
2. FPGA firmware
3. BMC firmware
The next sections list the server components that require firmware updates.
Figure 5-1 Remote Supervisor Adapter II firmware. separately on each x3950 M2 node
FPGA firmware
This update is in the form of a bootable diskette or ISO image. If you have the
RSA II configured and connected to your network, you can use the Remote
Console feature to mount a remote diskette IMG file or ISO file, and boot the
x3850 M2 or x3950 M2 from the mounted IMG or ISO file.
At the completion of the FPGA update on the primary node, you will be prompted
to press Enter to power off all nodes in the same partition. See Example 5-1 on
MNFPGA.EXE v2.60
------------------------------------
| Node 1 0x604E / 0x604F |
------------------------------------
*************************************************************************
* *
* DC Power Cycle IS required to Reload the FPGAs *
* *
* >>>>> Remove the diskette <<<<< *
* *
* Press the [Enter] key to automatically Power off *
* the Athena System and power back ON within 38 seconds *
* >>> FPGAs will reload during the Power off phase <<< *
*************************************************************************
On x3950 M2 multinode complex, if you perform the update from the primary
node, then all other nodes are automatically updated as shown in Example 5-2.
firmware image
IPMI Version= 2.0 2.0
major Revision= 2 2
minor Revision = 32 32
manufacturer ID = 2 2
product ID= 77 77
build Name= A3BT31B A3BT31B
Firmware and the image have the same version and build name.
Start programming…
Writing to Address: 0x0007FF80......OK.
Download to Flash OK.
BMC initialization…OK.
BMC Firmware and SDRs updated successfully!!!
Do you want to clear the SEL (y or n)?y
firmware image
IPMI Version= 2.0 2.0
Firmware and the image have the same version and build name.
Start programming…
Writing to Address: 0x0007FF80……OK.
Download to Flash OK.
BMC initialization…OK.
BMC Firmware and SDRs updated successfully!!!
Please remove the disk from the drive and restart the system
D :\>
System BIOS
In a x3950 M2 multinode partition, you can update the BIOS of all nodes from the
primary node. You are presented with a menu from which you can select:
Update the BIOS on the current system.
Update all nodes in a multinode x3950 M2 partition from the primary node.
The BIOS update process involves a sequence of erasing the current flash
image, updating the new flash image, requesting confirmation of the serial
number (SN) and model type for each node in consecutive order (from first node
to last node).
On x3950 M2 multinode complex, if you perform the update from the primary
node, then all other nodes are automatically updated as shown in Example 5-3.
Additional devices
For additional components with upgradable firmware, we recommend you apply
the latest versions of firmware on each. This includes:
LSI1078 Integrated SAS/SATA RAID controller
ServeRAID MR10k SAS/SATA RAID controller
ServeRAID MR10M SAS/SATA RAID controller
Network adapters
Fibre Channel HBAs
iSCSI HBAs
SAS HBAs
5. In the System Summary window (Figure 5-7), check that all memory
installed is visible to the server, as expected.
6. Check that the memory cards installed in the server are detected as shown in
the Memory Settings window (Figure 5-8), which also shows memory array,
scrub on every boot, and the scrub rate at run time. (To open the Memory
Settings window, select Advanced Setup in the Configuration/Setup Utility
window, shown in Figure 5-2 on page 254.)
IBM ServerProven program also extends to testing and publishing the IBM
System x options (for example, network adapters, storage adapters, storage
expansions, storage subsystems and other common options.
In addition, we also recommend that you check the particular operating system
vendor’s hardware compatibility list (HCL). Operating system vendors typically
publish and update their HCL as new models of servers are introduced by the
server hardware vendors. The commonly referenced Web pages of the operating
system vendors are:
VMware
http://www.vmware.com/pdf/vi35_systems_guide.pdf
http://www.vmware.com/pdf/vi3_systems_guide.pdf
http://www.vmware.com/pdf/vi35_io_guide.pdf
http://www.vmware.com/pdf/vi3_io_guide.pdf
Microsoft
http://www.windowsservercatalog.com/
http://www.windowsservercatalog.com/item.aspx?idItem=dbf1ed79-c158-c
428-e19d-5b4144c9d5cd
http://www.microsoft.com/whdc/hcl
Red Hat
https://hardware.redhat.com/
Novell SUSE Linux
http://developer.novell.com/yessearch/Search.jsp
Solaris™
http://www.sun.com/bigadmin/hcl/
http://www.sun.com/bigadmin/hcl/data/systems/details/3406.html
At the time of writing this book, the VMware operating systems are supported on
x3850 M2 and x3950 M2 as indicated in Table 5-1. The shaded table cells
indicate Yes, they are supported.
VMware ESX 3.5 Yes Yes with Yes with No Yes with
patch Update 1 Update 2
ESX350-200
802301-BGa
At the time of writing this book, the Windows operating systems are supported on
x3850 M2 and x3950 M2 as indicated in Table 5-2 on page 260. The shaded
table cells indicate Yes, they are supported.
Microsoft Windows Server 2003 Yes Yes, with Yes, with Planned Planned
R2 x64 Datacenter Edition SP2 or SP2 or
Unlimited Virtualization later later
Microsoft Windows Server 2003 Yes Yes, with Yes, with Planned Planned
R2 x64 Datacenter Edition SP2 or SP2 or
Unlimited Virtualization with High later later
Availability Program
Table 5-3 Red Hat Enterprise Linux OS support (current and planned)
Operating System x3850 M2 x3950 M2 x3950 M2 x3950 M2 x3950 M2
1-node 2-node 3-node 4-node
At the time of writing this book, the SUSE Linux operating systems are supported
on x3850 M2 and x3950 M2 as indicated in Table 5-4 on page 263. The shaded
table cells indicate Yes, they are supported.
At the time of writing this book, only Solaris 10 was supported on x3850 M2 and
x3950 M2 as indicated in Table 5-5. The shaded table cells indicate Yes, they are
supported.
Prerequisites
Before you configure ESXi:
Ensure you have local or remote (through RSA II Remote Control) video and
keyboard console access to the server with the IBM ESXi flash drive installed.
Download the following three VMware ESXi guides:
– Getting Started with ESX Server 3i Installable
http://www.vmware.com/pdf/vi3_35/esx_3i_i/r35/vi3_35_25_3i_i_get_
start.pdf
– ESX Server 3i Embedded Setup Guide
http://www.vmware.com/pdf/vi3_35/esx_3i_e/r35/vi3_35_25_3i_setup.
pdf
Figure 5-11 Example of x3850 M2 BIOS Start Options with USB Disk enabled
4. Exit the BIOS utility and save you settings when prompted.
Figure 5-13 ESXi starting to boot from IBM customized USB Flash Disk with IBM
Customized ESXi image
Figure 5-14 The x3850 M2 is successfully booting the IBM Customized ESXi on the
USB Flash Disk, and detecting processor sockets and total memory installed
Figure 5-15 The Customize System menu when you press F2 at the DCUI ESXi
interface
4. Set the host name (for example DNS FQDN) and DNS servers IP Addresses
by selecting Configure Management Network. See Figure 5-16 on
page 270. You might want to set the ESXi host management IP Address (this
interface should be separate to the interfaces you would set in VI Client for
Virtual Machine traffic) to static rather than DHCP, depending on your
management network policies.
7. Test that the network interface settings are correctly configured to ensure that
you can connect to the ESXi host for further configuration using VMware VI
Client or VirtualCenter. By default, this test tries to ping your Default Gateway,
DNS server IP addresses and performs DNS resolution of your ESXi host
name. See Figure 5-19.
Figure 5-21 The VI Client Configuration tab for ESXi host displaying Health Status basic alerts for the
ESXi’s server major subsystems
10.In the Hardware panel on the left of the Configurator tab in VI Client, check
that ESXi can successfully detect each processor’s sockets and cores, total
memory installed, network adapters installed, and storage adapters installed.
Figure 5-22 on page 275 shows the storage adapters detected.
11.ESXi should be able to detect all internal hard drives in the x3850 M2 and any
external hard drives connected through the x3850 M2’s external SAS
connector to EXP3000 enclosures. To confirm this:
– Check Storage sensors in the Hardware panel of the Configurator tab,
Hardware pane, under Health Status.
– Check the ESXi Web Interface accessible by using the address format:
https://<IP Address of ESXi host management interface>
Click Browse datastores in this host’s inventory link as shown in
Figure 5-24 on page 277.
12.ESXi has an ESXi Edition license (serial number) with limited hypervisor
management features. See Figure 5-24 on page 277. Perform one of the
following actions:
– If you intend to use VirtualCenter to manage all your ESXi hosts under
Licensed Features → License → Source Edit, configure the ESXi host’s
license source to point to the IP address of your designated License
Server (this is typically configured before or during the installation of
VMware VirtualCenter Server).
– If you use host-based license files, configure the ESXi host’s license
source to point to a license file.
13.If you intend to use VirtualCenter to manage all your ESXi hosts, we
recommend that you exit the VI Client sessions that are directly connected to
the individual ESXi hosts and connect to your VirtualCenter server using
VI Client (by using your VirtualCenter IP address, add the appropriate ESXi
hosts (using DNS registered host names and ESXi host root password)) to
VirtualCenter and make all configuration changes to ESXi hosts, HA/DRS
Clusters through VirtualCenter.
14.If you intend to use VMware VirtualCenter 2.5 Update 1 to manage ESXi
hosts in a High Availability (HA) Cluster, you have to enable swap space on
the ESXi hosts before adding them to an HA Cluster.
To enable Swap:
a. In the VirtualCenter Server, select the ESXi host.
b. In the Configuration tab page, click Advanced Settings.
c. Choose ScratchConfig. See Figure 5-25 on page 278.
See the following knowledge base articles from VMware regarding ESXi:
“Limited configurations are supported for VMware HA and ESX Server 3i
hosts” (KB 1004656)
http://kb.vmware.com/kb/1004656
“ESXi 3 Hosts without swap enabled cannot be added to a VMware High
Availability Cluster” (KB 1004177)
http://kb.vmware.com/kb/1004177
Detailed VMware ESX configuration guidance is beyond the scope of this book.
See the VMware ESXi setup and configuration guidelines listed in “Prerequisites”
on page 264.
4. In the VMware VMvisor Boot Menu, select the ThinESX Installer and press
Enter to boot the installer.
5. The ESX Server runs through its boot process until the Welcome window
opens, as shown in Figure 5-27.
9. When the installer completes the operation, the Installation Complete window
opens.
10.Remove the CD and reboot the host.
11.After the host reboots, the direct console window opens and allows you to set
up your ESX Server host configuration, as described in “Boot ESXi hypervisor
and customize settings” on page 266.
For detailed information about installing ESX Server 3i, see the ESX Server 3i
Embedded Setup Guide:
http://www.vmware.com/pdf/vi3_35/esx_3i_e/r35/vi3_35_25_3i_setup.pdf
Prerequisites
Before you install and configure ESX 3.5 Update 1:
Configure any RAID arrays using the MegaRAID Manager tool.
Figure 5-29 Booting from VMware ESX 3.5 Update 1 from CD or ISO file
Figure 5-31 Choosing the Partition Layout for ESX 3.5 Update 1.
The default boot loader setting from VMware ESX 3.5 Update 1 installer
configures the boot loader to be installed on the MBR of the disk, which you
selected earlier, on which to install VMware ESX 3.5 Update 1.
7. Network Configuration enables you to select the available Network Adapters
to use for the default vSwitch and vSwif0 interface for shared Virtual Machine
and Service Console access. You may modify the physical Network Adapter
that is configured for the vSwitch and the vSwif interface.
8. Select the time zone as appropriate to your location.
9. Under Account Configuration, you may configure the Root Password and add
new users to the ESX 3.5 Udpate 1 system.
10.In the summary window of all the selections you have made during the
installation process, confirm the selections and click Next to begin the
installation.
13.Press Alt+F1 to get the service console logon prompt (Figure 5-33); to return
to the startup window press Alt+F11.
14.To logon to the ESX 3.5 Service Console, type root and the password you set
earlier.
Tip: If you are using ServerGuide™ to install Windows, you do not have to
obtain these drivers separately.
Figure 5-35 Windows Setup detects the LSI MegaRAID SAS RAID Controller Driver
Prerequisites
Before you install and configure Windows Server 2008:
Configure any RAID arrays using the MegaRAID Manager tool.
Ensure you have local or remote (through RSA II remote control) video and
keyboard console access to the server you will be installing onto.
Download the appropriate Windows Server 2008 installation instructions:
– Installing Windows Server 2008 (32-bit) on x3950 M2 and x3850 M2:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5074895
– Installing Windows Server 2008 x64 on x3950 M2 and x3850 M2:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5074896
Figure 5-37 Selecting the disk on which to install Windows Server 2008
Prerequisites
Before you install and configure Red Hat Enterprise Linux 5 Update 1:
Configure any RAID arrays using the MegaRAID Manager tool.
Ensure you have local or remote (through RSA II Remote Control) video and
keyboard console access to the server you will be installing onto.
Download the RHEL 5 Update 1 installation guide:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5074155
If you are planning to boot from the internal SAS drives or remote storage (for
example SAN LUN) and will be installing using the RHEL5 Update 1
installation CD/DVD-ROM, you do not have to insert a driver diskette for SAS
controller drivers for local SAS drives, or host bus adapter drivers for remote
Figure 5-38 Red Hat Enterprise Linux 5 Update 1 select partition window
Post-Installation Information
If you are using a 32-bit versions of Red Hat Enterprise Linux 5, you must install
the kernel-PAE (Physical Address Extension) kernel to detect more than 4 GB of
memory.
Prerequisites
Before you install and configure SUSE Linux Enterprise Server 10 SP1:
Configure any RAID arrays using the MegaRAID Manager tool.
Ensure you have local or remote (through RSA II Remote Control) video and
keyboard console access to the server onto which you will be installing.
Download the following SLES 10 SP1 installation guide:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073088
If you plan to boot from the internal SAS drives or remote storage (for
example SAN LUN) and will be installing using the SLES10 Service Pack 1
installation CD/DVD-ROM, you do not have to insert a driver diskette for SAS
4. Continue with the installation process. Refer to the installation instructions for
additional information.
Post-Installation
If you are using a 32-bit version of SUSE Enterprise Linux 10, then install the
kernel-PAE kernel to access more than 4 GB of memory.
Chapter 6. Management
This section explains system management capabilities of the x3850 M2 and
x3950 M2 servers. It addresses the various subsystems, which are implemented
to help you manage and service your servers, and also describes embedded
features to help you reduce energy costs.
The BMC stores information related to up to 512 events. After 512 events have
been stored, the log must be cleared before further events are recorded. A
first-in-first-out algorithm is not used here. The time it takes to fill the BMC log
area depends of the kind of events that have occurred.
Tip: We recommend you save the BMC log before clearing it, in case the
service history is required. The logged events in the BMC can be transferred
in-band to the Remote Supervisor Adapter II (RSA II) or to the operating
system through a driver.
The BMC communicates with the RSA II through the Intelligent Platform
Management Bus (IPMB) and controls the components about a whole set of
standardized and system-specific sensors. The components communicate with
the BMC through an embedded Inter-Integrated Circuit (I²C) bus interface.
The BMC protects the server and powers the server off or prevents DC power-on
in any of the following circumstances:
At temperature of out-of-specification
Invalid or wrong component installation
Power voltage faults
The BMC and the x3850 M2 and x3950 M2 baseboard management firmware
enable the following features:
Environmental monitoring for:
– Fans
– Temperature
– Power supply
– Disk drives
– Processor status
– NMI detection
– Cable, card, component presence
– Voltages and Power Good settings for battery and system components
System LED Control (power, HDD activity, error, and others)
Fan speed control
Power, and reset control
Interface to subassemblies to provide the vital product data (VPD), which
contains information about components and the system, such as:
– Field replacement unit (FRU) information
– System firmware code levels like BIOS or BMC
Non-maskable interrupt (NMI), system management interrupt (SMI)
generation
Button handling (Power, Reset, Locate, Remind, and others)
Serial redirect
Serial over LAN
Machine check error capture.
Scalability partition management for multinode support
Platform event filtering (PEF) and alert policies
Local and remote update of BMC, PFGA firmware, and BIOS
The switch port should allow different port speeds and negotiation modes
to guarantee the Broadcom network interface at 1 Gbps and the BMC at
the 10/100Mb/s rate is still working after the operating system is loaded.
You can access the BMC System Event Log (SEL) through the menu shown in
Figure 6-1 on page 303, or by using tools such as OSA SMBridge, SvcCon,
Dynamic System Analysis (DSA), or IBM Director.
An example event log is shown in Figure 6-3. You can select Get Next Entry and
Get Previous Entry to page through the events.
The total number of events recorded is listed in the Entry Number field.
Tip: You might notice that several events have a time stamp of 1970. These
events are defined in the IPMI specification at Intel, and do not include time
stamp information. It fills the respective bytes in the event raw data with empty
values, which are interpreted by the SEL as a date in 1970.
The BMC makes the events available to the installed RSA II. The RSA II also
maintains a separate log, which you can view in the RSA II. Both the RSA II and
BMC-based events are listed. However, the reverse is not true: you cannot view
the events recorded by the RSA II in the BMC event log.
Figure 6-4 BIOS: BMC User Account Settings menu and individual user settings
Select the UserID you want to enable or disable and make the required settings
for username, password, and the privileges level.
The IBM Redbooks technote, Enabling Serial Over LAN for a Remote Windows
Text Console using OSA SMBridge, TIPS0551 also provides helpful information:
http://www.redbooks.ibm.com/abstracts/tips0551.html
OSAa IPMI device Required for in-band communication with IBM Director and
driver BMC system utilities
OSA IPMI library This is the OSA IPMI mapping layer. Includes the BMC
Mapping Layer, which maps the dot.commands to IPMI
commands. Required for in-band communication with IBM
Director.
Microsoft IPMI Microsoft IPMI driver interface, support for new IBM update
tools in Windows OS
The device drivers must be installed in a specific order or the installation can fail:
1. Either of the following drivers can be installed first:
– OSA IPMI device driver, then OSA IPMI library (mapping layer)
– Microsoft IPMI driver (any third-party driver must be uninstalled first)
2. IBM IPMI ASR service
Note: The IBM System x Web page, does not have OSA IPMI driver and layer
software available for Linux, Novell, and Windows operating systems to
download. For instructions, see the following sections.
You may use both drivers, but the OSA IPMI and the Microsoft IPMI drivers
cannot coexist. If the OSA driver is already installed, you must uninstall it before
installing the Microsoft IPMI driver.
Notes:
On multinode x3950 M2 configurations, the Microsoft IPMI device driver is
the only supported IPMI driver. If you have the OSA IPMI device driver
installed you must remove that and replace it with the Microsoft IPMI
device driver. The Microsoft IPMI driver is not installed in Windows Server
2003 by default. See the readme file in the driver package for more details.
The Microsoft IPMI device driver is required for UXSP, wFlash (embedded
in any IBM update packages for Windows), and online Dynamic System
Analysis (DSA) software.
3. Click Next to continue the installation. When it completes, you are prompted
to reboot the server, the installer does not do this automatically.
4. After the reboot open Device Manager again. Enable Show hidden devices
in the submenu View. The ISMP device driver is listed in the section System
devices as shown in Figure 6-7.
Note: If the Microsoft IPMI driver is installed, you do not have to install the
OSA IPMI mapping layer.
Note: The OSA IPMI driver should not be removed using Device Manager.
Doing so only partially removes the driver, which prevents the installation of a
new driver version or re-installation of the previously installed version. If an
attempt has already been made to remove the driver using Device Manager,
then follow the driver removal instructions listed in this section and reboot the
system.
2. Select Management and Monitoring Tools and click Details. The window
shown in Figure 6-10 on page 313 opens.
3. Add a check mark to Hardware Management.
4. Click OK.
Windows reminds you that any third-party IPMI driver must be uninstalled.
See Figure 6-11. Ensure that the OSA IBM driver is uninstalled, as described
in “OSA IPMI driver uninstall in Windows Server 2003” on page 311.
5. Click OK.
6. Click Next.
7. Click Finish.
The libraries for OpenIPMI are part of the most recent versions of Linux operating
systems. As a result, IBM does not supply these drivers.
Note: SUSE Linux Enterprise Server 10 (and later) and Red Hat Enterprise
Linux 5 (and later) are shipped with the OpenIPMI driver natively.
For older operating systems SUSE Linux Enterprise Server 9 (and earlier) and
Red Hat Enterprise Linux 4 (and earlier), see the OpenIPMI Web site:
http://openipmi.sourceforge.net/
Note: Problems have been reported with versions of ipmitool prior to v1.8.9.
See the following location for details:
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5069538
To build ipmitool:
Note: The installation must be run with root permissions to overlay the existing
ipmitool utility in /usr/local/bin.
The adapter includes two components: the main board has an embedded video
chip ATI RN50 and the networking interface; and the daughter card is the RSA II
adapter that is connected by a separate internal cable to the Intelligent Platform
Management Bus (IPMB) on the Serial IO/PCI-X board.
Base card
providing
power, video
and networking
Service processor
Ethernet port for
daughter card
management
The RSA II communicates with the BMC and periodically polls for new events.
The RSA II catches events of the BMC and translates them to more user-friendly
event information. The event information is then stored in an assigned space in
the NVRAM of the RSA II.
The card contains a real-time clock (RTC) timer chip, however the RSA II timer
can be synchronized with the system BIOS time or by a Network Time Protocol
(NTP) server.
The RSA II LAN settings are available in the server BIOS. To change the
IP settings in the server BIOS:
1. Boot or reset the server and press F1 when prompted to enter Setup.
2. Select the Advanced Setup → RSA II Settings. The RSA II menu opens, as
shown in Figure 6-14 on page 320.
Tip: You may also modify these settings by connecting to the service
processor port using a cross-over cable and opening a Web browser to the
card’s default IP address: 192.168.70.125
3. Set your static IP settings. You may also set up the RSA IP for each DHCP
server. You can change host name in the Web interface. Enter the following IP
settings in the DHCP Control field of the RSA II Settings menu in BIOS:
– Try DHCP then Use Static IP (default)
– DHCP Enabled
– Use Static IP
Configure the RSA II LAN interface to fit in your LAN segment. Use the right
and left arrow keys for selection.
4. In the OS USB Selection option, select either Other OS (for Windows) or
select Linux OS. Use the right and left arrow keys to make the selection.
The purpose of this selection is to prevent a known problem with Linux and its
generic human interface device (HID) driver. Linux cannot establish USB
communication with the RSA II using the generic HID (which Windows uses).
By selecting Linux OS here, it makes the RSA II appear as an OEM HID
instead of generic HID, which then functions properly.
5. Select Save the Values and Reboot RSA II, and then press Enter.
6. Exit the utility.
Tip: This process can take up to 20 seconds before you can ping the new
IP address of the RSA II adapter.
You can view the assigned address from the BIOS in the Advanced Settings
menu.
Note: In the Web interface, you will have to change the login credentials.
The home window of the Web interface opens. See Figure 6-17 on page 324.
Using remote media enables USB support after the server is powered up and the
RSA II is initialized. During installation of the operating system or in the operating
system support, USB is required. The correct setting for earlier USB support has
can be done in the RSA II Settings menu, shown in Figure 6-14 on page 320.
Tip: You can mount more than one remote drive concurrently. This means you
may add a CD-ROM and a disk drive to your remote managed server, or use
ISO and diskette image files.
The preferences link allows you to specify up to twelve key-stroke sequences and
enable mouse synchronization (that is, ensure the mouse pointer on the remote
system precisely follows the local mouse pointer).
The Video Speed selector, shown in Figure 6-20 is used to limit the bandwidth
that is devoted to the Remote Console display on your computer. Reducing the
Video Speed can improve the rate at which the Remote Console display is
refreshed by limiting the video data that must be displayed.
You may reduce, or even stop, video data to allow more bandwidth for Remote
Disk. Move the slider left or right until you find the bandwidth that achieves the
best results. The changes do not require a restart of RSA II or the server.
Check the installed firmware in the RSA II Web interface. Select the task
Monitors → Vital Product Data and scroll down to ASM VPD, as shown in
Figure 6-21.
After you install the operating system, also install the driver or service for the
RSA II SlimLine adapter.
To determine if the installation was successful, check the services for the IBM
Remote Supervisor Adapter II by selecting Start → All Programs →
Administrative Tools → Services. Scroll to the service IBM RSAII and verify
that the Status indicates Started (Figure 6-23).
Note: If you have not already done so, change the setting OS USB Selection
to Other OS in the system BIOS, as shown in Figure 6-14 on page 320.
The command returns the following two libusb entries (if your version numbers
are different, that is okay):
libusb-0.1.6-3
libusb-devel-0.1.6-3
Review the appropriate readme file of the RPM Package for prerequisites and
installation steps.
To start the daemon manually, use the command ibmspup. To stop the daemon,
enter ibmspdown.
The command returns the following two libusb entries (if your version numbers
are different, that is okay):
libusb-0.1.6-3
libusb-devel-0.1.6-3
If the command does not return these two libusb entries, install them. Both libusb
RPMs are in the /VMware/RPMS/ subdirectory on your VMware CD.
Other ports are fixed and cannot be changed, as listed in Table 6-4.
You may save this list in ASCII text format by clicking the Save Log as Text File
button on the bottom of the event log table. After saving it, you may clear the
event log by clicking the button Clear Log.
You may filter the events at severity (Error, Warning, Info), source, and date. The
events are displayed in different colors.
The IBM Systems Software Information center has additional support and
guidance for using the IBM Director:
http://publib.boulder.ibm.com/infocenter/eserver/v1r2/index.jsp?topic=/
diricinfo_all/diricinfoparent.html
Systems that are not IBM systems can also be monitored with metering devices,
such as PDU+ and sensors. Active Energy Manager is part of a larger
energy-management implementation that includes hardware and firmware
components.
The Active Energy Manager is also available in a stand-alone version, which runs
on top of Embedded Director.
To download the extension to IBM Director and the stand-alone version, go to:
http://www.ibm.com/systems/management/director
AEM Server also communicates with the Director Server to provide event filtering
and event actions that support IBM Director event action plans, and
communicates with Director Console to display status and to allow the user to
perform operations. You must install AEM Server on the IBM Director
management server. Power data is collected only while the Active Energy
Manager Server is running. When you install AEM, the server starts running. It
runs when the IBM Director Server is running. By default, it collects data on the
BladeCenter chassis and rack-mounted servers every minute, but the collection
interval can be configured on the Manage Trend Data window in the a graphical
user interface (GUI).
Managed systems
Active Energy Manager can monitor power consumption for selected rack
servers, BladeCenter chassis and blade servers, and intelligent PDUs.
You can implement each rack server, or modular system, which enables power
capping features to provide the power measurement information to the AEM.
Older server or expansion units, that do not have this feature can be
implemented by using iPDUs. See 6.8, “Power Distribution Units (PDU)” on
page 357. The iPDUs monitor the power usage on their outlets and provide the
captured information by the integrated SNMP out-of-band support.
The power capping feature in the BIOS of your x3850 M2 and x3950 M2, as
shown in Figure 6-24 on page 338, is enabled by default. This feature is available
for single node x3950 M2 systems. For details see section 3.1.3, “Processor
(CPU) configuration options” on page 99.
A new Active Energy Manager task icon is added in the IBM Director
Management console after installation of the AEM extension (see Figure 6-25).
You use the Active Energy Manager task on the IBM Director Console menu to
launch the Active Energy Manager GUI. Use the GUI to view and monitor power
consumption on various rack-mounted servers, BladeCenter chassis and iPDUs
in the IBM Director environment (see Figure 6-26 on page 339).
When you install the AEM extension in your IBM Director environment, its task is
added to IBM Director Console. Start the Active Energy Manager task by
dragging and dropping the task icon onto any of the following targets:
If the target is a blade server, the display consists of the BladeCenter chassis
containing the blade with the targeted blade preselected.
If the target is a BladeCenter chassis, the display consists of the chassis with
the chassis itself preselected.
If the target is a rack-mounted server like your x3850 M2 or x3950 M2, the
display consists of the server with the server itself preselected.
If the target is a group of rack-mounted servers or BladeCenter chassis, or
both, the display consists of all chassis or rack-mounted servers in the group
with the first chassis or Start Active Energy Manager for all managed
systems.
If the target is an iPDU, the current date for the PDU and the capacity and
usage for the load groups is displayed with the iPDU itself preselected.
When you start AEM for only a single managed system or for a group of
managed systems, the Active Energy Manager window opens and displays a
tree of only the selected objects. See Figure 6-27 on page 340. The left panel in
Active Energy Manager contains only the selected managed systems, those
chassis that contain the systems that have also been selected, and the managed
systems in a selected group.
When you select a system, Active Energy Manager displays the entire tree for all
managed systems. The advantage of starting Active Energy Manager in this
manner is that you see only the subset of managed systems that you are
interested in. The disadvantage is that you can manage only those managed
systems that were selected when you started Active Energy Manager. If you have
to manage other managed systems at a later time, you must start another Active
Energy Manager task for those additional managed systems.
Monitoring functions
These functions provide the general information about the power consumption of
the data center and possible anomalies. You control what information to collect
about power and temperatures, how often to collect it, and how long to store the
data. You can view graphs or tables of data trends, and also export the tabular
data as a spreadsheet, XML, or HTML formats.
Power trending
The Energy Scale architecture provides continuous power usage data collection.
The power usage data can be displayed from Active Energy Manager. Data
administrators can use the information to project the power consumption of the
data center at various times of the day, week, or month and use that information
to identify anomalies, manage loads when electrical demands or costs are high
and, if the system supports power capping, determine appropriate times and
levels for power caps (see “Power capping” on page 346).
In Figure 6-28 on page 342, the top chart shows an example of the power
trending information, set at custom intervals. Figure 6-27 on page 340 shows an
example of power trending in the last hour. Various controls on the panel help
define the interval and the data to display.
The trending graph can also display events that are monitored by the Active
Energy Manager. Details about each event recorded can be displayed by
dragging the mouse pointer over the event symbol.
The types of events monitored by the Active Energy Manager include power
component failures, offline and online actions, power management actions, and
power and thermal critical warnings.
Thermal trending
The thermal sensors are primarily based on the digital thermal sensors available
on the systems. All of the logical sensors are the result of firmware running on
the thermal and power management device (TPMD) using the raw data provided
by the hardware, and converting it into values that are then fed into the control
loops.
iPDU support
Section 6.8, “Power Distribution Units (PDU)” on page 357 briefly describes on
how iPDUs work with the AEM. An iPDU can record similar information to other
power monitored hardware. The power trending data in Figure 6-28 on page 342
is from an iPDU.
You can also display detailed information about each load group in the iPDU:
Name assigned to the load group
Average output watts which is the DC power currently being consumed as
reported by the power meter, or shown as two dashes (--) if power data is not
available
Minimum and maximum output watts for the load group over the last polling
interval
Amps used in relation to capacity
Figure 6-29 on page 344 shows how information for an iPDU and its load groups
is presented.
Watt-Hour Meter
The Watt-Hour Meter displays the amount of energy used for a given system or
group of systems over a specified period of time, and calculates the
corresponding cost of that energy, shown in Figure 6-30 on page 345. It gives a
visual indication of the number of watt-hours consumed by the target object or
objects over the specified time period and provides a comparison to what would
have been consumed had nameplate power been drawn over that entire period.
Management functions
The management functions are Power Saver and Power Capping. They provide
methods to reduce energy consumption by dropping the processor voltage and
frequency. The functions are available for a 90-day trial use. You must purchase a
license and install a key to use these functions beyond the trial period.
One possible use for Power Saver would be to enable it when workloads are
minimal, such as at night, and then disable it in the morning. When Power Saver
is used to reduce the peak energy consumption, it can lower the cost of all power
used. At low processor utilization, the use of Power Saver increases processor
utilization such that the workload notices no performance effect. Depending on
workload, this can reduce the processor power usage by 20-30%.
Power Saver mode is available on the x3850 M2 and x3950 M2 systems to.
Set a power cap:
– Guarantees server does not exceed a specified number of watts
– If cap is reached, processor is throttled and voltage is reduced
– Available on P6 Blades and selected System x servers and blades
Drag a function onto a system:
– For example, drag Power Saver on to a server in middle pane
– Perform a task now, or schedule it for later
Power capping
Power Capping enforces a user-specified limit on power usage. You set and
enable a power cap from the Active Energy Manager interface. In most data
centers and other installations, when a machine is installed, a certain amount of
power is allocated to it.
Generally, the amount is what is considered to be a safe value, which is often the
label power for the system. This means that a large amount of reserved, extra
power is never used. This is called the margined power. The main purpose of the
power cap is not to save power but rather to allow a data center operator the
ability to reallocate power from current systems to new systems by reducing the
margin assumed for the existing machines. Thus the basic assumption of power
capping allows an operator to add extra machines to a data center, which
previously had all the data center power allotted to its current systems.
Power capping provides the guarantee that a system does not use more power
than assigned to it by the operator.
Requirements
Integrating your x3850 M2 or x3950 M2 server as a Level-0 managed system
does not require that the drivers or the IBM Director agent be installed. Level-0
systems are managed by the embedded service processor or optional installed
RSA II adapter out-of-band through the service processor LAN interface
communication.
1
Level-1 systems with a LSI1078-based MegaRAID controller require an installed
LSI-MegaRAID-Provider, after the installation of the IBM Director agent.
LAN interface
Systems that have an embedded service processor or installed RSA II can be
managed and monitored if the communication to the LAN interface is
guaranteed.
The RSA II adapter in the x3850 M2 and x3950 M2 replaces the embedded
BMC, so the LAN interface for the RSA II must be configured only to manage it by
the IBM Director. Section 6.2.2, “RSA LAN configuration in BIOS” on page 319
discusses how to set a valid network configuration.
To access the system, right-click the system name and select Request access.
Enter the service processor login credentials to be able to manage the RSA II
adapter from IBM Director.
Supported tasks
Core services allow the following tasks to be performed:
Inventory information
Upgrading to a Level-2 by deployment of the IBM Director agent
Management of events by using of event action plans
Monitoring hardware status
Deployment of system update package
Remote management (if SSH service started)
Running of command-line scripts
Windows installation
Install the software as follows:
1. Extract the package:
unzip dir5.20.2_coreservices_windows.zip
2. Change to the extraction folder:
cd /install_files_directory/FILES
3. Copy the response file (\directory\FILES\coresvcs.rsp) to another location.
4. Open the coresvc.rsp with an ASCII editor and change the contents as
commented in this file. Save it with a new filename, such as:
responsefile.rsp
5. Select Start → Run, and then type:
\directory\FILES\dir5.20.2_coreservices_windows.exe /s /a
installationtype rsp=responsefile.rsp option
– directory is the installation files folder or cd-rom path
CD:\coresvc\agent\windows\i386\
– /s is optional to hide file extraction dialog
– installationtype is either:
• unattended: installation progress, no user intervention
• silent: hide progress
– responsefile.rsp is path and name of the created response file
– option indicates:
• waitforme (ensure complete installation)
After IBM Director detects the system, that system is listed under Groups as:
Level1: IBM Director Core Services Systems
To unlock the system, right-click the system name, select Request Access, and
provide your operating system logon credentials. See Figure 6-32. The lock
disappears and the system can now be managed from IBM Director Console.
The LSI MegaRAID Provider is supported on Level-1 systems with the following
operating systems:
Red Hat Enterprise Linux, version 4.0
Red Hat Enterprise Linux, version 5.0
SUSE Linux Enterprise Server 9 for x86
SUSE Linux Enterprise Server 10 for x86
Microsoft Windows
The IBM Director Core Services must be installed before you can install the
LSI MegaRAID Provider. Download this extension from:
http://www.ibm.com/systems/management/director/downloads.html
The use of the IBM Director agent requires installation of the service processor
driver. The x3850 M2 and x3950 M2 is shipped with the RSA II so installing the
BMC driver is not necessary.
To unlock the system, right-click the system name, select Request Access, and
provide your operating system logon credentials. See Figure 6-32 on page 351.
The Embedded hypervisor ESX 3i does not have a service console, unlike the
Installable version 3.5 has. This means an IBM Director agent cannot be installed
to manage this system as a Level-2 Agent system in your system with the
embedded hypervisor running. Instead, the Embedded hypervisor relies on CIM
and SNMP for remote management.
You do not have to possess special operating system knowledge, do not have to
maintain user accounts and passwords, and do not have to install additional
security tools, antivirus tools, or the need to backup the operating system.
Newer systems such as the x3850 M2 and x3950 M2 servers have power
supplies that regulate the power usage by active electronic parts and by the
system-adjusted firmware. As a result, the efficiency can now reach a level
above 90%. New power supplies have a power factor to reduce the ratio between
the amount of dissipated (or consumed) power and the amount of absorbed (or
returned) power.
However this is not the only approach to optimize your yearly IT environment
costs in your data center. More components such as processor cores, memory,
or I/O components for each system and rack can result in less space necessary
in the rack for your IT solution.
Another EIST feature allows you to control the power usage on a system. This
feature is known as power capping and can be enabled in the BIOS of the x3850
M2 and x3950 M2 server systems. This feature must be enabled so it works in
IBM Active Energy Manager as described in “Active Energy Manager (power
capping)” on page 100. IBM added this feature as a key component of the IBM
Cool Blue™ portfolio within Project Big Green.
Project Big Green is an IBM initiative that targets corporate data centers where
energy constraints and costs can limit their ability to grow.
We discuss these in detail in 6.8, “Power Distribution Units (PDU)” on page 357.
6.7.3 Virtualization
Intelligent applications are developed to help a system be more efficient and fully
loaded. Virtualization technologies in the various NUMA-aware operating
systems are enabled by Intel Virtualization Technology architecture in Xeon
processors, as follows:
VMware: ESX Server now in ESX 3.5 and ESXi 3.5
Red Hat Enterprise Linux 5 with Xen
SUSE Linux Enterprise Server 10 with Xen
Microsoft: Windows 2003 with Microsoft Virtualization Server
Microsoft: Windows 2008 and the Hyper V
This allows the usage of efficient multicore systems by running multiple virtual
machines in parallel on one system.
These help to quickly and simply deploy, protect, and manage your
high-availability IBM System x3850 M2 and x3950 M2 and non-server equipment
such as expansions for storage, tape backup, or non-IBM hardware.
All enterprise PDUs are designed in 1U full-rack size and can be mounted
vertically inside the rack.
The intelligent PDU+ models have the following power management features:
Monitored power draw at the breaker level
Monitoring / measurement of power data
Advanced remote monitoring capability
Detailed data-logging for statistical analysis and diagnostics
Includes an Environmental Monitoring Probe to provide both temperature and
humidity values
Integrates with IBM Active Energy Manager for consolidated rack power
monitoring
Comprehensive power management and flexible configuration through a Web
browser, NMS, Telnet, SNMP, or HyperTerminal
SNMP v1 support
71762MX 6090 Ultra Density Enterprise PDU+ 9 3 Varies by line cord, see Table 6-6
71763MU 6091 Ultra Density Enterprise PDU+ 9 3 60A/208V/3 phased fixed cable
71762NX 6050 Ultra Density Enterprise PDU 9 3 Varies by line cord, see Table 6-6
71763NU 6051 Ultra Density Enterprise PDU 9 3 60A/208V/3 phase fixed cable
39M2816 6030 DPI® C13 Enterprise PDU+ 0 12 Varies by line cord, see Table 6-6
39M2818 6080 DPI C19 Enterprise PDU+ 6 3 Varies by line cord, see Table 6-6
39Y8941 6010 DPI C13 Enterprise PDU 0 12 Varies by line cord, see Table 6-6
39Y8948 6060 DPI C19 Enterprise PDU 6 0 Varies by line cord, see Table 6-6
a. Last letter: U = available in U.S. and Canada; X = available worldwide
Second last letter: N = non-monitored (PDU);, M = monitored (PDU+)
Table 6-6 lists the available detachable line cords and the power source they
support. Only those PDUs in Table 6-5 without fixed line cords support these.
EMEA, AP, LA 40K9611 N/A IBM DPI 32A cord IEC 309 3P+N+G 250V 32A
EMEA, AP, LA 40K9612 N/A IBM DPI 32A cord IEC 309 P+N+G 250 V 32A
EMEA, AP, LA 40K9613 N/A IBM DPI 63A cord IEC 309 P+N+G 250 V 63A
US, Canada 40K9614 6500 4.3m, 30A/208V, NEMA L6-30P 208 V 30A
US, Canada 40K9615 6501 4.3m, 60A/208V, IEC 309 2P+G 208 V 60A
Australia, NZ 40K9617 N/A IBM DPI 32A cord IEC 309 P+N+G 250 V 32A
Korea 40K9618 N/A IBM DPI 30A cord IEC 309 P+N+G 250 V 30A
PDU
PDUs are engineered to split a power source to different power outlets so that
the power can be distributed to several systems. PDUs are either single-phase or
three-phase, and which one you use depends on the power sources you have to
attach it to.
Tip: We recommend you distribute the power cables to load groups that are
joined to different phases, to ensure even greater power redundancy.
Intelligent PDU
Intelligent PDUs (iPDU, with the IBM designation “PDU+”) can be remotely
managed and have an Ethernet or serial port for management. The PDU can be
accessed from a browser interface or through Telnet. Section 6.8.5, “Intelligent
PDU power management Web interface” on page 364 provides a description of
the Web interface.
The iPDU collects power and temperature data at the power outlets and can
send this information through SNMP. This data is collected once per second. You
can also view this data in table form from the iPDU Web interface or by using
tools such as IBM Active Energy Manager. They report power and thermal
trending also for devices that is plugged into their individual load groups.
Additionally, Active Energy Manager monitors how many amps are being
consumed by an iPDU overall, and how this compares to the maximum current
that the iPDU can support.
Note: IBM DPI C13+ and IBM DPI C19 PDU+ PDUs were available before the
announcement of Active Energy Manager. Any existing versions of these
iPDUs in the field must be upgraded to the November 7th, 2007 version of the
iPDU firmware so that Active Energy Manager can support them.
Figure 6-33 IBM Enterprise DPI C13 PDU+ UTG connector, part number: 39M2816
Figure 6-34 IBM Enterprise DPI C19 PDU+ UTG connector, part number: 39M2818
The outlets are protected by six branch type circuit breakers 20A. Each of the
C19 outlets on the front represents one load group, to six load groups overall.
Every one of the C13 outlets at the back of the PDU+ is shared with load group 1,
3, or 5 at the C19 outlets.
UTG inlet
C19 outlet
Operating mode DIPswitch
Figure 6-35 IBM Ultra Density Enterprise PDU+, part number: 71762MX
The bottom view in Figure 6-35 shows the back of PDU+, which has 3x
IEC-320-C13 outlets.
The outlets are protected by six branch type circuit breakers 20A. Each of the
C19 outlets on the front side represents one load group, to nine load groups
overall. Load group 1,4,7 each are shared with one C13 outlets at the rear side of
the PDU+.
The Web interface, in Figure 6-36 on page 365, shows the actual power data
measurement at the load groups, which are refreshed once per second.
The iPDU stores the measurement of power data and reports it to a graph,
shown in Figure 6-37 on page 366, and presents a chart of power consumption
tendencies.
The measured results are reported by the built-in SNMP support to the IBM
Active Energy Manager.
DSA Preboot is an NVRAM-based version of the of the DSA tool, which is used
by the Technical Support teams to collect the following information when
determining the cause of errors:
System and component level information,
Operating system driver information,
Hardware event logs of various components,
Operating system event logs
To run DSA Preboot, press F2 during POST when prompted. The system first
starts the memory test menu as shown in Figure 6-38 on page 368.
To leave the memory test window use the keyboard’s right arrow key to move to
the menu Quit and select Quit to DSA. The DSA Preboot is then started.
Commands:
gui - Enter GUI Environment.
cmd - Enter Command Line Environment.
copy - Copy DSA results to removable media.
exit - Quit the program.
Note: This will reboot the system.
help - Display this help message.
Note: The DSA Preboot updates stand-alone systems and all nodes that are
partitioned to a multinode system. See Chapter 4, “Multinode hardware
configurations” on page 195.
3. The DSA Preboot flash update menu is displayed, shown in Example 6-3 on
page 370.
Commands:
update - Update/Recover your embedded usb chip
help - Display this help message
exit - Quit program.
Note: This will reboot the syste,-
Read the instructions for and then download the *.sh for Linux operating system.
Example 6-6 DSA Preboot: flash progress of the Update process in Linux
srv-253-217:/usr/dsa # ./lflash64 -z
Node count: 1
Flashing node: 0
Waiting for device to become available:
.......
Device flashing in progress:
...........................................................................
...........................................................................
...........................................................................
..............................................................
DSA key has been flashed successfully
srv-253-217:/tmp/dsa #
cmd The command line environment interface you can use to perform the set
of diagnostics, collect the system information, and show the date in the
local text viewer. The data and results can be sent to the IBM FTP server
to diagnose the information. See “Use of the interactive menu” on
page 373.
copy Use this command to save the captured archive to any USB device.
view Displays the collected data on the local console in text viewer.
To exit viewer, type :x then press Enter.
exectest Presents a menu in which you can select a test to execute. Use
this command to execute the specific test.
getextendedresults Retrieves and displays additional diagnostic results for the last
diagnostic test that was executed.
transfer Transfers collected data and results to IBM. To transfer data and
results to IBM, a network connection is required. You see:
*******************
Attempted data upload to IBM by means of an unencrypted
channel will proceed in 10 seconds. Press any key to
cancel the attempted upload.
*******************
..........
Configuring Network Controllers.
Transferring collected data to IBM Service.
Example 6-7 DSA Preboot command line: mounted USB storage devices
>copy
1: SanDisk Cruzer Micr
2: USB 2.0 Flash Disk
Enter Number (type x to exit)
Note: The example shows two mounted USB devices. Number 1 in this
example is an attached Hypervisor key. You cannot use this device to save
any captured data.
3. Unmount the USB storage device. The data is now available for reviewing by
you and by IBM Service.
To retrieve the DSA logs that the DSA Preboot collected, enable this device
temporarily and make it show up as a drive in your operating system. Do this by
pressing Ctrl+End on your keyboard during POST when the IBM logo appears.
Although no message indicates that it is enabled, the device shows up in BIOS
and the operating system. See Figure 6-41, Figure 6-42 on page 375, and
Figure 6-43 on page 375.
If you boot the system, you can view the added USB devices in the BIOS. See
Figure 6-41.
In the Linux operating system, shown in Figure 6-43, you see the USB device
listed, which can be mounted as storage device.
Check that a partition is available so you can then use the following command:
mount /dev/sdb1 /mnt/dsa
linux-x3850m2-2:~ #
Copy the files with the extension *.xml.gz to a local disk. Provide the log files to
the IBM technical support team if requested. After you reboot or power cycle your
server this device is hidden again.
You start the Dynamic System Analysis GUI in the main menu by using the
command gui. The GUI window opens, shown in Figure 6-44.
After adding tests you select, start the tests by clicking the button Start Tests.
The Status column shows the result of this test. For details about the results,
click Get Status Details.
System Information
Use the system information window to get an overview about your system or your
multinode partition. See Figure 6-46 on page 378.
Note: If your system indicates an error and you are able to boot to the DSA
Preboot diagnostics, we recommend you review each of these entries.
Help system
The embedded Help (Figure 6-47) describes prerequisites, handling instructions,
known problems, and provides tips for using the DSA Preboot.
The publications listed in this section are considered particularly suitable for a
more detailed discussion of the topics covered in this book.
IBM Redbooks
For information about ordering these publications, see “How to get Redbooks” on
page 394. Note that some of the documents referenced here might be available
in softcopy only.
Tuning IBM System x Servers for Performance, SG24-5287
IBM eServer xSeries and BladeCenter Server Management, SG24-6495
Virtualization on the IBM System x3950 Server, SG24-7190
IBM BladeCenter Products and Technology, SG24-7523
Building an Efficient Data Center with IBM iDataPlex, REDP-4418
Enabling Serial Over LAN for a Remote Windows Text Console using OSA
SMBridge, TIPS0551
Product publications
These product publications are also relevant as further information sources:
System x3850 M2 and x3950 M2 Installation Guide
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073028
System x3850 M2 and x3950 M2 User’s Guide
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073029
System x3850 M2 and x3950 M2 Problem Determination and Service Guide
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073027
System x3850 M2 and x3950 M2 Rack Installation Instructions
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5073030
IBM ScaleXpander Option Kit Installation Instructions
http://www.ibm.com/support/docview.wss?uid=psg1MIGR-5075330
Online resources
These Web sites are also relevant as further information sources:
Intel
See the following Web addresses for more information:
Execute Disable Bit and Enterprise Security
http://www.intel.com/technology/xdbit/index.htm
Intelligent Platform Management (IPMI) Interface Specification
http://www.intel.com/design/servers/ipmi/ipmiv2_0_rev1_0_markup_2.pdf
Intel Xeon Processor 7000 Sequence
http://www.intel.com/products/processor/xeon7000/index.htm?iid=servp
roc+body_xeon7000subtitle
Server Processors
http://www.intel.com/products/server/processors/index.htm?iid=proces
s+server
Intel 64 Architecture
http://www.intel.com/technology/intel64
Red Hat
See the following Web addresses for more information:
What Every Programmer Should Know About Memory
http://people.redhat.com/drepper/cpumemory.pdf
Red Hat Enterprise Linux Server Version comparison chart
http://www.redhat.com/rhel/compare/
Red Hat Hardware Catalog
https://hardware.redhat.com/
Other
See the following Web addresses for more information:
IPMItool
http://ipmitool.sourceforge.net
OpenIPMI
http://openipmi.sourceforge.net/
TrouSerS
http://trousers.sourceforge.net/
Java Downloads for all operating systems
http://www.java.com/en/download/manual.jsp
Solaris OS: Hardware Compatibility Lists
http://www.sun.com/bigadmin/hcl/
HCL for OpenSolaris, Solaris OS
http://www.sun.com/bigadmin/hcl/data/systems/details/3406.html
Index 397
RSA II 208, 248, 327 IBM Systems Director Active Energy Manager 51
SAS disk drives 152 iDataPlex 57–58
ServeRAID 150 initial ROM size 193
system 208 Initialization Scrub Control 118, 123
flash upgrade 209 installation
four-node configuration 216 EXP3000 145
FPGA IBM Director 349
updating 248 IPMI driver 309
front panel 3 processors 93
front-side bus 37 Red Hat Enterprise Linux 293–295
ServeRAID-MR10k 132
ServeRAID-MR10M 137–138
G SUSE Linux Enterprise Server 295–297
Gigabit Ethernet controller 44
VMware ESX 281–285
Gigabit Ethernet ports 4
VMware ESXi 264–281
global hot spares 130
Windows Server 2003 285–289
Grid Computing 58
Windows Server 2008 289–293
integrated virtualization 19
H Intel 64 Technology 5, 37
hard disks 127 Intel Core Architecture 29
hardware compatibility list (HCL) 258 Intel Execute Disable Bit 108
hardware prefetcher 108 Intel SpeedStep 101
heat-sink 34 Intel Virtualization Technology 109
High Performance Clusters 58 Intel VT support 5
home node 68 Intel Xeon 7000 family 33
hot spares 130 interleaving 39
hot-add internal disks 127
memory 41, 74, 120 internal SAS cable 125
hot-swap IP Prefetcher 109
backplane 125 iPDU 357, 360
disk bays 45 Active Energy Manager support 343
fans 45 IPMI device driver 308
memory 41, 119 IPMI library 308
PCI 192 iTBBU 130
HPMA 122
HSS-IB ports 29
Hurricane 4 chipset 5, 29 K
key (ScaleXpander chip) 201, 204
Hyper-Threading Technology 36
hypervisor model 4
L
L1 cache 35, 92
I L2 cache 35, 92
iBBU 137
L3 cache 92
IBM BladeCenter 58
L4 cache 30
IBM Director 334, 346–355
latency 29, 67
Active Energy Manager 334
LEDs 3
agents 348
legacy mode 37
LSI provider 352
light path diagnostics 6, 34, 47, 93
IBM iDataPlex 58
line cords for PDUs 359
Index 399
NUMA 66 manual partitioning 227
Linux 77 power off 231
Linux performance 77 reorder 223
reset 231
standalone boot 222
O starting 228
Open Fabric Manager 58
status 224
OpenIPMI 308
Partitions Configure buttons 222
operating system installation
patrol read 130, 165
Red Hat Enterprise Linux 293–295
PCI Device Boot Priority 155, 189
SUSE Linux Enterprise Server 295–297
PCI Express subsystem 43, 188
VMware ESX 281–285
bridges 29–30
VMware ESXi 264–281
card installation order 189
Windows Server 2003 285–289
hot-plug support 192
Windows Server 2008 289–293
initial ROM size 193
operating systems 257
installation 138
Red Hat Enterprise Linux 261
PCI slot information 192
Solaris 263
ROM size 193
SUSE Linux Enterprise Server 262
scan order 188
VMware ESX 259
slots 4
VMware ESXi 259
Southbridge 188
Windows Server 2003 259
x8 slots 43
Windows Server 2008 259
PCI-X slots 60
optical drive 7
PDUs 357
optical media 34, 93
features 358
options
intelligent PDUs 360
disk drives 127
line cords 359
EXP3000 143
part numbers 359
memory 111
PFA error 165
PDUs 359
Planar SAS 189
processors 91
ports
scalability 202
Ethernet 4
order of installation, cards 189
scalability 211
orientation of processors 96
TCP/IP ports for BMC 316
OSA SMBridge 305
TCP/IP ports for RSA II 332
overview 1
positioning 53, 56–87
IBM BladeCenter 57
P iDataPlex 57
PAE switch 289, 293 POST watchdog 306
Partition Control 221 power capping 100
Partition Reorder 223 power management 356
partitions 15, 220 power off partition 231
automatic partitioning 227 power supplies 4, 7, 45
clearing all 232 PowerExecutive. See Active Energy Manager
configuration steps 225 prefetcher
configure 222 adjacent sector 107
creating 226 processor hardware 108
defined 197, 199 prerequisites for multinode 201
delete a partition 232 primary node 197
Index 401
SAS onboard controller 124 virtual disks 136
See also WebBIOS ServerProven 258
BIOS utility 154 SFF-8087 internal SAS cable 125
firmware update 148 SFF-8088 external SAS port 126
IBM Director provider 352 slots, PCI 188
scalability 199 SMART error 165
applications 79 SMBridge 305
LED 3 SMP 66
ports 29, 211 SMP expansion ports 4
scalable system 198 snoop filter 30
system 64 snoop filter lookup table 28
VMware ESX 68 socket 34
Scalability Upgrade Option 2 8, 202 Soft-NUMA 82
scale out 85 Solaris support 263
scale up 84 Southbridge 29, 60, 188
ScaleXpander key 7, 197, 204 SPEC CPU2006 65
ScaleXpander Option Kit 11, 61, 198, 202 SpeedStep 101
scan order, PCI Express 188 SQL Server 2005 80, 87
scheduling, VMware ESX 22, 68 SQL Server 2008 87
scrub control 123 hot-add CPU 84
scrubbing, memory 40 overview 83
secondary nodes 197 Resource Governor 83
serial port 4, 44 SRAT table 67
server consolidation 54 Standalone Boot 222
ServeRAID-MR10k 42, 128 standalone mode 197
See also WebBIOS standalone system 196
battery backup unit 130 standard memory 6
data scrubbing 130 start a partition 228
features 130 storage expansion unit 142
firmware update 150 Streams benchmark 109
hot spares 130 stripe sizes 128
IBM Director provider 352 Supply Chain Management 55
installation 132–133 support
iTBBU 130 VMware, applications running on 26
migrating arrays 133 SUSE Linux Enterprise Server
part numbers 132 boot from internal drives 295
patrol read 130 features 79
RAID levels 129 installation 295–297
WebBIOS 158 LSI MegaRAID Provider 352
ServeRAID-MR10M 135 NUMA 77
See also WebBIOS RSA II driver 330
battery backup 137 scaling 77
block diagram 135 support 262
cache 135 system event log 304
external x4 port 135 system management features 47
firmware update 150
iBBU 137
IBM Director provider 352
T
target applications 54
installation 137–138
Index 403
WebBIOS 159 BIOS settings 286, 290
Access Policy 171 boot from internal drives 286
adapter BIOS 165 driver diskette 288
Adapter Properties 162 hot-add memory 74, 120
alarm 165 installation 285, 287–289
background initialization 164 IPMI drivers 309
battery backup 163 LSI MegaRAID Provider 352
cache flush interval 166 NUMA aware 73
Check Consistency 170 PAE switch 289
cluster mode 164 power options 103
coercion mode 165 prerequisites 286
configuration wizard 179 RSA II driver 329
consistency check 165, 167 scalability 73
data scrubbing 165 support 259
deleting a virtual disk 170 Windows Server 2008
Disable BGI 172 hot-add memory 120
Disk Cache Policy 172 installation 289–293
drive rebuild 175 IPMI driver 314
drive states 168 logical drives 292
Event Information panel 187 LSI MegaRAID Provider 352
factory defaults 164 PAE switch 293
Fast Initialize 170 power management 104
firmware information 175 prerequisites 289
functions 161 RSA II driver 329
IO Policy 172 support 259
main window 160 Windows Server Resource Manager 83
Patrol Read 165
PFA error 165
policies 171
X
X3 technology, comparison 31, 62
RAID level migration 173
x3850 M2
Read Policy 171
compared with x3850 59
rebuild a drive 175
features 2
rebuild rate 164
x3950 M2
reconstruction rate 165
can be x3850 M2 with kit 198
schedule consistency check 167
compared with x3950 62
ServeRAID-MR10k 158
features 2
Slow Initialize 170
X4. See eX4
SMART error 165
XAPIC 105
spinup delay 166
X-Architecture 1
spinup drive count 166
XceL4v cache 30
stop on error 166
XD, Execute Disable Bit 108
Stripe Size 185
Xeon 7000 family 33
Unconfigured good 183
virtual disk properties 168–169
virtual drive states 168
wizard 179
WHEA support 7
Windows Datacenter Server 15–19
Windows Server 2003
Understand the IBM The x3950 M2 server is the System x flagship server and
System x3950 M2 implements the fourth generation of the IBM X-Architecture. INTERNATIONAL
and IBM x3850 M2 It delivers innovation with enhanced reliability and availability TECHNICAL
features to enable optimal performance for databases, SUPPORT
Learn the technical enterprise applications, and virtualized environments. ORGANIZATION
details of these The x3950 M2 features make the server ideal for handling
high-performance complex, business-critical On Demand Business applications
servers such as database serving, business intelligence, transaction BUILDING TECHNICAL
processing, enterprise resource planning, collaboration INFORMATION BASED ON
See how to applications, and server consolidation. PRACTICAL EXPERIENCE
configure, install, Up to four x3950 M2 servers can be connected to form a
manage multinode single-system image comprising of up to 16 six-core IBM Redbooks are developed by
complexes processors, up to 1 TB of high speed memory and support for the IBM International Technical
up to 28 PCI Express adapters. The capacity gives you the Support Organization. Experts
ultimate in processing power, ideally suited for very large from IBM, Customers and
relational databases. Partners from around the world
create timely technical
This IBM Redbooks publication describes the technical information based on realistic
details of the x3950 M2 scalable server as well as the x3850 scenarios. Specific
recommendations are provided
M2 server. We explain the configuration options, how to help you implement IT
2-node, 3-node and 4-node complexes are cabled and solutions more effectively in
implemented, how to install key server operating systems, your environment.
and the management tools available to systems
administrators.
For more information:
ibm.com/redbooks