QV1203 StudentGuide
QV1203 StudentGuide
QV1203 StudentGuide
cover
Student Notebook
ERC 3.1
The information contained in this document has not been submitted to any formal IBM test and is distributed on an “as is” basis without
any warranty either express or implied. The use of this information or the implementation of any of these techniques is a customer
responsibility and depends on the customer’s ability to evaluate and integrate them into the customer’s operational environment. While
each item may have been reviewed by IBM for accuracy in a specific situation, there is no guarantee that the same or similar results will
result elsewhere. Customers attempting to adapt these techniques to their own environments do so at their own risk.
© Copyright International Business Machines Corporation 2007, 2009. All rights reserved.
This document may not be reproduced in whole or in part without the prior written permission of IBM.
Note to U.S. Government Users — Documentation related to restricted rights — Use, duplication or disclosure is subject to restrictions
set forth in GSA ADP Schedule Contract with IBM Corp.
V3.1.0.1
Student Notebook
TOC Contents
Course Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Brief Description
This course is part of a PowerHA curriculum designed to prepare students to support
customers using PowerHA. This course is an introductory course designed to prepare
students to install and configure a highly available cluster using PowerHA Version 5.5 on
an IBM pSeries server running AIX V6.1 or V5.3.
Audience Description
This course is intended for AIX technical support personnel and AIX system
administrators.
Course Objectives
On completion of this course, students should be able to:
- Describe the capabilities of PowerHA for AIX
- Configure AIX networks for PowerHA
- Configure PowerHA for AIX with a two-node standby cluster with one resource
group (one active node and one standby node)
- Start and stop cluster services
- Move a resource group from one node to the other.
Course Prerequisites
Students attending this course are expected to have AIX system administration, TCP/ IP,
LVM storage and disk hardware implementation skills. These skills are addressed in the
following courses (or can be obtained through equivalent education and experience):
- AHQV011: AIX Problem Determination I: Boot Issues
- AHQV012: AIX Problem Determination II: LVM Issues
- AHQV511: TCP/IP on AIX Problem Determination I: Configuration
Course Topics:
- Introduction to PowerHA for AIX
- Configuring Networks for PowerHA
- Configuring Applications for PowerHA
- Installing PowerHA
- Initial PowerHA Configuration
Course Agenda
(Agenda times are just estimates.)
(1:00) Welcome
(2:00) Unit 1 - Introduction to PowerHA for AIX
(1:00) Exercise 1 - Exploring PowerHA
(2:00) Unit 2 - Configuring Networks for PowerHA
(1:00) Unit 2 - Continued
(2:00) Exercise 2- Configure Networks for PowerHA
(1:00) Unit 3 - Configuring Applications for PowerHA
(1:00) Unit 4 - Installing PowerHA
(2:00) Unit 5 - Configuring PowerHA
(3:00) Exercise 3- Configure PowerHA
References
http://www-03.ibm.com/systems/p/library/hacmp_docs.html
PowerHA manuals
http://www.software.ibm.com/webapp/set2/sas/f/hacmp/home.html
Service packs for PowerHA
http://w3.tap.ibm.com/w3ki03/display/hacmp/PowerHA+for+AIX
PowerHA Wiki
Contains links to many PowerHA resources.
http://www.ibm.com/common/ssi:
IBM Sales Manual
Select country and language;
Choose the HW & SW desc (Sales Manual, RPQ); Click
Advanced Search;
Enter PowerHA in the Title: field;
Click Search at the bottom of the page
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-1
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
http://www-03.ibm.com/support/techdocs/atsmastr.nsf/Web/Flashes
Techdoc Flash web page
Search for PowerHA or HACMP to find latest on supported
hardware and other technical developments which have not yet
been added to the manuals.
http://www.ibm.com/developerworks/forums/forum.jspa?forumID=1611&start=0
PowerHA (Formerly known as HACMP) Technical Forum
A highly technical forum focussing on concepts, and concerns
with using PowerHA for High Availability.
This is a good place to post questions as this forum is
monitored by the PowerHA support group.
Uempty
Unit Objectives
After completing this unit, you should be able to:
• Describe basic concepts and
terminology of PowerHA (formerly HACMP)
• Describe the components of a PowerHA cluster
• Explain how PowerHA operates in some typical cluster
configurations
• Discuss considerations and limitations of PowerHA
Objectives
In this unit we introduce PowerHA , describe the concepts, components, abilities and
limitations of PowerHA, and provide an overview of basic PowerHA cluster
configurations.
PowerHA terminology
In this course we will use the following terminology:
- PowerHA will mean any version and release of the PowerHA product.
- PowerHA x will mean version x and any release of that version
- PowerHA x.y will mean a specific version and release
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-3
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
PowerHA characteristics
IBM’s PowerHA product is a mature and robust technology for building a high
availability solution. A high availability solution based upon PowerHA provides
automated failure detection, diagnosis, recovery and reintegration. With an appropriate
application, PowerHA can also work in a concurrent access or parallel processing
environment, thus offering excellent horizontal scalability.
Uempty
A Highly Available Cluster
Fundamental Concepts
clstrmgr clstrmgr
urce
Reso up
gro Shared Storage
Node Node
Fallover
A Cluster is comprised of
•Topology (nodes, networks, network adapters)
•Resources (items being made available: applications, volume groups, etc.)
Fundamental concepts
PowerHA is based on the fundamental concepts of cluster, resource group, and cluster
manager (clstrmgrES).
Cluster
A cluster is comprised of nodes, networks and network adapters. These objects are
referred to as topology objects.
Resource group
A resource group is typically comprised of an application, network address, and a
volume group using shared disks. These objects are referred to as resource objects.
clstrmgrES
The cluster manager daemons together are the software components that
communicate with each other to control on which node a resource group is activated or
where the resource group is moved on a fallover based on parameters set up by the
administrator. The cluster manager runs on all the nodes of the cluster.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-5
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
N tw
Communication
on or
N
e
-IP k
Interface
n
tio
ica
un e
m evic
m
Co D
No
de
r
st e
Cl u
• Topology components
- Cluster:
• Nodes (Power Systems servers)
• Networks (connections between the nodes)
• Communication interfaces (e.g.: Ethernet or token-ring network adapters)
• Communication devices (e.g.: /dev/hdisk for heartbeat on disk non-IP network).
• Nodes
In this context, any IBM Power Systems server which is a member of a PowerHA
cluster.
• Networks
Networks consist of IP and non-IP networks. The non-IP networks ensure cluster
monitoring can be done if there is a total loss of IP communication. Non-IP networks are
strongly recommended to be configured in PowerHA, and are required to provide true
high availability. In PowerHA 5.4 the importance of non-IP networks is emphasized by
the issuance of warnings when a cluster verification is performed and non-IP networks
are missing.
Uempty
Nodes
All Power Systems servers work with PowerHA in IBM
IBM
pSeri
p5-550/550
server
HC
R
server
U6
pSer
ies
I
B
M
IBM
server
Supported nodes
As you can see, the range of systems that supports PowerHA is, well - everything. The
only requirement is that the system should have at least four adapter slots spare (two
for network adapters and two for disk adapters). The internal Ethernet adapter fitted to
most entry level Power Systems servers cannot be included in the calculations. It
should be noted that even with four adapter slots free, there is still be a single point of
failure as the cluster is only able to accommodate a single TCP/IP local area network
between the nodes.
PowerHA 5 works with Power Systems servers in a “no-single-point-of-failure” server
configuration. For a current list of systems that are supported with the version of
PowerHA you wish to use, please see the Sales manual at:
http://www.ibm.com/common/ssi.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-7
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Ethernet / Etherchannel
Token Ring PC
PC Server
FDDI
Server Server
Server
Non -IP
Traffic Flow
Heartbeat on Disk
Server
RS232/422
Target mode SSA
Target mode SCSI
Uempty
Un-Supported/Limited Support Networks
• Un-supported networks:
í Serial Optical Channel Converter (SOCC)
í Serial Line Internet Protocol (SLIP)
í Fibre Channel Switch (FCS)
í 802.3 Ethernet (et interface type)
í Virtual IP Address (VIPA) interface
í Aggregate IP interface (ml0) with SP Switch2
• Limited support:
í Internet Protocol Version 6 (IPv6)
• Un-supported prior to HACMP v5.5
• Supported for service addresses only in HACMP v5.5
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-9
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Resources
- Application Server - the application itself must be identified. The use of the term
Application Server in PowerHA describes the start/stop methods (scripts) for the
application. It is an object that points to the start/stop methods.
- Service IP Address/Label - IP address or hostname users use to connect to the
application. This IP address/hostname becomes a resource in the resource group
as it must be associated with the same node that is running the application. More
than one Service IP Address may be configured for a resource group.
- Volume Group - if the application requires shared disk storage, this storage is
contained within volume groups.
- Filesystem - an application often requires that certain filesystems be mounted.
In addition to the resources listed in the figure, you can also associate the following with
a resource group:
- NFS mounts - NFS filesystems to be mounted by the node running the application
- NFS exports - NFS filesystems to be exported by the node running the application
- Attributes - that can be assigned to control various aspects of the group’s behavior
Uempty
Resource Groups
Nodes where the •Application server
resource group can run •Service IP address
•Volume group
NODE LIST
•File system
How PowerHA will •and so forth
choose which node to RESOURCES
run the resource group
RUN POLICIES
ce G roup
Resour
Resource group
A resource group is a collection of resources treated as a unit by PowerHA, including:
- The resources needed by an application
- The nodes they can potentially be activated on
- The policies the cluster manager should use to decide which node to choose during
startup, fallover, and fallback.
A cluster may have more than one resource group (usually one for each application),
thus allowing for very flexible configurations.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-11
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Solution Components
Ethernet / Etherchannel
PC EMC
DS4000
RS/6000
RS/6000
Server Server
IBM
SAN
DS8000 RS/6000
Fibre
Uempty
AIX's Contribution to High Availability
9Object Data Manager (ODM)
9System Resource Controller (SRC)
9Logical Volume Manager (LVM)
9Journaled File System (JFS/JFS2)
9Online JFS Backup (splitlvcopy)
9Work Load Manager (WLM)
9Quality of Service (Qos)
9External Boot
9Software Installation Management (installp, smit, websmit)
9Reliable Scalable Cluster Technology (RSCT, RMC)
AIX contributions
AIX offers many availability benefits, for example, logical volume manager mirroring of
all data types, the journaled filesystem, and AIX’s capability to execute an online JFS
backup. AIX has a dynamic, multi-threaded kernel which allows configuration changes
to the operating system while the computer is running.
How often have you had to reboot an AIX system in order to configure a new hardware
device or extend a filesystem?
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-13
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
RS/6000
RS/6000
DS4000
IBM
SSA
Maximum 25m
Host
System
SCSI DS8000
Controller RS/6000
Uempty
Some Clusters Do Not Have Shared Disks
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-15
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Additional Features of PowerHA
OLPW Configuration
webSMIT Assistant
Verification
Auto tests
clstrmgrES
CSPOC
DARE SNMP
RG in
WPAR
Tivoli Application
NetView Monitoring
Additional features
- C-SPOC simplifies administration by providing SMIT menus to perform
administration tasks across all nodes in the cluster: including LVM, start/stop cluster
services, move a RG to another node or bring a RG offline or online.
- Dynamic Automatic Reconfiguration Event (DARE): Configuration changes can
be made to the cluster while the cluster is running.
- PowerHA plug-ins are available to monitor your cluster from Tivoli NetView.
- Application monitoring should be used to monitor the cluster’s applications and
restart them should they fail.
- Resource Group in WPAR: Beginning in PowerHA 5.5, PowerHA integrates with
the WPAR feature in AIX 6. A resource can be run within a WPAR.
- Verification is provided at PowerHA startup time, as part of synchronization, as a
manual process and a daily Automatic Cluster Configuration Monitoring function.
- The two node configuration assistant facility allows you to configure an PowerHA
cluster with very little input.
- Administration is made easier by the use of Online Planning Worksheets (OLPW)
and a WebSMIT interface.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-17
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Customized
Pre-event scripts
Application
PowerHA core events
•Application start script
•Application stop script
•Application monitor (optional) Customized
post-event scripts
Customization required
Application start and stop scripts: At a minimum, you must provide application start
and stop scripts to allow PowerHA to control your application.
Application monitors: It is highly recommended that you also configure application
monitors. The monitors can be custom scripts, or you can do process monitoring using
RMC.
PowerHA Smart Assists: These are a priced feature which provide start, stop and
monitor scripts for three popular applications (DB2, Oracle and Websphere). In
PowerHA 5.4 and later there is an API provided that allows 3rd party application
vendors to write Smart Assists.
Pre- and post-event scripts: PowerHA is shipped with event scripts (PowerHA core
events) which handle failure processing. In most cases, these core events will do all
that is needed. However, if you need to provide some special fallover behavior for your
environment, you can further customize PowerHA using pre- and post-event scripts.
Uempty
Just What Does PowerHA Do?
Additional monitoring
ERROR NOTIFICATION (via the AIX errdemon):
- Quorum loss: Moves affected RG to another node if quorum is lost
- Optional: You can add "automatic error notification" to monitor for disk failures
- Optional: You can monitor for anything that creates an entry in the errlog
APPLICATION MONITORS (via RMC or via a script which the user creates):
- Optional: Provides a mechanism to monitor that your application is running.
If not running, PowerHA tries to restart it. If it still fails, PowerHA moves the RG to
another node.
USER DEFINED EVENTS (via Resource Monitoring and Control (RMC)):
- Optional: UDEs can be created to use RMC to monitor for almost any possible
event in AIX. (For example, a file system full or a process not running, etc.)
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-19
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
• How the cluster responds to a failure depends on what has failed, what
the resource group's fallover policy is, and if there are any resource
group dependencies
• The cluster's configuration is determined by the application's
requirements
• Typically another component of the same type takes over duties of
failed component (for example, another node takes over from a failed
node)
UNIX Software Service Enablement
Uempty
What Happens When a Problem is Fixed?
?
• How the cluster responds to the recovery of a failed component depends
on what has recovered, what the resource group's fallback policy is, and
what resource group dependencies there are
• The cluster's configuration is determined by the application's
requirements
• Cluster administrator may need to indicate/confirm that the fixed
component is approved for use
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-21
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
node1 node2
A A
One node is primary
node1 node2 node1 node2
(no change)
A A
Standby
In standby configurations one (or more) nodes have no workload. The application
(resource group (RG)) normally runs on a primary or home node and the node with no
workload is the secondary or standby node.
- The startup policy indicates the home node
- The fallover policy indicates how to select the fallover node
- The fallback policy indicates automatic or manual fallback to the primary node
when the primary node recovers
There are several drawbacks:
- One node is not used (this is ideal for availability but not for utilization)
- A second outage on fallback
This can be extended to more nodes in two ways
- All nodes except one have applications and one node is the standby node.
- The resource group could be configured to have multiple backup nodes. If the
primary node fails, the RG is moved to the first standby node. If that fails, the second
standby node would be used. And so forth.
Uempty
Standby: Without Fallback
A node2 returns
node1 fails
node1 node2
A
Minimize downtime A
A
node1 node2
Minimize downtime
A resource group can be configured to not fall back to the primary node (or any other
higher priority node) when it recovers. This avoids the second outage which results
when the fallback occurs.
The cluster administrator can request that PowerHA move the resource group back to
the higher priority node at an appropriate time or it can simply be left on its current node
indefinitely (an approach which calls into question the terms primary and secondary, but
which is actually quite a reasonable approach in many situations).
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-23
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Mutual Takeover
A B
node1 fails node1 node2 node2 fails
Very common A B
B A
node1 node2
node1 node2
Mutual takeover
An extension of the primary node with a secondary node configuration is to have two
resource groups, one failing from right to left and the other failing from left to right. This
is referred to as mutual takeover.
Mutual takeover configurations are very popular configurations for PowerHA since they
support two highly available applications at a cost which is not that much more than
would be required to run the two applications in separate stand-alone configurations.
Issues
Each cluster node probably needs to be somewhat larger than the stand-alone nodes
as they must each be capable of running both applications, possibly in a slightly
degraded mode, should one of the nodes fail. If you are involved in the design and
planning of a cluster, it is important that these issues get discussed. If not properly
sized, the application performance may be unacceptable in a fallover situation when
both applications are running on one node.
Uempty
Concurrent: Multiple Active Nodes
Concurrent mode
PowerHA also supports resource groups in which the application is active on multiple
nodes simultaneously. In such a resource group, all nodes run a copy of the application
and share simultaneous access to the disk. This style of cluster is often referred to as a
concurrent access cluster or concurrent access environment.
The application must be specifically designed to support concurrent access.
Service addresses
Since the application is active on multiple nodes, each node must have its own address
for clients to connect. There is no PowerHA service address resource for concurrent
resource groups. The client systems must be configured to select which server address
to communicate with, and be prepared to switch to another address should the one that
they’re dealing with stop functioning (presumably, because the node has failed). It is
also possible to configure an IP multiplexer between the clients and the cluster which
redistributes the client sessions to the cluster nodes, although care must be taken to
ensure that the IP multiplexer does not itself become a single point of failure.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-25
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Terminology
A clear understanding of the above concepts and terms is important as they appear
over and over again both in the remainder of the course and throughout the PowerHA
documentation, log files and smit screens.
Uempty
Cluster Requirements and Capabilities
• Applications:
– Must be manageable via scripts (start/restart and stop)
– Can be restarted via monitoring
• Resource groups:
– Must be serviced by at least two nodes
– Can have different policies
– Can be migrated (manually or automatically) to rebalance loads
• Clusters:
– Must have at least one IP network and should have one non-IP
network
– Need not have any shared storage
– Can have any combination of supported nodes *
– For disaster recovery, the cluster may be split across two sites
•May or may not require replicating data (PowerHA/XD).
Importance of planning
Planning, designing, configuring, testing and operating a successful PowerHA cluster
requires considerable attention to detail. In fact, a careful methodical approach to all the
phases of the cluster’s life cycle is probably the most important factor in determining the
ultimate success of the cluster.
Methodical approach
A careful methodical approach takes into account the relevant points above, and many
other issues which are discussed in this course and in the PowerHA documentation.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-27
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Disaster Recovery: Limited Distance
•Localized separation of data centers (sites)
– Limited distance: PowerHA for AIX
• Each site sees same LUNs (half local, half remote)
• LVM mirroring is used to copy data to remote site
• PowerHA Cross-site LVM mirroring capability can and should be used
• Max distance limited by bandwidth and latency of the SAN being used
WAN
LVM
1 1’
Mirroring
SAN
North South
Campus Campus
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-29
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
GLVM
Metro Mirror
1 1’
HAGEO*
Data Replication
Toronto Brussels
* HAGEO has been replaced by PowerHA/XD GLVM V5.5
UNIX Software Service Enablement
Uempty
Things PowerHA Does Not Do
TSM
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-31
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
When is PowerHA
Not the Correct Solution?
• Zero downtime required
– Maybe a fault tolerant system is the correct choice
– 7x24x365, PowerHA occasionally needs to be shut down for
maintenance
– Life-critical environments
• Security issues
– Too little security
•Lots of people with the ability to change the environment.
– Too much security
•C2 and B1 environments may not allow PowerHA to function as
designed.
• Unstable environments
– PowerHA cannot make an unstable and poorly managed
environment stable
– PowerHA tends to reduce the availability of poorly managed
systems
UNIX Software Service Enablement
Zero downtime
If you truly need zero down time, PowerHA is not the solution. It takes a few minutes for
fallover and PowerHA may need to be occasionally shut down for maintenance, An
example of zero down time may be the intensive care room. Also PowerHA is not
designed to handle many failures at once.
Security issues
PowerHA may not be able to operate in some highly secure environments.
Unstable environments
The prime cause of problems with PowerHA is poor design, planning, implementation,
or administration. If you have an unstable environment, with poorly trained
administrators, easy access to the root password, and a lack of change control and
documented operational procedures, PowerHA is not the solution for you.
With PowerHA, the only thing more expensive than employing a professional to plan,
design, install, configure, customize, and administer the cluster is employing an
amateur.
Uempty
What Do We Plan to
Achieve in this Course?
Your mission in this course is to build a two-node highly available
cluster using two previously separate System p systems, making a
single application highly available in a standby configuration.
This course focuses on configuring the networks for PowerHA.
Configuring shared storage will be covered in the next course.
A A
Goals
During this course you will design, plan, configure, customize, and administer a
two-node high availability cluster running PowerHA 5.5 on an IBM Power Systems
server. You will learn how to build a standby environment for one application.
Some classroom environments will involve creating the cluster on a single Power
Systems server between two LPARs. Although this is not a recommended configuration
for production, it provides the necessary components for a fruitful PowerHA
configuration experience.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-33
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Implementation process
• Work as a team. It can’t be stressed enough that it will be necessary to work with others
when you build your PowerHA cluster in your own environment. Practice here will be
useful.
• Look at AIX environment.
- For storage, configure adapters and LVM components required for application
- For networks, configure communication interfaces, devices, name resolution via
/etc/hosts and service address(es) for the application(s)
- For application build start and stop script and test outside of the control of PowerHA
• Install the PowerHA software and reboot.
• Configure the topology and resource groups (and resources).
• Synchronize, start and test.
Uempty
Sources of PowerHA Information
• PowerHA Web Site:
- http://www-03.ibm.com/systems/p/ha/
• PowerHA manuals are available online – READ THEM!
- http://www-03.ibm.com/systems/p/library/hacmp_docs.html
• Sales Manual (specifications, supported hardware, etc.):
- http://www.ibm.com/common/ssi
• Techdocs Flash web site (HW support and technical updates):
- http://www-03.ibm.com/support/techdocs/atsmastr.nsf/Web/Flashes
• /usr/es/sbin/cluster/release_notes
• USSE PowerHA live virtual classes:
- http://bvrgsa.ibm.com/projects/t/tse/Roadmap/HACMP.html
• HA and HACMP Wiki:
- http://w3.tap.ibm.com/w3ki03/display/hacmp/Home+of+The+HA+and+HACMP+Wiki
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-35
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Review (1 of 2)
Uempty
Review (2 of 2)
4. True or False?
Resource Groups may be moved from node to node.
5. True or False?
PowerHA/XD is a solution for building geographically
distributed clusters.
6. Which of the following capabilities does PowerHA not
provide (select all that apply)?:
a. Time synchronization.
b. Automatic recovery from node and network adapter
failure.
c. System Administration tasks unique to each node.
Backup and restoration.
d. Fallover of just a single resource group.
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-37
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Exercise
Exploring PowerHA
Notes:
Uempty
Unit Summary 1 of 2
• PowerHA concepts:
– Topology: network components
• IP networks (network adapter)
• Non-IP networks (network device)
– Resources: The entities being made available
• Application server - Object that points to application start and stop scripts
• Service IP address - Address/hostname used to access application
• Storage - Volume groups, file systems, etc.
• NFS mounts / NFS exports
– Resource group (RG): A collection of resources that PowerHA
controls as a group
– Fallover: Move resource group to another node when there is a failure
– Fallback: Move resource group to "normal" node when recovery
occurs
– Resource group policies: How PowerHA will handle a resource group
in event of a failure
• Startup policy: determines target node for the resource group when PowerHA is started
• Fallover policy: determines target node when there is a failure
• Fallback policy: determines target node when recovery occurs
– Event: A change of status within a cluster
• The PowerHA system is event-driven
• When the Cluster Manager detects a change in cluster status, it initiates pre-defined
event handling and any user-defined customized processing
© Copyright IBM Corp. 2007, 2009 Unit 1. Introduction to PowerHA for AIX 1-39
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit Summary 2 of 2
• PowerHA provides these basic functions:
– Monitor networks (communication adapters/devices) for failure (RSCT)
– Monitor applications for failure (application monitors)
– Monitor many other failures and events (AIX errdemon and RMC)
– Move resource groups (fallover / fallback) based on resource group policies
• Some customization is required:
– Application start and stop scripts (required)
– Application monitors (optional, but highly recommended)
– Pre- and post-event scripts (optional, if customized fallover is needed)
• Basic PowerHA configurations:
– Serial access to shared data and resources:
application runs on one node at a time
• Standby (with or without automatic fallback) - standby node takes over RG in case of failure
• Mutual takeover - multiple active nodes; RG moved to another active node in case of failure
– Concurrent access to shared data and resources:
application runs in parallel across several nodes
• PowerHA is not the right solution if:
– Zero downtime is required
– Highly secure environment is required
– Systems are unstable or poorly managed
References
http://www-03.ibm.com/systems/p/library/hacmp_docs.html
PowerHA manuals
http://www-947.ibm.com/systems/support/supportsite.wss/brandmain?brandind=
5000025
System p Support Site (Fix Central)
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-1
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit Objectives
After completing this unit, you should be able to:
• Discuss how PowerHA uses networks
• Describe PowerHA network topology components and terminology
• Explain why a non-IP network is essential for any PowerHA cluster
• Describe the basic PowerHA network configuration rules
• Describe persistent IP addresses and how they can be used
• Explain and configure IP Address Takeover (IPAT):
– IPAT via IP aliasing
– IPAT via IP replacement
– Select the appropriate style of IPAT for a given context
• Describe how the AIX boot sequence changes when IPAT is
configured in a cluster
• Discuss the importance of consistent IP addressing and labeling
conventions
• Explain how user systems are affected by IPAT related operations
• Describe the ARP cache issue and several ways it can be addressed
• Describe the limited IPv6 support available in PowerHA v5.5
Unit objectives
This unit discusses networking in the context of PowerHA.
Uempty
How PowerHA Uses Networks
PowerHA uses networks to:
1. Configure and make highly available an application access address
(service address)
2. Monitor network interfaces to detect failures and co-ordinate event
processing (RSCT)
3. Internode configuration, verification and synchronization (clcomd)
4. Provide cluster status (SNMP)
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-3
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
PowerHA Configures
Highly Available Service Addresses
• PowerHA configures application access addresses
– The application access address is called the Service Address
– Configured on top of an IP interface
– Can use either an alias or replacement method
– Address can be moved to a new interface when the underlying
interface fails
Uempty
PowerHA Monitors
Networks and Detects Failures
• NICs/devices are monitored by Topology Services
– Via heartbeats
• The following failures are detected:
– Network Interface Card (NIC) failure
– Network failure
– Node failure
IP network
en0 en1 en0 en1
non-IP network
node1 node2
Other failures
PowerHA uses AIX features such as the error notification facility to respond to other
failures (for example, the loss of a volume group can trigger a fallover) but PowerHA
does not detect them directly.
In addition, Cluster administrators can configure their own error notifications and even
User Defined Events to cause PowerHA to take action on almost any incident in AIX.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-5
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
node1 node2
Heartbeat packets
PowerHA’s primary monitoring mechanism is for Topology Services to send heartbeat
packets. They are sent from every NIC and to every NIC and to and from non-IP
devices.
Heartbeating pattern
IP and non-IP heartbeat packets are sent in a ring or loop. In a two node cluster this
amounts to heartbeat packets going in both directions.
Uempty
Heartbeat Rings
• For redundancy, there should be at least two NICs on each physical network
• Each set of NICs will be organized into a heartbeat ring
(for example: en0 on all nodes in ring one; en1 on all nodes in ring two)
• Heartbeats are sent in one direction around each ring
• Addresses on different subnets are used
(for example: en0 addresses on 10.0.1/24 and en1 on 10.0.2/24)
One physical network
en0 en1
node3 10.0.1.3 node3 10.0.2.3
UNIX Software Service Enablement
Heartbeat rings
To prevent the NIC from being a SPOF, at least two NICs are needed on each network.
In order to detect and diagnose failures, you must configure the sets of NICs into
separate subnets. For example, the first set of NICs (en0 in the drawing) are in one
subnet (10.0.1 subnet in the drawing) and the second set of NICs (en1 in the drawing)
are in a second subnet (10.0.2 subnet in the drawing).
This allows RSCT to send heartbeats to a specific NIC on a node and thus to be able to
differentiate between failures on one or the other NIC.
We will cover the rules for configuring these addresses used for heartbeating in a few
minutes.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-7
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
node1 node2
Detection
The heartbeat patterns just discussed are sufficient to detect a failure in the sense of
realizing that something is wrong. They are not sufficient to diagnose a failure in the
sense of figuring out exactly what is broken.
For example, if the en1 interface on node1 fails as in the visual above, node1 stops
receiving heartbeat packets via its en1 interface, and node2 stops receiving heartbeat
packets via its en1 interface. node1 and node2 both realize that something has failed,
but neither of them have enough information to determine what has failed.
Uempty
Failure Diagnosis Example
• When a failure is detected, PowerHA (RSCT Topology
Services) uses specially crafted packet transmission patterns
to diagnose the actual failure by ruling out other alternatives
• Example:
1. RSCT on node1 notices that heartbeat packets are no longer
arriving via en1 and notifies node2 (which has also noticed that
heartbeat packets are no longer arriving via its en1)
2. RSCT on both nodes send diagnostic packets between various
combinations of NICs (including out via one NIC and back in via
another NIC on the same node)
3. The nodes soon realize that all packets involving node1's en1 are
vanishing but packets involving node2's en1 are being received
4. DIAGNOSIS: node1's en1 has failed
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-9
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
node1 node2
UNIX Software Service Enablement
Partitioned cluster
Since each node is, in fact, still very alive, the result is that the applications are now
running simultaneously on both nodes. If the shared disks are also online to both nodes,
then the result could be a quite massive data corruption problem. This situation is called
a partitioned cluster. It is, clearly, a situation which must be avoided.
Uempty
Clusters Should Have Multiple Networks
• Multiple networks protect against a partitioned cluster
– There must be more than one network to distinguish
between: Failure of a network VS failure of a node
• Second network should be a non-IP network
– There must be a non-IP network to distinguish between:
Failure of a node's IP subsystem VS failure of a node
• Therefore, ALWAYS INCLUDE A NON-IP NETWORK!
non-IP network
node1 node2
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-11
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Node recovery
In contrast, a node is not considered to have recovered until the PowerHA daemons
have been started up on the node. This behavior allows the user to reboot the node and
implement any necessary repair action without the PowerHA declaring failures and/or
perform automatic reintegration.
The reintegration of a component might trigger quite significant actions. For example, if
a node is reintegrated which has a high priority within a resource group then, depending
on how the resource group is configured, the resource group might fallback.
Uempty
PowerHA Networking Support
• Supported IP networking technologies:
– Ethernet
•All speeds
•Not the IEEE 802.3 frame type which uses et0, et1 ...
– Etherchannel
– FDDI
– Token-Ring
– ATM and ATM LAN Emulation
– SP Switch 1 and SP Switch 2
– Infiniband
• Supported non-IP network technologies:
– Heartbeat over Disks (diskhb)
•Requires Enhanced Concurrent Volume Group and
PowerHA 5.x
– RS232/RS422 (rs232)
– Target Mode SSA (tmssa)
– Target Mode SCSI (tmscsi)
UNIX Software Service Enablement
Supported IP networks
PowerHA supports all of the popular IP networking technologies (and a few which are
possibly not quite as popular). Note that the IEEE 802.3 Ethernet frame type is not
supported.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-13
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Network Types
PowerHA categorizes all networks:
• IP networks:
• Non-IP networks:
IP network attribute
The default for this attribute is public. Oracle uses the private network attribute
setting to select networks for Oracle inter-node communications. This attribute is not
used by PowerHA itself. See the HACMP for AIX, Planning Guide for more information.
Non-IP networks
PowerHA uses non-IP networks for:
- Alternative non-IP path for PowerHA heartbeat and messaging
• Differentiates between node/network failure
• Eliminates IP as a single point of failure
Uempty
PowerHA Topology Components
• PowerHA uses some unique terminology to describe the type and
function of topology components under its control
IP lab
IP el s
Ne IP addres
two
rk
vancouver-service 192.168.5.2
Interface
Interface
Network
Network
Interface
Network
Card
Card
Card
Card
ne non
tw -IP
or
ks
Communication Device
Serial Serial
Port
non IP - rs232 Port
non IP - tmssa
node1 node2
non IP - diskhb
nod
en
am
e
Terminology
- node: An IBM system p server operating within an PowerHA cluster.
- node name: The name of a node from PowerHA’s perspective.
- IP label: For IP networks, the name specified in the /etc/hosts file or by the Domain
Name Service for a specific IP address.
- IP network: A network which uses the TCP/IP family of protocols.
- non-IP network or serial network: A point-to-point network which does not rely on
the TCP/IP family of protocols.
- communication interface: A network connection onto an IP network (slightly better
definition coming shortly).
- communication device: A port or device connecting a node to a non-IP network
(slightly better definition coming shortly).
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-15
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
PowerHA Address Terms
•A service address is:
–Used by clients to access an application
–Configured by PowerHA and kept highly available
–Configured either by replacement or by alias
•A service label is the IP label associated with a service address
•A service interface is an interface configured with a service address
•A persistent address is
–Configured by PowerHA by alias
–Kept highly available on the interfaces of a single node
•A persistent label is the IP label associated with a persistent address
UNIX Software Service Enablement
More terminology
• Service IP label / address: An IP label or address used by client systems to access
services running within the cluster. Used with IP Address Takeover (IPAT).
• Applications should use the service IP label / address: Non-service IP addresses
should not be used by client systems to contact the cluster’s applications. A client
system which gets into the habit of connecting to its application using a non-service IP
address will not be able to find its application after a fallover to a different node.
• Service interface: A communications interface used by clients that is configured with a
service IP label / address (either by alias or by replacement).
• Non-service (boot) IP label / address: An IP address which is stored in the AIX ODM
and configured onto a NIC at boot time.
• Non-service interface: A communications interface that is configured with a
non-service (boot) IP address. It can be used as a backup for a service IP address.
• Persistent IP label / address: An IP address monitored by PowerHA but it stays on the
node on which it is configured. It is implemented as an alias and PowerHA will attempt
to keep this IP label / address highly available on the same node.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-17
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
IP labels
# netstat -i
Name Mtu Network Address Ipkts Ierrs Opkts Oerrs Coll
lo0 16896 link#1 5338 0 5345 0 0
lo0 16896 127 localhost 5338 0 5345 0 0
lo0 16896 ::1 5338 0 5345 0 0
tr0 1500 link#2 0.4.ac.49.35.58 76884 0 61951 0 0
tr0 1500 192.168.1 vancouverboot1 76884 0 61951 0 0
tr1 1492 link#3 0.4.ac.48.22.f4 476 0 451 13 0
tr1 1492 192.168.2 vancouverboot2 476 0 451 13 0
tr2 1492 link#4 0.4.ac.4d.37.4e 5667 0 4500 0 0
tr2 1492 195.16.20 db-app-svc 5667 0 4500 0 0
Uempty
Node Name Considerations (2 of 2)
• AIX hostname, LPAR name and PowerHA node name are separate names
– They are often assigned the same value, but they do not need to be
(although it's less confusing if they are the same)
– Except if using DLPAR with PowerHA, then these three names must be
the same
• The AIX hostname is stored in the inet0 object in the ODM
(hostname and uname -n)
• The LPAR name is the name of the logical partition
(lparstat –i | grep Name or uname -L)
• The HACMP node name is the name of the node to PowerHA and is stored
in the HACMPnode object in the ODM
(get_local_nodename)
IP labels
Each IP address used by an PowerHA cluster almost certainly has an IP label
associated with it. In non-PowerHA systems, it is common for the system’s only IP label
to be the same as the system’s hostname. This is rarely a good naming convention
within a PowerHA cluster as there are just so many IP labels to deal with, and having to
pick which one gets a name that is the same as a node’s hostname is a pointless
exercise.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-19
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
• General:
- Each node must have at least one non-routed connection with all other nodes.
- Do not place intelligent switches, routers, or other network equipment that can filter
or delay packets between cluster nodes.
- All addresses on this network must use the same subnet mask.
• Heartbeating over IP interfaces:
- Each interface on a node must be on a different subnet
If there are multiple interfaces on the same subnet, PowerHA can not select which
interface will be used for heartbeating, so it cannot reliably monitor all interfaces.
- There must be at least one subnet in common with all nodes
• Heartbeating over IP aliases:
- You specify an IP address offset to be used for heartbeating. PowerHA then
configures a set of IP addresses and subnets for heartbeating.
- You must reserve a unique address and subnet range to be used for
heartbeating. These addresses need not be routed in your network.
- Other addresses can be on any subnet. However, there may be some routing
issues if the service address is in the same subnet as the boot addresses.
Uempty
Boot Address Example
• All addresses have the same subnet mask: 255.255.255.0
• en0 addresses are on the 192.168.10/24 subnet
• en1 addresses are on the 192.168.11/24 subnet
192.168.11.2 (ODM)
192.168.10.1 (ODM) 192.168.11.1 (ODM) 192.168.10.2 (ODM)
node1 node2
Examples
The visual shows some boot IP address examples. We’ll see the service IP address
examples later, when we discuss IPAT.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-21
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Persistent IP Addresses
• An IP address associated with a particular node
• Useful for administrative purposes:
– Provides highly available IP address associated with a particular node
– Allows external monitoring tools (for example, Tivoli) and administrative
scripts to reach a particular node
• Configured, via IP aliasing, after node synchronization
• Configured at boot time, even if cluster services are not running
• PowerHA will strive to keep a node's persistent address available on
that node -- never moved to another node
• Maximum of one persistent address per network per node
• Configuration rules:
– IPAT via Aliases: persistent address can not be in any boot subnet
– IPAT via Replacement: persistent address must be in the service
subnet
• Example:
– Boot addresses are in 192.168.10/24 and 192.168.11/24 subnets
– node1 persistent address: 9.47.87.101
– node2 persistent address: 9.47.87.102
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-23
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Ser
eel ss
Pal d ce
IIP errvviice
abdr
NF
Sy
Fil tem
S
e
s
SSe
ex
ro e
NF
G lu m
S po
up
mo rts
Vo
er
un
ts ion Serv
licat
App
Reso
u
Grou rce
p
Uempty
Two Ways to Implement IPAT
• IPAT via IP Aliases:
– PowerHA adds the service address to the NIC using AIX's IP aliasing
feature. The NIC now has two addresses assigned: boot address and
service address.
ifconfig en0 alias 9.47.87.157
– This is the default
– Requires more subnets than IPAT via IP replacement
– Allows multiple service addresses to use the same NIC, so may
require less hardware that IP replacement
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-25
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Requirements
• The following networks support IPAT via IP aliasing:
- Ethernet
- Token-ring
- FDDI
- SP switch 1 / SP switch 2
• No service IP labels on the network require hardware address takeover (HWAT)
Uempty
IPAT via IP Aliasing in Operation
• When the resource group comes up on a node, PowerHA
aliases the service IP label onto one of the node's available
(that is, currently functional) NICs (ODM)
9.47.87.22 (alias)
192.168.11.2 (ODM)
192.168.10.1 (ODM) 192.168.11.1 (ODM) 192.168.10.2 (ODM)
Operation
PowerHA uses AIX’s IP aliasing capability to alias service IP labels included in resource
groups onto network interfaces on the node which runs the resource group. With
aliasing, the boot IP address (stored in the ODM) is still present.
Note that one advantage of sorts of IPAT via IP aliasing is that the boot IP addresses do
not need to be routable from the client/user systems.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-27
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
9.47.87.22 (alias)
192.168.10.1 (ODM) 192.168.11.1 (ODM) 192.168.10.2 (ODM) 192.168.11.2 (ODM)
Interface failure
If a NIC fails, PowerHA moves the service IP addresses to another NIC, which is still
available, on the same network. If there are no remaining available NICs on the node
for the network, then PowerHA initiates a fallover for that resource group. Note:
PowerHA will choose among NICs that have been defined to PowerHA Topology.
User perspective
Since existing TCP/IP sessions generally recover cleanly from this sort of
failure/move-IP-address operation, users might not even notice the outage if they don’t
happen to be interacting with the application at the time of the failure.
Uempty
IPAT via IP Aliasing After a Node Fails
• If the node fails, PowerHA moves the resource group to a new
node and aliases the service address onto one of the new
node's currently functional NICs
Clients should use the service address to connect to the application. If they use a boot address, or
even a persistent address, they may not be able to connect after a failure!
(If possible, don't give them the boot addresses!)
9.47.87.22 (alias)
192.168.10.2 (ODM) 192.168.11.2 (ODM)
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-29
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
IPAT via IP Aliasing Summary
• Configure each node's NICs with boot addresses (each on a
different subnet)
• Assign service addresses to resource groups as appropriate
– Must be on separate subnet from boot addresses
– There is a total limit of 256 IP addresses known to PowerHA and
64 resource groups. Within those overall limits:
• There is no limit on the number of service addresses in a resource
group
• There is no limit on the number of resource groups with service
addresses
• PowerHA aliases service addresses to NICs based on
resource group rules and available hardware
• Hardware address takeover (HWAT) is not supported when
using IPAT via IP aliasing
• IPAT via IP aliasing requires gratuitous ARP support
Advantages
• It can require fewer physical interfaces than IPAT via Replacement
Probably the most significant advantage to IPAT via IP aliasing is that it supports
multiple service IP addresses on the same NIC and allows a node to easily support
quite a few resource groups. In other words, IPAT allows you to share several service
addresses on one interface.
Disadvantages
IPAT via alias is usually the best choice, but there are cases when it cannot be used:
• IPAT via IP aliasing does not support hardware address takeover (HWAT)
You will rely on gratuitous ARP as the means of resetting the ARP entries on IPAT. If
HWAT is required, you must use IPAT via replacement.
• IPAT via IP aliasing can require a lot of subnets
You must have a subnet for each interface and a subnet for each service IP label.
• IPAT via aliasing not supported for ATM networks
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-31
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
History
In the beginning, IPAT via IP replacement was the only form of IPAT available. IPAT via
IP aliasing became available when AIX could support multiple IP addresses associated
with a single NIC via IP aliasing. Because IPAT via IP aliasing is more flexible and
usually requires less network interface cards, IPAT via IP replacement is no longer the
recommended method.
Many existing cluster implementations still have IPAT via Replacement. When
upgrading to versions of PowerHA that support IPAT via Aliasing, you should consider
converting IPAT via Replacement configurations to Aliasing only if there is another
reason compelling you to do so. Otherwise, leave the IPAT via Replacement
configuration as it is. Any new implementations should strongly consider using IPAT via
Aliasing.
This visual gives a brief overview of IPAT via IP replacement. A detailed discussion can
be found in the Appendix.
Uempty
Service Address Examples
subnet mask = 255.255.255.0
Boot addresses on Boot addresses on Valid service Valid service addresses for
first node second node addresses for IPAT via IPAT via IP Replacement
IP Aliasing
192.168.5.1 192.168.5.25 192.168.7.1 192.168.5.3 and 192.168.5.97
192.168.6.1 192.168.6.25 9.37.202.213 OR
192.161.22.1 192.168.6.3 and 192.168.6.97
...
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-33
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
IP Naming Conventions
• PowerHA clusters tend to have quite a few IP labels and other
names associated with them
• Adopt appropriate labeling and naming conventions
Suggestions:
• Node-resident labels should include the node's name:
For example: node1-if1, node1-if2, node2-boot1, node2-boot2,
node1-if1, node1-if2 …
• Service IP labels that move between nodes should describe the
application rather than the node:
For example: web1-svc, infodb-svc, app1, prod1, …
• Persistent IP labels should include the node name (since they won’t
be moved to another node) and should identify that they are
persistent:
For example: node1-per, node2-per, node1adm, node1admin, …
• Why?
– Conventions avoid mistakes
– Mistakes reduce availability!
Uempty
Name Resolution
• All of the cluster's IP labels must be defined in every cluster
node's /etc/hosts file:
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-35
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
sw1
sw2
Shared Storage
Heartbeat on disk
appB 192.168.3.20
appA 192.168.3.10
Etherchannel history
Etherchannel is a “trunking” technology that allows grouping several Ethernet links.
Traffic is distributed across the links, providing higher performance and redundant
parallel paths. When a link fails, traffic is redirected to the remaining links within the
channel.
Uempty
Other Configurations: Virtual Ethernet
Virtual I/O Server (VIOS1) AIX Client LPAR 1 Virtual I/O Server (VIOS2)
Frame1 ent1 ent0 ent2 ent5 ent0 ent5 ent2 ent1 ent0
(phy) (phy) (virt) (virt) (virt) (virt) (virt) (phy) (phy)
Hypervisor
Hypervisor
Virtual I/O Server (VIOS1) AIX Client LPAR 2 Virtual I/O Server (VIOS2)
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-37
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Hypervisor
Virtual I/O Server (VIOS1) AIX Client LPAR Virtual I/O Server (VIOS2)
Additional information
Single adapter Ethernet networks in PowerHA require the use of a netmon.cf file. This
will be discussed in a few minutes.
Note that there does not have to be link aggregation at the VIO Server level. You could
use a single NIC in each VIO server and rely on the other VIO server for redundancy (in
a Shared Ethernet Adapter Failover configuration, for example).
Uempty
Other Configurations: Single Adapter Nodes
• Single IP Adapter nodes may appear to reduce cluster cost
• The cost reduction is an illusion:
1. A node with only a single adapter on a network is a node with a single
point of failure - the single adapter
2. Clusters with unnecessary single points of failure tend to suffer more
outages
3. Unnecessary outages cost (potentially quite serious) money
• PowerHA needs at least two NICs per network for failure diagnosis
• Clusters with fewer than two NICs per IP network, though supported,
are not recommended
– If you have a single adapter network, PowerHA will issue warnings when
cluster verification is run
• Single adapter networks are considered OK if redundancy is provided
via another mechanism, such as:
– Virtual Ethernet, backed with two VIOS LPARs in failover configuration
– Etherchannel
• If you have a single adapter network with redundancy:
– Configure /usr/es/sbin/cluster/netmon.cf file
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-39
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Talk to Your Network Administrator
•Explain how PowerHA uses networks
•Ask for what you need:
–IPAT via IP aliasing:
•Service IP address(es) (for client connections)
•One boot IP address for each NIC (for heartbeating)
–One subnet per NIC on each node
(for example: three NICs per node requires three subnets)
–Must not be in any service subnet
–These addresses do not need to be routable
–IPAT via IP replacement:
•Service IP address(es), must be in one subnet (for client connections)
•One boot address for each NIC (for heartbeating)
–One subnet per NIC on each node
(for example: three NICs per node requires three subnets)
–One boot subnet must be the same as the service address subnet
•Service address subnet must be routable, others need not be routable
–Persistent IP address for each node (for admin access, optional)
•Cannot be on any boot subnet
•For IPAT via Replacement: must be in the service subnet
•Ask early (getting subnets assigned may take some time)
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-41
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Changes to AIX Boot: 5.3, 5.4 and 5.4.1
• The startup sequence of AIX networking is changed when IPAT
/etc/inittab
is enabled
/etc/inittab /sbin/rc.boot
cfgmgr
/sbin/rc.boot /etc/rc.net
cfgmgr modified for IPAT: exit 0
/etc/rc.net /etc/rc
config NICs, mount FSs
hostname, static /usr/es/sbin/cluster/etc/harc.net
routes config NICs, config persistent
addresses, rc.net -boot,
/etc/rc start selected daemons
mount FSs /etc/rc.tcpip
changed to run level a, not run
/etc/rc.tcpip /etc/rc.nfs
start network
changed to run level a, not run
daemons
Figure 2-42. Changes AIX Boot: 5.3, 5.4 and 5.4.1 QV1203.1
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-43
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Changes to AIX Boot: 5.5
•telinit -a no longer needed for IPAT
/etc/inittab /etc/inittab
/sbin/rc.boot /sbin/rc.boot
cfgmgr cfgmgr
/etc/rc.net /etc/rc.net
config NICs, not modified
hostname, static /etc/rc
routes mount FSs
/usr/es/sbin/cluster/etc/harc.net
/etc/rc config persistent addresses,
mount FSs cleanup NFS, start selected
daemons
/etc/rc.tcpip /etc/rc.tcpip
start network not modified
daemons /etc/rc.nfs
not modified
/etc/rc.nfs
start NFS
daemons
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-45
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Common TCP/IP Configuration Problems
• Subnet masks are not consistent for all HA network adapters
• Boot IP addresses on one node are placed on the same subnet
• Service and boot IP addresses are placed in the same subnet in
IPAT via IP aliasing networks
• Service and boot IP addresses are placed in different subnets in
IPAT via IP replacement networks
• Ethernet frame type is set to 802.3. This includes Etherchannel.
• Ethernet speed is not set uniformly or is set to autodetect
• The contents of /etc/hosts are different on the cluster nodes
• A different version of perl than is used by the PowerHA
verification tools (resulting in what appears to be a network
communications problem)
Configuration problems
The visual shows some common IP configuration errors to watch out for.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-47
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
What About the User's Computer?
• An IPAT operation renders ARP cache entries on client systems
obsolete
• Client systems must (somehow) update their ARP caches
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-49
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Gratuitous ARP
• AIX supports a feature called gratuitous ARP
– AIX sends out a gratuitous (that is, unrequested) ARP update
whenever an IP address is set or changed on a NIC
• Other systems on the local physical network are expected to
update their ARP caches when they receive the gratuitous ARP
packet
• Remember: only systems on the cluster's local physical network
must respect the gratuitous ARP packet
• So ARP update problems have been minimized
• Required if using IPAT via Aliasing
Gratuitous ARP
Whenever an IP address associated with a NIC changes, AIX broadcasts out a
gratuitous (in other words, unsolicited) ARP update. This gratuitous ARP packet is
generally received and used by all systems on the cluster’s local physical network to
update their ARP cache entries.
Uempty
What If Gratuitous ARP Is Not Supported?
• Some types of physical networks do not support gratuitous ARP
(for example: ATM)
• Some systems may ignore gratuitous APR broadcasts
(although this is rare)
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-51
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Configuration Rules: Non-IP Network
• Non-IP networks are strongly recommended in order to provide an
alternate communication path between cluster nodes in the event of an
IP network failure or IP subsystem failure
• You must provide a non-IP communication path, possibly via
intermediate nodes, between every pair of nodes in the cluster
• With more than three nodes you can configure the non-IP network
topology using one of the following layouts:
– Mesh: Each node is connected to all other nodes. This is the most robust, but,
requires the most hardware.
– Ring or loop (RECOMMENDED): Each node is connected to its two adjacent
neighbors. Each node, has two non-IP connections for heartbeating.
– Star (NOT RECOMMENDED): One node is connected to all other nodes. This is the
least robust; the center node becomes a single point of failure for all the associated
networks.
net_rs232_04
Non-IP networks
Non-IP networks are point-to-point: each connection between two nodes is considered
a network and a separate non-IP network label for it is created in PowerHA.
For example, the visual shows four RS232 networks, in a ring configuration, connecting
four nodes to provide full cluster non-IP connectivity.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-53
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
IP Version 6 (IPv6)
• IPv6 is not supported prior to PowerHA 5.5
• Limited support for IPv6 in PowerHA 5.5:
– Only for service addresses and persistent addresses
• NICs must have IPv4 boot address used for heartbeating
• IPv6 addresses allowed in netmon.cf
– ether network type only
– Only IPAT via aliases networks
– AIX 6.1D or later
– IPv6 address can not be used for HMC (DLPAR/CuoD)
– IPv6 service address can not be included in any WPAR enabled
resource groups
– No NFS v4 exports using IPv6 service address (NFS limitation)
Uempty
IP Version 6 History
• Why IPv6?
– Main reason for redesigning IP was the predicted IPv4 address
exhaustion
– Network Address Translation (NAT) provides relief but has its own
drawbacks
• How fast is IPv6 being adopted?
– A 2008 study by Google indicated that implementation was still less
than one percent of Internet-enabled hosts in any country
• The leaders were Russia (0.76%), France (0.65%), Ukraine (0.64%),
Norway (0.49%), and the United States (0.45%).
• Why IPv6 now for PowerHA?
– In US, the June 2003 “Stenbit memo” signaled the intent for
Department of Defense to transition to IPv6 by 2008.
– Other countries have similar initiatives.
– All IBM products must support or plan to support IPv6
• Realistic outlook is that IPv4 and IPv6 will coexist for decades
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-55
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Configuring IPv6
Service and Persistent Addresses
• Configured through the same set of SMIT panels
• New field for “prefix length” – mandatory for IPv6
[Entry Fields]
* IP Label/Address [] +
Prefix Length [] #
* Network Name [] +
Alternate HW Address to accompany IP Label/Address []
[Entry Fields]
* Node Name ppstest1
* Network Name [] +
* Node IP Label/Address [] +
Prefix Length [] #
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-57
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
en1
192.9.201.10 AIX boot address
fe80::c862:65ff:fee7:e2f7 Link local address
2001::1 Service address
Uempty
IPv6 Persistent Address
• AIX configures boot address (CuAt) - RSCT uses this for
heartbeat
• autoconf6 adds link local address
• PowerHA adds the IPv6 persistent address as an alias
• PowerHA adds the IPv6 service address as an alias
• PowerHA moves service and persistent address after failure
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-59
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
IPv6 References
• Documentation, reference, standards, or other publications, URLs,
etc. related to IPv6:
– Ipv6 @ Wikipedia:
http://en.wikipedia.org/wiki/Ipv6
– IPv6 Information Page:
http://www.ipv6.org/
– IPv6 Forum:
http://www.ipv6forum.com/
– “Format for Literal IPv6 Addresses in URL’s”:
http://www.ietf.org/rfc/rfc2732.txt
– IPv6 on AIX White Paper:
http://www.ibm.com/servers/aix/whitepapers/aix_ipv6.pdf
Uempty
Review 1 of 4
1. How does PowerHA use networks (select all which apply)?
a. Provide client systems with highly available access to the cluster's
applications
b. Detect and diagnose failures
c. Provide cluster status
d. Communicate between cluster nodes
e. Monitor network performance
2. Using information from RSCT, PowerHA only directly handles
three types of failures: ______________, ______________,
______________.
3. True or False?
• Heartbeat packets must be acknowledged or a failure is assumed to have
occurred.
4. True or False?
• Clusters should include a non-IP network.
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-61
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Review 2 of 4
5. True or False?
• Clusters must always be configured with a private IP network for PowerHA
communication.
6. Which of the following are true statements about communication
interfaces? (select all that apply)
a. Has an IP address assigned to it using the AIX TCP/IP SMIT screens
b. Might have more than one IP address associated with it
c. Sometimes but not always used to communicate with clients
d. Always used to communicate with clients
7. True or False?
• Persistent node IP labels are not supported for IPAT via IP replacement.
8. True or False?
• Each NIC on each physical IP network on each node is required to have an
IP address on a different logical subnet (unless using heartbeat over IP
alias).
9. True or False?
• A single cluster can use both IPAT via IP aliasing and IPAT via IP
replacement.
Uempty
Review 3 of 4
10.True or False?
• All networking technologies supported by PowerHA support IPAT via IP
aliasing.
11.True or False?
• All networking technologies supported by PowerHA support IPAT via IP
replacement.
12.If node1 has NICs with the IP addresses 192.168.20.1 and
192.168.21.1 and node2 has NICs with the IP addresses
192.168.20.2 and 192.168.21.2, then which of the following are
valid service IP addresses if IPAT via IP aliasing is being used?
(select all that apply)
a.(192.168.20.3 and 192.168.20.4) OR (192.168.21.3 and 192.168.21.4)
b.192.168.20.3 and 192.168.20.4 and 192.168.21.3 and 192.168.21.4
c. 192.168.22.3 and 192.168.22.4
d.192.168.23.3 and 192.168.24.3
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-63
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Review 4 of 4
13. If node1 has NICs with the IP addresses 192.168.20.1 and
192.168.21.1 and node2 has NICs with the IP addresses
192.168.20.2 and 192.168.21.2 then which of the following are
valid service IP addresses if IPAT via IP replacement is being
used? (select all that apply)
a. (192.168.20.3 and 192.168.20.4) OR (192.168.21.3 and 192.168.21.4)
b. 192.168.20.3, 192.168.20.4, 192.168.21.3 and 192.168.21.4
c. 192.168.22.3 and 192.168.22.4
d. 192.168.23.3 and 192.168.24.3
14. True or False?
• Clients are required to exit and restart their application after a fallover.
15. True or False?
• All client systems are potentially directly affected by the ARP cache issue.
16. True or False?
• clinfo must not be run both on the cluster nodes and on the client
systems.
Uempty
What We Are Going to Do in This Class
student hdisk0
rootvg client
en1 en2 en0
laptops Virtual disk - client only
Node Name = node1 Home node name = node1 Node Name = node2
en1 (boot) = node1-if1
Startup Policy = home node en1 (boot) = node2-if1
192.168.1.1 Fallover Policy = next prio node 192.168.1.2
en2 (boot) = node1-if2 Fallback Policy = never fallback en2 (boot) = node2-if2
192.168.2.1 Service IP = appA-svc 192.168.2.2
Persistent = node1-per 192.168.3.10 Persistent = node1-per
192.168.3.1 192.168.3.2
Application server = appA
en0 (login) = appA start script = /tmp/qv20/startA en0 (login) =
non-IP network
hdisk1 hdisk2
hdisk0 hdisk0
rootvg rootvg
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-65
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
= node1
Node Name = node1
Home node name = node1 Node Name = node1
= node2
en0 (boot)
en1 = node1-if1
= node1-if1
Startup Policy = home node en1 (boot) = node2-if1
= node2-if1
192.168.1.1
192.168.1.1 Fallover Policy = next prio node 192.168.1.2
192.168.1.2
en1 (boot)
en2 = node1-if2
= node1-if2 Fallback Policy = never fallback en2 (boot) = node2-if2
= node2-if2
192.168.2.1
192.168.2.1 ServiceIP
Service IP = appA-svc 192.168.2.2
192.168.2.2
Persistent = node1-per
= node1-per 192.168.3.10
192.168.3.10 Persistent = node1-per
= node2-per
192.168.3.1
192.168.3.1 192.168.3.2
192.168.3.2
Application server = appA
en0 (login) = 9.47.8x.x1 appA start script = /tmp/qv20/startA en0 (login) = 9.47.8x.x2
non-IP network
hdisk1 hdisk2
hdisk0 hdisk0
rootvg rootvg
Most commonly, PowerHA clusters have two NICs per node. Boot addresses often use
private addresses (not routable). Service and persistent addresses are usually routable.
In our class environment, all the addresses used by PowerHA are private addresses.
And we have three interfaces on each node. This is NOT required by PowerHA. Why
did we do it this way?
- We need multiple clusters (one for each team)
- If all the clusters were on the same physical network, we would need different
cluster addresses for each team. This made the lab instructions confusing.
- Instead we isolated each cluster: Each cluster, and one client, are within a p520
system.
- Cluster addresses (boot, service and persistent) are all private addresses and on a
private network. This would be pretty unusual in a production cluster, but is very
useful for our classroom.
- The third NIC (en0) is connected to the external network and is not used by
PowerHA at all. It is only used by us to connect to the LPARs. This “external login”
address is the only routed address in our configuration. This would not be normal for
a production cluster.
Uempty
What We Are Going to Do in This Exercise
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-67
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit summary (1 of 2)
• PowerHA uses networks to:
– Provide highly available client access to applications in the cluster
– Detect and diagnose NIC, node, and network failures using RSCT
– Communicate with PowerHA daemons on other nodes
– Provide cluster status using SNMP
• All PowerHA clusters should have a non-IP network
– Differentiate between node, IP subsystem and network failures
– Prevent cluster partitioning
• PowerHA networking terminology
– Service IP label/address: HA address used by client to access application
– Boot (non-service) IP label/address: Applied to NIC at boot time; stored in
AIX ODM
– Persistent IP label/address: Node bound HA address for admin access to a
node
– Communication interface: NIC used in an IP network
– Communication device: Device used in non-IP network
– Communication adapter: X.25 adapter used in a HA communication link
– IP Address Takeover (IPAT): Moves service IP address to working NIC
after a failure
• IPAT via aliasing: Adds the service address to a NIC using IP aliasing
• IPAT via replacement: Replaces the non-service address with the service address
Uempty
Unit summary (2 of 2)
• PowerHA has very specific requirements for subnets
– All addresses on one network must use the same subnet mask
– IPAT via aliasing
• NICs on a node must be on different subnets (boot addresses)
• There must be at least one subnet in common with all nodes
• Service addresses must be on different subnet than any boot address
• A service address can be on same subnet with another service address
– IPAT via replacement
• NICs on a node must be on different subnets (boot addresses)
• Each service address must be in same subnet as one of the boot addresses
• Multiple service addresses must be in the same subnet
– Heartbeating over IP alias (any form of IPAT)
• Service and non-service addresses can coexist on the same subnet, or be on
separate subnets
• One subnet per NIC required for heartbeating; these do not need to be routed
• PowerHA can update local clients’ ARP cache after IPAT
– Gratuitous ARP (default)
– clinfo on clients
– clinfo on server nodes
– Hardware address takeover (HWAT),
requires IPAT via IP Replacement
© Copyright IBM Corp. 2007, 2009 Unit 2. Configuring Networks for PowerHA 2-69
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
References
http://www-03.ibm.com/systems/p/library/hacmp_docs.html
PowerHA manuals
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-1
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit Objectives
After completing this unit, you should be able to:
Uempty
How to Define an Application to PowerHA
• Create start/stop scripts; make executable; copy to all nodes; test on
all nodes
• Create resources in PowerHA
– Application server: start and stop scripts (and optionally monitor(s))
– Service address
• Create Resource Group
– Participating Nodes
– Run time policies: where to run
• Add resources to the group
– Application Server
– Service Address
– Volume Group
• Synchronize
Resource Group
Node 1 Node 2 Node 3
Shared
Disk
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-3
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Application Considerations 1 of 2
• Automation
– No manual intervention required
• Dependencies
– Using names unique to one node
– Node based licensing
– Dependencies with other applications
• Interference
– Conflicts with PowerHA
• Automation
One key requirement for an application to function successfully under PowerHA is that
the application must start and stop without any manual intervention. Since the cluster
daemons call the start and stop scripts, there is no option for interaction.
• Dependencies
Consider what dependencies your application has that might prevent it from running in
a cluster environment. For example:
i. Referring to a locally attached device
ii. Hard coding such as /dev/tty0 which may not be the same on another node
iii. Using a hostname which is not the same on other nodes
iv. One application must be up before another one
v. Applications that must both run on the same node
• Interference
An application may execute properly on both the primary and standby nodes. However,
when PowerHA is started, a conflict with the application or environment could arise that
prevents PowerHA from functioning successfully. Two areas to look out for are: Using
IPX/SPX Protocol and manipulating network routes.
Uempty
Application Considerations 2 of 2
• Robustness
– Application can restart after hardware failure
– Is there cleanup that needs to be done by your scripts?
• Implementation
– Data storage
– Effective scripts
– Using crontab or inittab
• Monitoring using PowerHA
– Startup monitor: Checks that app has started
– Long running monitor: Checks that app is still running
– If not running: attempt to restart on this node; fallover to other
node
– Used to be overlooked
– Nearly mandatory for
•Unmanaged resource groups
•Non-disruptive Startup/Upgrade
• See Applications and HACMP in the Planning Guide
UNIX Software Service Enablement
• Robustness
An application under PowerHA should meet robustness characteristics, such as
successful start after hardware failure, survival of real memory loss, and so forth.
• Implementation
- Where should data be stored, private storage or shared storage?
- Writing effective scripts (We will look at script considerations in a few minutes)
- Using inittab and crontab (inittab is processed before PowerHA is started and
crontab is local to a each node. Time must be synchronized if using crontab.)
• Monitoring using PowerHA
- Startup monitor:
• Checks if application is already running. Prevents multiple instances when
returning from unmanaged state or during non-disruptive startup or upgrade.
• Checks if application has started. If not, restart or fallover to another node.
- Long running monitor:
• Checks periodically that application is running. If not, restart or fallover to another
node.
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-5
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Node Node
1 2
LVM LVM
odm odm
pvid
Device
access shared access
pvid
Device
hdisk hdisk
Adapter disks Adapter
VGDA
rootvg
rootvg pvid rootvg
rootvg
private private
Uempty
Writing Start and Stop Scripts 1 of 2
•Start script issues
– Script should perform any preparation needed to start the
application.
– Script must handle any errors or cleanup from previous
termination. Is data recovery needed? You should assume data is
in an unknown state.
– Start script is run in background. If it fails to start the application,
PowerHA is unaware, even if the script returns an error.
Consider an application startup monitor.
– Script should check if application is already running and not start
application if it is already running (unless multiple instances are
desired).
– When starting an application with multiple instances, only start the
instances applicable for each node. Certain database startup
commands read a configuration file and start all known databases
at the same time; this may not be what is desired.
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-7
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Resource Group Policies
• Three initial policies
– Startup Policy
– Fallover Policy
– Fallback Policy
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-9
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Startup Policy
•Online On Home Node Only
– Resource group will only be started on the first node in the nodelist. If
unavailable, the resource group will not be started automatically.
Uempty
Online on All Available Nodes
• Application runs on all available nodes concurrently
– No fallover/fallback – just less/more nodes running the
application
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-11
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Fallover Policy
•Fallover To Next Priority Node In The List
– Example: If the nodelist is node2,node1,node3 and node2 fails, PowerHA
will attempt to start the resource group on node1. If node1 is not up, PowerHA
will try node3.
•Fallover Using Dynamic Node Priority
– Fallover node is determined dynamically. You can choose:
•cl_highest_mem_free (choose node with most available memory)
•cl_highest_idle_cpu (choose node with most available CPU cycles)
•cl_lowest_disk_busy (choose node with least disk activity)
– Example:
• Dynamic Node Priority Policy is set to cl_highest_mem_free
• The resource is running on node1 and node1 fails
• PowerHA will chose node2 or node3 based on which node has the most free
memory
– Only useful for clusters with three or more nodes. For two node clusters, Next
Priority Node and Dynamic Node Priority are the same.
•Bring Offline (On Error Node Only)
– Only for Online on All Available Nodes resource groups (concurrent
access)
– The resource group goes offline only on the error node but remains online on
the other nodes
UNIX Software Service Enablement
Uempty
Fallback Policy
•Never Fallback
– Resource group will NOT fallback when a higher priority node
joins the cluster
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-13
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Online using node – Fallover to next priority node – Never fall back
distribution policy – Fallover using Dynamic Node
Priority
Online on first available – Fallover to next priority node – Never fall back
node (OFAN) – Fallover using Dynamic Node – Fall back to higher
Priority priority node
Online on all available – Bring offline (on error node – Never fall back
nodes only)
Valid combinations
PowerHA allows you to configure only valid combinations of startup, fallover, and
fallback behaviors for resource groups.
Uempty
Dependent Applications/Resource Groups
RG 1
Parent RG
Child RG Child RG
RG 2 RG 3
Parent RG Parent RG
Child RG
RG 4
•Parent/Child Dependency
– Online: Parent will be started before the child. If the parent cannot be started, the child will
not be started
– Offline: Child will be stopped before the parent
– A resource group can have more than one child or more than one parent
– One resource group can be the parent of another resource group
– A child resource group can be a parent of a third resource group
•Location Dependency
– A resource group may only be on the same node/site or on a different node/site than
another resource group
•Implemented as Run-Time Policy
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-15
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Review
1. True or False
Applications are defined to PowerHA in a configuration file that lists
what binary to use.
2. What policies would be the best to use for a 2-node “active-active” cluster
using IPAT to minimize both applications running on the same node?
a. home, next, never
b. first, next, higher
c. distribution, next, never
d. all, error, never
e. home, next, higher
3. Which type of data should not be placed in private data storage?
a. Application log data
b. License file
c. Configuration files
d. Application binaries
4. True or False
Start and stop scripts which are tested on one node can be counted
on to perform correctly under PowerHA. No further testing is needed.
Uempty
Unit Summary
• To define an application to PowerHA, you must:
– Create an application server resource (start and stop scripts)
– Create a resource group (node list, policies, resources)
• Considerations for putting an application under PowerHA control
– Automation
– Dependencies
– Interference
– Robustness
– Implementation details
– Monitoring
– Shared storage requirements
• Considerations for start and stop scripts
– Environment
– Multiple instances
– Script location
– Error handling
– Coding issues
• Resource group policies control how PowerHA manages the
application
– Startup policy (with optional Settling timer)
– Fallover policy
– Fallback policy (with optional Delayed fallback)
• Resource group dependencies
– Parent/child dependencies control order of startup and shutdown
– Location dependencies control whether resource groups are started on same node or
same site
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 3. Configuring Applications for PowerHA 3-17
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
References
http://www-03.ibm.com/systems/p/library/hacmp_docs.html
PowerHA manuals
Unit Objectives
After completing this unit, you should be able to:
• State where installation fits in the implementation process
• Describe how to install PowerHA 5.5
• List the prerequisites for PowerHA 5.5
• List and explain the purpose of the major PowerHA
components
Uempty
Steps for Successful Implementation
• Proper planning is critical to a successful implementation
– Special care should be taken when installing PowerHA on a system
which is in production
Uempty
First Step in Installation: Documentation
• PowerHA manuals can be obtained from the CD or from:
http://www-03.ibm.com/systems/p/library/hacmp_docs.html
• HACMP for AIX 5L V5.5 Planning Guide
SC23-4861-11 11/2008
– Contains Planning Worksheets: print and fill in by hand
– Online Planning Worksheets application
• Can be used in planning, then applied to create the cluster
• Can be used to document an existing cluster
• Runs on AIX or Windows (Java 1.3.0 or later)
• Install from CD or from /usr/es/sbin/cluster/worksheets directory
• Instructions in the Planning Guide
• HACMP for AIX 5L V5.5 Installation Guide
SC23-5209-02 11/2008
• Release notes:
– On the CD as README_hacmp5.5
– Installed as /usr/es/sbin/cluster/release_notes
– There may be additional notes, such as:
README5.5.0.UPDATE, release_notes_xd, etc.
UNIX Software Service Enablement
• PowerHA manuals:
HACMP for AIX: Concepts and Facilities Guide SC23-4864
HACMP for AIX: Planning Guide SC23-4861
HACMP for AIX: Installation Guide SC23-5209
HACMP for AIX: Administration Guide SC23-4862
HACMP for AIX: Troubleshooting Guide SC23-5177
HACMP for AIX: Programming Client Applications SC23-4865
HACMP for AIX: Master Glossary SC23-4867
• PowerHA/XD manuals:
HACMP/XD GLVM Planning and Administration Guide SA23-1338
HACMP/XD: Metro Mirror Planning and Administration Guide SC23-4863
• PowerHA Smart Assist manuals:
HACMP for AIX: Smart Assist for WebSphere User's Guide SC23-4877
HACMP for AIX: Smart Assist for Oracle SC23-5178
HACMP for AIX: Smart Assist for DB2 SC23-5179
HACMP for AIX: Smart Assist Developer's Guide SC23-5210
PowerHA Filesets
Here are some of the PowerHA 5.5 filesets:
cluster.adt.es cluster.es.cfs
+ 5.5.0.0 ES Client CLINFO Samples + 5.5.0.0 ES Cluster File System Support
+ 5.5.0.0 ES Client Clstat Samples cluster.es.plugins
+ 5.5.0.0 ES Client Include Files + 5.5.0.0 ES Plugins - Name Server
+ 5.5.0.0 ES Client LIBCL Samples + 5.5.0.0 ES Plugins - Print Server
+ 5.5.0.0 ES Web Based Monitor Demo
cluster.doc.en_US.es + 5.5.0.0 ES Plugins - dhcp
+ 5.5.0.0 HAES PDF Documentation-U.S. English cluster.es.worksheets
+ 5.5.0.0 HAES Web-based HTML Documentation - + 5.5.0.0 Online Planning Worksheets
U.S. English cluster.hativoli
cluster.es.client
+ 5.5.0.0 ES Client Libraries
+ 5.5.0.0 HACMP Tivoli Client
+ 5.5.0.0 ES Client Runtime + 5.5.0.0 HACMP Tivoli Server
+ 5.5.0.0 ES Client Utilities cluster.license
+ 5.5.0.0 Web based Smit + 5.5.0.0 HACMP Electronic License
cluster.es.server cluster.man.en_US.es
+ 5.5.0.0 ES Cluster Test Tool
+ 5.5.0.0 ES Server Diags + 5.5.0.0 ES Man Pages - U.S. English
+ 5.5.0.0 ES Server Events cluster.msg.en_US
+ 5.5.0.0 ES Server Utilities + 5.5.0.0 HACMP CSPOC Messages - U.S.
+ 5.5.0.0 ES Two-Node Configuration Assistant English
+ 5.5.0.0 ES Base Server Runtime [...]
cluster.es.cspoc
+ 5.5.0.0 ES CSPOC Commands
+ 5.5.0.0 ES CSPOC Runtime Commands
•Your requirements will determine what you
+ 5.5.0.0 ES CSPOC dsh install
•Reboot is required after initial installation
UNIX Software Service Enablement
• Fileset considerations
- smit install_latest will not show the msg filesets: use install_all and select
the filesets.
- cluster.es contains both client and server components. You can install either or
both. Normally cluster nodes get both and clients get the client filesets.
- The same filesets should be installed on all nodes or Verify will give warnings.
- You should install the documentation filesets on at least one non-cluster node
(ensuring that the PowerHA PDF-based documentation is available even if none of
the cluster nodes will boot could prove REALLY handy someday).
- Some of the filesets require other products such as Tivoli or NetView. You should not
install these filesets unless you have these products. HAView is never installed on
the cluster node, it is installed on the NetView server.
- The cluster.es.cfs fileset can only be used if GPFS is installed.
- You may not need the plug-ins.
• Reboot after install
The installation or upgrade of PowerHA filesets normally requires rebooting the nodes.
Starting in PowerHA 5.4 PTF 1, you can install subsequent updates without rebooting
the system or disrupting applications.
Uempty
Remember the Prerequisites
• Always check:
– HACMP for AIX: Installation Guide Complete list of prerequisites
– Release notes / README files Any updated prerequisites
• When using AIX Version 5.3:
– AIX V5.3 with Technology Level 9
– RSCT 2.4.10.0
• When using AIX Version 6.1:
– AIX V6.1 with Technology Level 2 Service Pack 1 and APAR IZ31208
– RSCT 2.5.2.0
• AIX filesets
– The Installation Guide lists a number of AIX filesets that are needed for
PowerHA in any case. Some of these are not part of a standard base install
and must be installed before you install PowerHA.
– Other prerequisites are only required for specific needs
• Enhanced concurrent mode Volume Groups:
– bos.rte.lvm (at required TL version)
– bos.clvm.enh (at required TL version)
• Online Planning Worksheets
– AIX 5L Java Runtime Environment
• And so forth
– See the Installation Guide and release notes
Prerequisites
Listed above are the minimum prerequisites. As time goes by, these will almost
certainly be superseded by later levels.
Always check the Installation Guide for a complete list of prerequisites. Then check the
release notes for any additional prerequisites or requirements.
Bevore
Uempty
Things to Check Before Configuration
• Code installation
– Correct filesets and prerequisites have been installed
– Documentation is installed and accessible
• Network setup
– /etc/hosts file is configured on all nodes correctly
– Name resolution works
– IP and non-IP networks are configured
– Subnets configured correctly
• The subnet mask is identical
• All interfaces are on different subnets
– Routing configured correctly
– Test connectivity
• Shared storage configured correctly
• You have a written plan describing configuration and testing
procedures!
Uempty
How Does PowerHA Fit into AIX?
Here are the layers of software on an PowerHA 5.5 cluster node:
Application Layer
Contains the highly available applications that use
PowerHA services
PowerHA Layer
Provides highly available services to applications
AIX Layer
Provides operating system services (SRC, snmpd)
Uempty
Cluster Manager
• Runs on each cluster node
– PowerHA 5.3 and later: always running, even if cluster services are not
started
• Implemented by the clstrmgrES subsystem
• Primarily responsible for responding to cluster events:
– Recover from software and hardware failures
– Respond to user-initiated events:
•Request to online/offline a node
•Request to move/online/offline a resource group
•And so forth
• A client to RSCT
• Provides snmp retrievable status information
• Started in /etc/inittab and always running
Cluster Secure
Communication Subsystem
• Runs on each cluster node
– PowerHA 5.3 and later: always running, even if cluster services not started
• Implemented by the clcomdES subsystem
• Provides communication infrastructure for configuration tools:
– C-SPOC
– Verification and synchronization
– Discovery
– WebSMIT
• PowerHA provides several security options:
– Connection Authentication
• Standard
– Authentication
– Uses /usr/es/sbin/cluster/rhosts file and PowerHA ODM files
• Kerberos (SP only)
– Authentication and privacy (encryption)
• Virtual Private Networks (VPN) using persistent labels
– Authentication and privacy (encryption)
– VPNs are configured within AIX
– PowerHA is then configured to use VPNs
– Message authentication or Message authentication and Encryption
• Provided via RSCT (ctcas subsystem)
• PowerHA provides methods for key distribution
UNIX Software Service Enablement
Uempty
Cluster Communication Daemon (clcomd)
• Provides secure node-to-node communications without use
of /.rhosts
clcomd basics
This daemon replaces a number of ad hoc communication mechanisms with a single
facility thus funneling all cluster communication through one point. This makes it
feasible to use a VPN to send traffic between nodes.
clcomd:
Standard Connection Authentication
•clcomd's authentication algorithm uses:
– Special rhosts file: /usr/es/sbin/cluster/etc/rhosts
– PowerHA ODM
• Upon receiving a connection request:
– If the rhosts file is missing: connection is rejected
– If the rhosts file exists, then check the PowerHA ODM
• If HACMPadapter and HACMPnode are populated:
– Does requestor match an entry in the ODM files?
»No: Request is rejected
»Yes: Connect back and ask for the node name
If node name matches the ODM entry: request is accepted
• If HACMPadapter and HACMPnode are empty: (new cluster)
– If the cluster rhosts file is empty: Allow any connection request
– If the rhosts file is populated: Allow requests only from addresses in the file
Uempty
RSCT
• Reliable Scalable Cluster Technology
• Included with AIX
• Provides:
– Scalability to large clusters
– Cluster Failure notification
– Coordination of changes
• Key components used by PowerHA:
– Topology Services
•Heartbeat services and reliable messaging service
– Group Services
•Coordinates/monitors cluster state changes
– RMC: Resource Monitoring and Control
•Provides process monitoring, dynamic node priority variables and
user-defined events
• Cluster Manager is an RSCT (group services) client
• What is RSCT?
Reliable Scalable Cluster Technology (RSCT) is a set of software components that
together provide a comprehensive clustering environment for AIX and Linux. RSCT is
the infrastructure used by a variety of IBM products to provide clusters with improved
system availability, scalability, and ease of use.
• What RSCT provides to PowerHA:
- Failure detection and diagnosis for topology components (nodes, networks and
network adapters)
- Notification to the cluster manager of events which it has expressed an interest in -
primarily events related to the failure and recover of topology components
- Coordination of the recovery actions involved in dealing with the failure and recovery
of topology components (in other words, fallovers, fallbacks and dealing with
individual NIC failures by moving or swapping IP addresses)
- Monitoring of many different AIX components to provide:
• Application monitoring
• CPU, memory and disk I/O values for fallover with dynamic node priority
• User Defined Events
Recovery
Commands
RSCT RSCT
Group ~
Topology
Services Services PowerHA Event
topsvcs grpsvcs Scripts
Group Membership
heartbeats messages Event Subscription
Voting Protocols between nodes
• Topology Services
Builds heartbeat rings to detect, diagnose and report state changes to Group Services.
Topology Services also transmits RSCT-related messages between cluster nodes.
• Group Services
Group Services reports failures to the cluster manager as it becomes aware of them
from Topology Services. The cluster manager then drives cluster-wide coordinated
responses to the failure through the use of the Group Services voting protocols.
• RMC Monitors
These monitors report state changes related to monitored entities to the RMC Manager.
• RMC Manager
Receives notification from the monitors. Notifies RSCT clients of those events in which
they have expressed an interest. The PowerHA cluster manager is an RSCT client and
registers with both the RMC manager and Group Services.
• Cluster manager
The cluster manager responds to events through the use of PowerHA’s recovery
programs and event scripts. Running the scripts across the nodes in the cluster
coordinated via Group Services.
Uempty
Heartbeat Rings
en0 on each node en1 on each node
25.8.60.3 25.8.65.3
Uempty
Cluster Information Daemon (clinfo)
• An SNMP-aware client to the cluster manager
• Provides
– A cluster information API to the PowerHA SNMP
manager
•Focused on providing PowerHA cluster information
•Easier to work with than the SNMP APIs
– Support for ARP cache issues
• Is used by:
– The clstat command
– Customer written utility/monitoring tools
• Implemented as the clinfoES subsystem
Uempty
Shared External Disk Access
• Provides two types of shared disk support:
– Serial access resource group:
•Varied on by one node at a time under the control of PowerHA
•LVM or RSCT ensures no access by 2 nodes at once
•Two types of volume groups:
– Non-concurrent VG (standard VG)
– Enhanced Concurrent VG running in non-concurrent mode
– Concurrent access resource group:
•Used by concurrent applications writing to raw logical volumes
•Two types of volume groups:
– SSA Concurrent Mode VG
(no longer supported in HACMP 5.4 and later)
– Enhanced Concurrent VG running in concurrent mode
• bos.clvm.enh fileset required for Concurrent/Enhanced
Concurrent
Review 1 of 2
1. What is the first step in implementing a cluster?
a. Order the hardware
b. Plan the cluster
c. Install AIX and PowerHA
d. Install the applications
e. Take a long nap
2. True or False?
PowerHA 5.5 is compatible with any version of AIX v5.x or v6.x.
3. True or False?
Each cluster node must be rebooted after the PowerHA software is
installed.
Uempty
Review 2 of 2
4. Which component detects an adapter failure?
a. Cluster manager
b. RSCT
c. clcomd
d. clinfo
5. Which component provides SNMP information?
a. Cluster manager
b. RSCT
c. clsmuxpd
d. clinfo
6. Which component is required for clstat to work?
a. Cluster manager
b. RSCT
c. clcomd
d. clinfo
7. Which component removes requirement for the /.rhosts file?
a. Cluster manager
b. RSCT
c. clcomd
d. clinfo
UNIX Software Service Enablement
Unit Summary
• Before you install PowerHA, you should
– Plan your cluster
– Configure AIX: shared storage, networks and application start/stop scripts
• Install PowerHA software
– Determine which PowerHA filesets you will need and determine prerequisites
– Install software on cluster nodes (and optionally on clients (clinfo))
• PowerHA 5.5 includes the following major components
– Cluster Manager (clstrmgr)
• Heart of PowerHA. Responds to events (changes in cluster status) and takes action
(fallover, fallback, etc)
– Cluster Secure Communication Subsystem (clcomd)
• Provides secure node-to-node communications without /.rhosts
– RSCT and RMC
• Failure detection; notification to cluster manager; co-ordination of recovery
– SNMP monitoring programs
•clstrmgr is an smux peer, managing the PowerHA MIB for the SNMP daemon
• PowerHA status is available to clinfo and SNMP manager programs
– Cluster Information Program (clinfo)
• Provides an API to access PowerHA status via SNMP
• Supports clstat group of commands and customer written utilities
– Highly Available NFS Server
• Availability enhancements to NFS
– Shared External Disk Access
• Serial access shared disks (one node at a time)
• Concurrent access shared disks
References
http://www-03.ibm.com/systems/p/library/hacmp_docs.html
PowerHA manuals
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-1
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit Objectives
After completing this unit, you should be able to:
• Build a standby and mutual takeover PowerHA 5.5 cluster
– Use the Initialization and Standard Configuration path
(standard path)
– Use the Two-Node Cluster Configuration Assistant
(two-node assistant)
• Configure PowerHA topology to include:
– IP address takeover via alias
– Non-IP networks (rs232, diskhb)
– Persistent address
• Verify, synchronize, and test a cluster
• Start cluster services
• Save cluster configuration
• Objectives
This unit will show how to configure a two-node hot standby cluster, with a heartbeat
over disk non-IP network, using both the standard and extended path and the two-node
configuration assistant. It will then demonstrate how to start and stop PowerHA cluster
services. It will also show the additional steps necessary to build a mutual takeover
cluster configuration. To accomplish this, you will see how to modify the configuration of
the cluster by adding a resource group, adding a persistent IP label, adding a heartbeat
on disk non-IP network and synchronize the changes. The final step is making a
snapshot backup of the cluster configuration.
You will be walked through the methods of configuring the cluster, starting with the
standard path and using the extended path where necessary followed by the simplest,
most limited method, i.e., the Two-Node Configuration Assistant. While using the
Initialization and Standard Configuration path, you will also see what additional
resources are required to build a cluster configuration to support a mutual takeover.
Uempty
What We Are Going to Achieve
A two node standby configuration (active / passive)
– Resource group appAgroup with node1 as its home (primary) node
and node2 as its backup node
– Application server appA in appAgroup
Fallover
IP network
node2
node1
appA appA node2-per
node1-per
Y Y
non-IP network
heartbeat over disk
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-3
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
The Topology Configuration
• Here's the key portion of the /etc/hosts file that we'll be using in this unit:
192.168.1.1 node1-if1 # node1's first interface IP label
192.168.2.1 node1-if2 # node1's second interface IP label
192.168.3.1 node1-per # persistent node IP label on node1
192.168.1.2 node2-if1 # node2's first interface IP label
192.168.2.2 node2-if2 # node2's second interface IP label
192.168.3.2 node2-per # persistent node IP label on node2
192.168.3.10 appA-svc # the IP label for the application
# normally resident on node1
• Hostnames: node1, node2
• node1's network configuration (defined via smit chinet) :
en0 - 192.168.1.1
en1 - 192.168.2.1
• node2's network configuration:
en0 - 192.168.1.2
en1 - 192.168.2.2
• These network interfaces are all connected to the same
physical network
• The subnet mask is 255.255.255.0 on this network
• An enhanced concurrent mode volume group “vgA" has been created to
support the appA application and will be used for a non-IP heartbeat over
disk network
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-5
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Configuration Methods
• PowerHA provides two menu paths with three methods
– Initialization and Standard Configuration
•Two-node cluster configuration assistant
– Limited configuration
» Only 2 node hot standby cluster
– Builds cluster configuration based on AIX configuration
» All adapter addresses are used
» All volume groups assigned to one resource group
– Creates everything needed for simple cluster (Topology,
Resources, Resource Group)
» No persistent addresses
» Creates non-IP network using Heartbeat over Disk ( if enhanced concurrent
mode volume group present)
•Standard configuration
– Topology done in one step
– You then must configure resource groups and synchronize
– Desirable means to create more than 2 node hot standby cluster
– Extended Configuration
• More steps but provides access to all the options
• Needed for persistent address and non-IP network configuration
UNIX Software Service Enablement
Uempty
Planning and Base Configuration Review
• In AIX, configure boot (non-service) addresses on all systems
– Addresses on same system are in different subnets
• Plan service IP Label (appA-svc)
– IPAT via Aliases: Different subnet than the boot subnets
– IPAT via Replacement: Same subnet as one boot subnet
• Ensure /etc/hosts:
– Contains boot, service and persistent address for all systems
– Is the same file on all systems.
• Create the volume group(s) to be used by the application(s) (vgA)
– Enhanced Concurrent Mode Volume group(s) recommended
• Plan communication path to the nodes (node1-if1, node2-if1)
• Plan Application Server name (appA)
• Ensure Application Server start and stop scripts
– Exist on both nodes
– Are executable
• Network configuration
Configure boot addresses in AIX and ensure that all the addresses (interface, service
and persistent) are in the /etc/hosts files. If you do this first, then SMIT can discover the
IP addresses / IP labels and populate the SMIT pick lists for you.
• Storage configuration
In the previous exercise, you created an enhanced concurrent volume group prior to
configuring the cluster. Another approach: once you have created the cluster topology,
you can create any volume groups using the PowerHA SMIT menus and C-SPOC will
automatically create the volume group on all your cluster nodes in one step.
• Communication path (boot address)
Both the 2-node assistant and the Standard Path use one of the boot addresses on
each node to identify the nodes. In the 2-node assistant, you must identify the second
node using one of the boot addresses (it assume this system is the first node). In the
Standard Path, you must identify all the nodes using the boot address.
• Application server
You will need to name the application server and identify the start/stop scripts. The
scripts must be executable and in the some location on both nodes.
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-7
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Almost There…
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-9
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
The Standard Configuration Method
Configuration Assistants
Configure an HACMP Cluster and Nodes
Configure Resources to Make Highly Available
Configure HACMP Resource Groups
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-11
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
[Entry Fields]
* Cluster Name [qv120Cluster]
Uempty
What Did We Get?
•# /usr/es/sbin/cluster/utilities/cltopinfo
Cluster Name: qv120Cluster
Cluster Connection Authentication Mode: Standard
Cluster Message Authentication Mode: None
Cluster Message Encryption: None
Use Persistent Labels for Communication: No
There are 2 node(s) and 2 network(s) defined
NODE node1:
Network net_ether_01
node1-if1 192.168.1.1
node1-if2 192.168.2.1
NODE node2:
Network net_ether_01
node2-if1 192.168.1.2
node2-if2 192.168.2.2
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-13
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Configuration Assistants
Configure an HACMP Cluster and Nodes
Configure Resources to Make Highly Available
Configure HACMP Resource Groups
Uempty
Configure Service Addresses
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-15
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Add appA-svc Service Address (1 of 3)
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-17
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
[Entry Fields]
* IP Label/Address [appA-svc] +
Prefix Length [] #
* Network Name [] +
ňņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņʼn
Ň CONTEXTUAL HELP Ň
Ň Ň
Ň Press Enter or Cancel to return to the application. Ň
Ň Ň
Ň Enter the prefix length for the service IP. For IPv4 service IP Ň
Ň the prefix length must, either, concur with the prefix length of the Ň
Ň underlying HACMP network, or left blank for its auto-population. Ň
Ň However, for IPv6 service IP, the user must supply appropriate prefix Ň
Ň length (ranging from 1 to 128). Ň
Ň Ň
F1=Help Ň F1=Help F2=Refresh F3=Cancel Ň
F5=Reset Ň F8=Image Enter=Do Ň
F9=Shell Ŋņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņŋ
Uempty
Add appA-svc Service Address (3 of 3)
Add a Service IP Label/Address (standard)
[Entry Fields]
* IP Label/Address [appA-svc] +
Prefix Length [] #
* Network Name [net_ether_02] +
ňņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņʼn
Ň Network Name Ň
Ň Ň
Ň Move cursor to desired item and press Enter. Ň
Ň Ň
Ň net_ether_01 (9.47.87.0/24) Ň
Ň net_ether_02 (192.168.2.0/24 192.168.1.0/24) Ň
Ň Ň
Ň F1=Help F2=Refresh F3=Cancel Ň
F1=Help Ň F8=Image F10=Exit Enter=Do Ň
F5=Reset Ň /=Find n=Find Next Ň
F9=Shell Ŋņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņŋ
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-19
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
• Application server
Continuing to follow the menus, the next step is to define the application server(s).
Uempty
Add appA Application Server (1 of 2)
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-21
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Configure Volume Groups (Optional)
• Shared volume groups can be created before you start PowerHA
configuration
• Once you have created the cluster topology, you can use
C-SPOC to create the volume groups and import them on all nodes
in the resource group
• If you choose to, follow the menus…
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-23
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Discovery
• Information for SMIT pick lists is created by the Discovery process
• Discovery is initially run when you create the cluster topology
• Discovery should be re-run when you make changes to the network
or LVM configuration
• After creating volume groups using C-SPOC:
– Prior to PowerHA 5.5: Run discovery to make VGs available in pick lists
– PowerHA 5.5: Discovery is run automatically
• Use Extended Configuration menu to run Discovery
Extended Configuration
Uempty
Adding the appAgroup Resource Group
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-25
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Adding Resources to appAgroup (1 of 2)
Configure HACMP Resource Groups
Move cursor to desired item and press Enter.
Add a Resource Group
Change/Show a Resource Group
Remove a Resource Group
Change/Show Resources for a Resource Group (standard)
+--------------------------------------------------------------------+
¦ Select a Resource Group ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ appAgroup ¦
| appBgroup ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F8=Image F10=Exit Enter=Do ¦
¦ /=Find n=Find Next ¦
+--------------------------------------------------------------------+
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-27
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Synchronize Your Changes
Initialization and Standard Configuration
Move cursor to desired item and press Enter.
Configuration Assistants
Configure an HACMP Cluster and Nodes
Configure Resources to Make Highly Available
Configure HACMP Resource Groups
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-29
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
What Do We Have at This Point?
• We DO have a cluster configured with the following:
– Two nodes defined
– One network defined with interface and service addresses
(Actually we have two networks:
net_ether_01 (en0) and net_ether_02 (en1 and en2) )
– Application server objects defined containing the start and stop
scripts for all the applications to be made highly available
– Volume groups defined to contain the data for the applications
– Resource group(s) with node priority, the service label(s),
application server(s) and volume group(s) for the applications
– All this has been synchronized
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-31
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Extended Topology Configuration Menu
Extended Topology Configuration
Move cursor to desired item and press Enter.
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-33
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Defining a Non-IP Network (1 of 3)
Configure HACMP Communication Interfaces/Devices
Move cursor to desired item and press Enter.
Add Communication Interfaces/Devices
Change/Show Communication Interfaces/Devices
Remove Communication Interfaces/Devices
Update HACMP Communication Interface with Operating System
Settings
+--------------------------------------------------------------------------+
¦ Select a category ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ Add Discovered Communication Interface and Devices ¦
¦ Add Predefined Communication Interfaces and Devices ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F8=Image F10=Exit Enter=Do ¦
F1¦ /=Find n=Find Next ¦
F9+--------------------------------------------------------------------------+
Don't risk a potentially catastrophic partitioned cluster by using cheap rs232 cables!
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-35
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
+--------------------------------------------------------------------------+
¦ Select a category ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ # Discovery last performed: (Feb 12 18:20) ¦
¦ Communication Interfaces ¦
¦ Communication Devices ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F8=Image F10=Exit Enter=Do ¦
F1¦ /=Find n=Find Next ¦
F9+--------------------------------------------------------------------------+
• Is it an interface or a device?
Now we need to indicate whether we are adding a communication interface or a
communication device. Non-IP networks use communication devices as end-points
(/dev/tty for example) so select Communication Devices to continue.
Uempty
Defining a Non-IP Network (3 of 3)
Press Enter and HACMP defines a new non-IP network with these communication
devices. Choose the hdisks for the Heartbeat over Disk network too.
Configure HACMP Communication Interfaces/Devices
Move cursor to desired item and press Enter.
Add Communication Interfaces/Devices
Change/Show Communication Interfaces/Devices
Remove Communication Interfaces/Devices
+--------------------------------------------------------------------------+
¦ Select Point-to-Point Pair of Discovered Communication Devices to Add ¦
¦ ¦
¦ Move cursor to desired item and press F7. Use arrow keys to scroll. ¦
¦ ONE OR MORE items can be selected. ¦
¦ Press Enter AFTER making all selections. ¦
¦ ¦
¦ # Node Device Device Path Pvid ¦
¦ node1 hdisk1 /dev/hdisk1 000b4a7cd1.. ¦
¦ node2 hdisk1 /dev/hdisk1 000b4a7cd1.. ¦
¦> node1 tty1 /dev/tty1 ¦
¦> node2 tty1 /dev/tty1 ¦
¦ node1 tmssa1 /dev/tmssa1 ¦
¦ node2 tmssa2 /dev/tmssa2 ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F7=Select F8=Image F10=Exit ¦
F1¦ Enter=Do /=Find n=Find Next ¦
F9+--------------------------------------------------------------------------+
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-37
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
ňņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņʼn
Ň Select a Network to Remove Ň
Ň Ň
Ň Move cursor to desired item and press Enter. Ň
Ň Ň
Ň net_diskhb_01 Ň
Ň net_ether_01 (9.47.87.0/24) Ň
Ň net_ether_02 (192.168.3.0/24 192.168.2.0/24 192.168.1.0/24) Ň
Ň Ň
Ň F1=Help F2=Refresh F3=Cancel Ň
Ň F8=Image F10=Exit Enter=Do Ň
F1=Help Ň /=Find n=Find Next Ň
F9=Shell Ŋņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņņŋ
• Remove a network
When you use the Standard Configuration path to create the cluster, all interfaces
found will be included in the PowerHA topology. If there are some interfaces that you
don’t want PowerHA to use, you can remove them from the configuration.
In our lab environment, we have an extra interface (en0) on each node that we use for
access to the cluster. Since en0 is not on the same physical network as en1 and en2,
PowerHA creates a separate network for it. Since it is a single adapter network,
PowerHA will give warnings about this network being a single point of failure every time
we synchronize the cluster. Further, since we are not actually using this network in
PowerHA, we can remove it from the PowerHA configuration and get rid of the
warnings.
Uempty
Defining Persistent Node IP Labels (1 of 3)
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-39
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
+--------------------------------------------------------------------------+
¦ Select a Node ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ node1 ¦
¦ node2 ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F8=Image F10=Exit Enter=Do ¦
¦ /=Find n=Find Next ¦
+--------------------------------------------------------------------------+
Uempty
Defining Persistent Node IP Labels (3 of 3)
Press Enter and then repeat for the node2 persistent IP label.
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-41
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Synchronize
smitty hacmp -> Extended Configuration
[Entry Fields]
* Verify, Synchronize or Both [Both] +
* Automatically correct errors found during [No] +
verification?
Uempty
Save Configuration: Snapshot
Snapshot Configuration
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-43
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Snapshot Configuration
Uempty
Two-Node Cluster Configuration Assistant
Two-Node Cluster Configuration Assistant
Type or select values in entry fields.
Press Enter AFTER making all desired changes.
[Entry Fields]
* Communication Path to Takeover Node [node2-if1] +
* Application Server Name [appA]
* Application Server Start Script [/usr/HTTPServer/bin/apachectl start]
* Application Server Stop Script [/usr/HTTPServer/bin/apachectl stop]
* Service IP Label [appA-svc] +
Prefix Length [] #
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-45
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Where Are We in the Implementation?
9Plan for network, storage, and application
– Eliminate single points of failure
9Define and configure the AIX environment
– Storage (adapters, LVM volume group, filesystem)
– Networks (IP interfaces, /etc/hosts, non-IP)
– Application start and stop scripts
9Install the HACMP filesets and reboot
9 Configure the HACMP environment
– Topology
•Cluster, node names, HACMP IP and non-IP networks
– Resources, resource group, attributes:
•Resources: Application Server, service label
•Resource group: Identify name, nodes, policies
•Attributes: Application Server, service label, VG, filesystem
– Synchronize
¾Start HACMP
• Test configuration
• Save configuration
UNIX Software Service Enablement
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-47
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Starting Cluster Services (2 of 4)
System Management (C-SPOC)
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-49
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Starting Cluster Services (4 of 4)
# smit clstart
Start Cluster Services
Type or select values in entry fields.
Press Enter AFTER making all desired changes.
[Entry Fields]
* Start now, on system restart or both now +
Start Cluster Services on these nodes [node1,node2] +
* Manage Resource Groups Automatically +
BROADCAST message at startup? true +
Startup Cluster Information Daemon? true +
Ignore verification errors? false +
Automatically correct errors found during Interactively +
cluster start?
• Startup choices
There are a few choices to make. For the moment we will just recommend the defaults
except selecting both nodes and turning on the Cluster Information Daemon.
• Remember the fast path
Notice the smit clstart fastpath. This is often much faster than working your way
through the menu tree.
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-51
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Removing a Cluster
•Stop cluster services (smitty cl_stop)
•Remove cluster using Extended Topology Configuration on first node
•Repeat on second node
•Remove contents of the cluster rhosts file - on both nodes
# > /usr/es/sbin/cluster/etc/rhosts
• Starting over
If you have to start over, you can:
- Stop cluster services on all nodes
- Use Extended Configuration, as shown above to remove the cluster from this
node
Note: This must be done separately on each node. (Once you’ve removed the
cluster configuration from a node, it is no longer in the cluster so you cannot
synchronize the change to the other nodes.)
- Remove the entries (but not the file) from /usr/es/sbin/cluster/etc/rhosts (on all
nodes)
Uempty
We're There!
• We’ve configured a two-node hot standby cluster containing:
– A RG(X) with node1 (primary), application, service address, VG
– A non-IP network, persistent address
• We’ve shown how to extend to a two-node mutual takeover cluster
– Add RG(Y) with node2 (primary), application, service address, VG
– Add persistent address
node1 node2
RG(X) X X
Y Y RG(Y)
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-53
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Review
1. In which of the top level HACMP menu choices is the menu for starting and
stopping cluster nodes?
a. Initialization and Standard Configuration
b. Extended Configuration
c. System Management (C-SPOC)
d. Problem Determination Tools
2. In which of the top level HACMP menu choices is the menu for defining a non-IP
heartbeat network?
a. Initialization and Standard Configuration
b. Extended Configuration
c. System Management (C-SPOC)
d. Problem Determination Tools
3. True or False?
It is possible to configure HACMP faster by having someone help you on the
other node.
4. True or False?
You must always specify exactly which filesystems you want mounted when
you put resources into a resource group.
Uempty
Exercise
Configure PowerHA
Notes:
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-55
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit Summary 1 of 2
• The steps to implement an HACMP cluster are:
– Plan
– Configure AIX for HACMP (networks, storage, applications)
– Install HACMP software and reboot
– Configure PowerHA
(topology, resources, resource groups, attributes)
– Synchronize
– Start PowerHA services
– Test
• There are two menu paths used for PowerHA configuration
– Initialization and Standard Configuration
• Two-Node Cluster Configuration Assistant
– Automates process, but you don't see the details
– Simple two-node, hot-standby, one resource group configuration
– Automatic synchronization
• Standard configuration
– Preferred method for most initial cluster configuration, forces you to see the details of your
configuration
– Includes the most commonly used options
– Synchronization must be done manually
– HACMP Cluster Test Tool
– Extended Configuration
• More steps, but provides access to all options
• Required for node persistent address and non-IP network configuration
Uempty
Unit Summary 2 of 2
• Pick one node as your administration node and make all changes
from that node
• Start HACMP services
– Cluster Management (C-SPOC) -> Manage HACMP Services ->
Start Cluster Services
• Removing a cluster (if you have to start over)
– Stop cluster services
– Remove the cluster
Extended Configuration -> Extended Topology Configuration
-> Configure an HACMP Cluster -> Remove an HACMP Cluster
– Remove the entries (but not the file) from /usr/es/sbin/cluster/etc/rhosts on
all nodes
© Copyright IBM Corp. 2007, 2009 Unit 5. Initial PowerHA Configuration 5-57
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Review (1 of 2) Solutions
Unit 1 - 2 of 2
Review (2 of 2) Solutions
4. True or False?
Resource Groups may be moved from node to node.
5. True or False?
PowerHA/XD is a solution for building geographically
distributed clusters.
6. Which of the following capabilities does PowerHA not
provide (select all that apply)?:
a. Time synchronization.
b. Automatic recovery from node and network adapter
failure.
c. System Administration tasks unique to each node.
Backup and restoration.
d. Fallover of just a single resource group.
AP Unit 2- 1 of 4
Review 1 of 4 Solutions
1. How does PowerHA use networks (select all which apply)?
a. Provide client systems with highly available access to the cluster's
applications
b. Detect and diagnose failures
c. Provide cluster status
d. Communicate between cluster nodes
e. Monitor network performance
2. Using information from RSCT, PowerHA only directly handles
three types of failures: Network interface card (NIC) failures,
Node failures, Network failures.
3. True or False?
• Heartbeat packets must be acknowledged or a failure is assumed to have
occurred.
4. True or False?
• Clusters should include a non-IP network.
Unit 2 - 2 of 4
Review 2 of 4 Solutions
5. True or False?
• Clusters must always be configured with a private IP network for PowerHA
communication.
6. Which of the following are true statements about communication
interfaces (select all that apply)?
a. Has an IP address assigned to it using the AIX TCP/IP SMIT screens
b. Might have more than one IP address associated with it
c. Sometimes but not always used to communicate with clients
d. Always used to communicate with clients
(Communication interfaces on private IP networks are not intended to be
used by clients.)
7. True or False?
• Persistent node IP labels are not supported for IPAT via IP replacement.
8. True or False?
• Each NIC on each physical IP network on each node is required to have an
IP address on a different logical subnet (unless using heartbeat over IP
alias).
9. True or False?
• A single cluster can use both IPAT via IP aliasing and IPAT via IP
replacement.
UNIX Software Service Enablement
AP Unit 2 - 3 of 4
Review 3 of 4 Solutions
10. True or False?
• All networking technologies supported by PowerHA support IPAT via IP
aliasing.
11. True or False?
• All networking technologies supported by PowerHA support IPAT via IP
replacement.
12. If node1 has NICs with the IP addresses 192.168.20.1 and
192.168.21.1 and node2 has NICs with the IP addresses
192.168.20.2 and 192.168.21.2, then which of the following are
valid service IP addresses if IPAT via IP aliasing is being used?
(select all that apply)
a. (192.168.20.3 and 192.168.20.4) OR (192.168.21.3 and 192.168.21.4)
b. 192.168.20.3 and 192.168.20.4 and 192.168.21.3 and 192.168.21.4
c. 192.168.22.3 and 192.168.22.4
d. 192.168.23.3 and 192.168.24.3
Unit 2- 4 of 4
Review 4 of 4 Solutions
13. If node1 has NICs with the IP addresses 192.168.20.1 and
192.168.21.1 and node2 has NICs with the IP addresses
192.168.20.2 and 192.168.21.2 then which of the following are
valid service IP addresses if IPAT via IP replacement is being
used (select all that apply)?
a. (192.168.20.3 and 192.168.20.4) OR (192.168.21.3 and 192.168.21.4)
b. 192.168.20.3, 192.168.20.4, 192.168.21.3 and 192.168.21.4
c. 192.168.22.3 and 192.168.22.4
d. 192.168.23.3 and 192.168.24.3
14. True or False?
– Clients are required to exit and restart their application after a fallover.
15. True or False?
– All client systems are potentially directly affected by the ARP cache issue.
16. True or False?
– clinfo must not be run both on the cluster nodes and on the client
systems.
AP Unit 3 -
Review Solutions
1.True or False
Applications are defined to PowerHA in a configuration file that lists
what binary to use.
2.What policies would be the best to use for a 2-node “active-active” cluster
using IPAT to minimize both applications running on the same node?
a. home, next, never
b. first, next, higher
c. distribution, next, never
d. all, error, never
e. home, next, higher
3.Which type of data should not be placed in private data storage?
a. Application log data
b. License file
c. Configuration files
d. Application binaries
4.True or False
Start and stop scripts which are tested on one node can be counted on
to perform correctly under PowerHA. No further testing is needed.
Unit 4- 1 of 2
Review 1 of 2 Solutions
1. What is the first step in implementing a cluster?*
a. Order the hardware
b. Plan the cluster
c. Install AIX and PowerHA
d. Install the applications
e. Take a long nap
*There is some dispute about whether the correct answer is b or e -.
A disconcerting number of clusters are implemented in the order a, b, c, d, e (how can
you possibly order the hardware if you do not yet know what you are going to build?) or
even just a, c, d (cluster implementers who skip step b rarely have time for long naps).
2. True or False?
PowerHA 5.5 is compatible with any version of AIX v5.x or v6.x.
Only AIX 5.3 and 6.1.
3. True or False?
Each cluster node must be rebooted after the PowerHA software is
installed.
AP Unit 4- 2 of 2
Review 2 of 2 Solutions
4. Which component detects an adapter failure?
a. Cluster manager
b. RSCT
c. clcomd
d. clinfo
5. Which component provides SNMP information?
a. Cluster manager
b. RSCT
c. clsmuxpd
d. clinfo
6. Which component is required for clstat to work?
a. Cluster manager
b. RSCT
c. clcomd
d. clinfo
7. Which component removes requirement for the /.rhosts file?
a. Cluster manager
b. RSCT
c. clcomd
d. clinfo
UNIX Software Service Enablement
Unit 5-
Review Solutions
1. In which of the top level HACMP menu choices is the menu for starting and
stopping cluster nodes?
a. Initialization and Standard Configuration
b. Extended Configuration
c. System Management (C-SPOC)
d. Problem Determination Tools
2. In which of the top level HACMP menu choices is the menu for defining a non-IP
heartbeat network?
a. Initialization and Standard Configuration
b. Extended Configuration
c. System Management (C-SPOC)
d. Problem Determination Tools
3. True or False?
It is possible to configure HACMP faster by having someone help you on the
other node.
4. True or False?
You must always specify exactly which filesystems you want mounted when
you put resources into a resource group.
AP Appendix C-
Review Solutions
1. For IPAT via replacement (select all that apply)
a. Each service IP address must be in the same subnet as one of the non-service
addresses
b. Each service IP address must be in the same subnet
c. Each service IP address cannot be in any non-service address subnet
2. True or False?
If the takeover node is not the home node for the resource group and the
resource group does not have a Startup policy of Online Using Distribution
Policy, the service IP address replaces the IP address of a NIC with an IP
address in the same subnet as the subnet of the service IP address.
3. True or False?
In order to use HWAT, you must enable and complete the ALTERNATE
ETHERNET address field in the SMIT devices menu.
4. True or False?
You must stop the cluster in order to change from IPAT via aliasing to IPAT via
replacement.
o Software prerequisites
o Configuration limitations
o Notes on functionality
o Documentation
o Feedback
======================================================
Enhancements and new features included in this release
======================================================
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-1
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
o Support for additional configurations with HACMP/XD SVC PPRC and VIO
See the release_notes_xd file for more details.
o WebSMIT Gateway
WebSMIT requires an HTTP server to enable a browser to connect
to and manage the HACMP cluster nodes. In prior releases the HTTP
server had to run on one of the cluster nodes which created
security concerns for some customers.
The WebSMIT gateway feature lets you run the HTTP server on a machine
outside of the cluster to be managed by WebSMIT. A simple registration
process is required to specify the cluster nodes to be managed
by the gateway.
AP IPv6 will require AIX 6.1 on all nodes. Please see the section
on Installation prerequisites for specific information.
=========================================
Installation and Migration Considerations
=========================================
o When migrating from a prior release you have several options to choose
from based on your overall upgrade plan and downtime requirements.
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-3
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
restart, and repeat for all nodes in the cluster. Although there
is an outage while the application moves between nodes, this is less
disruptive than the options above. If you are migrating from HACMP 5.3
or later you can use rolling migration.
==============================================
Software prerequisites
==============================================
==============================================
Configuration Limitations
==============================================
======================
Notes on Functionality
======================
-------------------------------------------------------------------------
Enable grace periods and restart nfsd after installing cluster.es.nfs.rte
-------------------------------------------------------------------------
This note applies only to 64 bit systems. You may ignore this note if all
of your cluster nodes are 32 bit.
The NFS daemon nfsd must be restarted on each cluster node with grace periods
enabled after installing cluster.es.nfs.rte before configuring NFSv4 exports.
This step is required otherwise NFSv4 exports will fail to export with the
misleading error message
The following commands enable grace periods and restart the NFS daemon.
chnfs -I -g on -x
stopsrc -s nfsd
startsrc -s nfsd
Please note that this will impact the availability of all exported filesystems
on the machine, therefore the best time to perform this step is when all
resource groups with NFS exports are offline or failed over to another node
in the resource group.
o SNMP
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-5
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
o COD
- When changing a resource group and reducing the number of cpu’s or memory,
the change does not take effect immediately when the dare event runs, it
only affects the next failover (where CuOD is used).
o WPAR naming
o HAGEO
o eRCMF
o Event emulation
AP stopsrc -s clcomdES
startsrc -s clcomdES
o WebSMIT
- The new gateway function cannot be used with nodes running prior releases
of HACMP. This support will require a change in a service pack for the
earlier releases.
- When registering multiple clusters for use with the Enterprise view, each
cluster must have a unique cluster ID.
- The gateway can only register a configured cluster - you cannot use
WebSMIT through the gateway to configure a cluster from scratch (though
you can do so with WebSMIT without the gateway).
IBM intends to lift these restrictions in upcoming service packs.
===================================
Documentation
===================================
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-7
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Online Documentation:
------------------------------------------------------
Viewing and installing the documentation files
------------------------------------------------------
You can view the PDF documentation without installing the product.
Simply copy the PDFs to your desired location or view them directly from
the media.
If you prefer you can install the documentation by following these steps:
2. Select the CD ROM drive from the picklist and press Enter.
Image cluster.doc.en_US.es
--------------------------
cluster.doc.en_US.es.html HACMP Web-based HTML
Documentation - U.S. English
cluster.doc.en_US.es.pdf HACMP PDF
Documentation - U.S. English
Image cluster.doc.en_US.glvm
----------------------------
cluster.doc.en_US.glvm.html HACMP GLVM HTML Documentation
- U.S. English
cluster.doc.en_US.glvm.pdf HACMP GLVM PDF Documentation
- U.S. English
Image cluster.doc.en_US.pprc
----------------------------
cluster.doc.en_US.pprc.html PPRC Web-based HTML
Documentation - U.S. English
cluster.doc.en_US.pprc.pdf PPRC PDF Documentation
- U.S. English
Image cluster.doc.en_US.assist
------------------------------
cluster.doc.en_US.assist.db2.html HACMP Smart Assist for
DB2 HTML Documentation
- U.S. English
cluster.doc.en_US.assist.db2.pdf HACMP Smart Assist for
DB2 PDF Documentation
- U.S. English
cluster.doc.en_US.assist.oracle.html HACMP Smart Assist for
Oracle HTML Documentation
- U.S. English
cluster.doc.en_US.assist.oracle.pdf HACMP Smart Assist for
Oracle PDF Documentation
- U.S. English
cluster.doc.en_US.assist.websphere.html HACMP Smart Assist for
WebSphere HTML
Documentation
- U.S. English
cluster.doc.en_US.assist.websphere.pdf HACMP Smart Assist for
WebSphere PDF
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-9
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Documentation
- U.S. English
5. Select all filesets that you wish to install and execute the
command.
/usr/share/man/info/en_US/cluster/HAES
o HTML documentation
==================
Product Man Pages
==================
Man pages for HACMP commands and utilities are installed in the
following directory:
/usr/share/man/cat1
Use
man [command-name]
========================
Accessing IBM on the Web
========================
AP http://www.ibm.com
Miscellaneous pages:
http://www.ibm.com/systems/p/ha/faq3.html
http://www.ibm.com/systems/p/library/hacmp_docs.html
========
Feedback
========
IBM welcomes your comments. You can send comments and questions via e-mail to:
hafeedbk@us.ibm.com
README5.5.0.UPDATE
===========================================
README for HACMP Version 5.5 Service Pack 2
===========================================
====================================================================
o APAR IZ31205: Asynchronous mirroring option for GLVM
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-11
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
====================================================================
Synchronous replication
With this mode, application writes do not complete until the data
has been written at both sites. Network bandwidth and latency
affect the application performance.
Asynchronous replication
The asynchronous replication option decouples the application
performance from the network. Remote writes are cached locally
and completed asynchronously, which can hide peeks in load
(until the cache fills). However, data can be lost on unplanned
site fallover.
====================================================================
o APAR IZ49934: varyonvg failed to bring up the async Vg on remote site
====================================================================
If there are stale partitions on the recovery node, customers may see
varyonvg command failures in hacmp.out. To address this problem,
retreive and install APAR IZ49934. This applies to Async GLVM (and
hence AIX 61) only.
====================================================================
o APAR IZ41204: Enabling Internet MIB tree for clstat or cldump to work
====================================================================
clstat or cldump will not start if the internet MIB tree is not
enabled in snmpdv3.conf file. This behavior is usually seen in
AIX 6.1 onwards where this internet MIB entry was intentionally
disabled as a security issue. This internet MIB entry is required to
There are two ways to enable this MIB sub tree(risc6000clsmuxpd) they
are:
1) Enable the main internet MIB entry by adding this line in
/etc/snmpdv3.conf file
VACM_VIEW defaultView internet - included -
But doing so is not advisable as it unlocks the entire MIB tree
2) Enable only the MIB sub tree for risc6000clsmuxpd without enabling
the main MIB tree by adding this line in /etc/snmpdv3.conf file
VACM_VIEW defaultView 1.3.6.1.4.1.2.3.1.2.1.5 - included -
Note: After enabling the MIB entry above snmp daemon must be restarted
with the following commands as shown below:
1) stopsrc -s snmpd
2) startsrc -s snmpd
After snmp is restarted leave the daemon running for about two minutes
before attempting to start clstat or cldump.
====================================================================
o APAR IZ37766: Loadable Password Algorithm
====================================================================
HACMP cluster-wide C-SPOC password administration does not support use
of the feature allowing passwords longer than 8 characters which became
available with the Loadable Password Algorithm as part of AIX 53 TL 7.
====================================================================
o APAR IZ01331: New options for netmon.cf for VIO environment
====================================================================
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-13
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
The netmon.cf file can be configured to help avoid false adapter death
events, especially in single adapter networks. RSCT will attempt to ping
the addresses in the file in an attempt to differentiate between a local
adapter failure and a network failure. In prior releases, all addresses
were used, regardless of the adapter which was considered to have failed
or which address responded to the ping request.
Parameters:
----------
!REQD : An explicit string; it *must* be at the
beginning of the line (no leading spaces).
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-15
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
9.12.4.11
!REQD en2 100.12.7.9
9.12.4.13
!REQD en2 100.12.7.10
AP
-- most adapters will use netmon in the traditional
manner, pinging 9.12.4.11 and 9.12.4.13 along with
other local adapters or known remote adapters, and
will only care about the interface’s inbound byte
count for results.
-- interface en2 will only be considered up if it
can ping either 100.12.7.9 or 100.12.7.10.
Important notes:
---------------
This APAR will only take effect *if* valid updates in
this new format are made to the netmon.cf file. As long
as you only use the netmon.cf file in the traditional
manner (or do not use it at all), then you can safely
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-17
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
For this same reason, customers who find they must take
advantage of this new functionality are encouraged to be
as generous as possible with the number of targets they
provide for each interface (up to the limit).
====================================================================
=
o APAR IZ43939: NON EXISTING SMIT PANELS FOR WPARS REFERENCED IN ADMIN
GUIDE
====================================================================
=
There is WPAR related information in the Administration Guide under the
AP topic “Using HACMP Workload Partitions” which runs from page 503 to 507.
This information is inaccurate and should be ignored.
====================================================================
Multi-node disk heartbeat.
====================================================================
o New features
o Prerequisite Software
o MNDHB Limitations
o Best Practices
==========================================
New features
==========================================
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-19
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
====================================
Installation Notes
====================================
When the requisite software is installed on all nodes you are now ready
to configure the multi-node disk heartbeat network(s). As with any
configuration change made with HACMP, you will do the configuration from a
single cluster node then verify and synchronize the configuration across the
cluster. The new network will become active after the synchronization.
==============================================
MNDHB Configuration and Usage
==============================================
o Overview
To begin using multi-node disk heartbeat and disk fencing you will
1) Plan the configuration including the disks, volume groups, and
networks.
2) Configure the networks using the MNDHB smit panels.
3) Configure the disk fencing options.
4) Verify and Synchronize the Changes.
5) View the configuration and monitor the network status.
During the configuration process HACMP will identify all the available
logical volumes which meet these requirements (and tell you which
volumes were found which did not meet the requirements and why).
If you do not have any logical volume available which meet these
requirements, the MNDHB smit options can be used to create them
for you. You can also use the CSPOC smit options to create the logical
volumes, however it is recommended to use the MNDHB smit paths to
ensure the logical volumes are created correctly.
Using the MNDHB smit paths will also create or modify the volume groups
and resource groups which will contain the logical volume used for the
MNDHB networks. You can manage the volume groups and resource groups
using the standard smit paths, however, using the MNDHB smit paths
will ensure all the correct configuration options are used.
To configure and use MNDHB networks use the SMIT fastpath: cl_manage_mndhb
smit hacmp
System Management (C-SPOC)
HACMP Concurrent Logical Volume Management
Concurrent Volume Groups (smitty cl_convg)
Manage Concurrent Volume Groups for Multi-Node Disk Heartbeat
smit hacmp
Extended Configuration
Extended Topology Configuration
Configure HACMP Networks
Manage Concurrent Volume Groups for Multi-Node Disk Heartbeat
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-21
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Note that multi-node disk heartbeat networks cannot be managed using the
traditional HACMP smit paths for adding or changing networks or adapters -
only the MNDHB smit paths should be used for managing these networks.
The initial smit menu consists of these options which are described
in detail below.
Use an existing Logical Volume and Volume Group for Multi-Node Disk HB
Create a new Volume Group and Logical Volume for Multi-Node Disk HB
Add a Concurrent Logical Volume for Multi-Node Disk Heartbeat
Show Volume Groups in use for Multi-Node Disk Heartbeat
Stop using a Volume Group for Multi-Node Disk Heartbeat
Configure failure action for Multi-Node Disk Heartbeat Volume Groups
Use an existing Logical Volume and Volume Group for Multi-Node Disk HB
SMIT fastpath: cl_enable_mndhb_vg
====================================================================
This menu option allows a user to use an existing Volume Group and
Logical Volume for a new MNDHB network. For example, an Oracle
configuration could have additional shared Logical Volumes in the
Volume Group that contains the Oracle voting disks, which could be used
for MNDHB.
HACMP will check the existing configuration and present a list of resource
groups that have concurrent volume groups associated with them.
+--------------------------------------------------------------------------
+
¦ Select a Volume Group for Multi-Node Disk Heartbeat ¦
¦ ¦
¦ Move cursor to desired item and press Enter. Use arrow keys to scroll.
¦
¦ ¦
¦ #Resource Group Volume Group ¦
¦ shared_storage ccvg ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ Esc+8=Image Esc+0=Exit Enter=Do ¦
¦ /=Find n=Find Next ¦
AP
+--------------------------------------------------------------------------
+
Choose the Volume Group you wish to add the Logical Volume to.
You will be presented with a picklist containing the logical volumes in
that Volume Group which are candidates for MNDHB. If any volumes are found
which do not meet requirements, they will be listed along with the reason
they are not valid candidates.
+--------------------------------------------------------------------------
+
¦ Select Logical Volume to use for Heartbeat ¦
¦ ¦
¦ Move cursor to desired item and press Enter. Use arrow keys to scroll.
¦
¦ ¦
¦ [MORE...39] ¦
¦ # logical volume data_lv1 allowed to span more than one disk (-u
¦
¦ # logical volume data_lv2 allowed to span more than one disk (-u s
¦
¦ # logical volume mnhb_lv2 is already used for multi-node disk hear
¦
¦ # logical volume mnhb_lv3 is already used for multi-node disk hear
¦
¦ mndhb_lv1 ¦
¦ [BOTTOM] ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ Esc+8=Image Esc+0=Exit Enter=Do ¦
¦ /=Find n=Find Next ¦
+--------------------------------------------------------------------------
+
After selecting the logical volume you will proceed to the next menu:
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-23
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
[Entry Fields]
Existing Logical Volume to reserve for heartbeat [mndhb_lv_02] +
VOLUME GROUP name mndhb_vg_01
Network Name [net_diskhbmulti_02] +
Resource Group Name rg_diskhbmulti_02 +
Node Names racb1,racb2,racb3,rac>
After configuring the Logical Volume to use for MNDHB, you must
perform a verification and synchronization operation to activate the new
MNDHB.
Create a new Volume Group and Logical Volume for Multi-Node Disk Heartbeat
SMIT fastpath: cl_add_mndhb_vg
====================================================================
If you do not already have a volume group and shared logical volume for use
with MNDHB, this menu option can used to create the MNDHB network and the
associated volume group and resource group configuration.
The first step is to choose a physical volume from the pick list:
+--------------------------------------------------------------------------
+
¦ Physical Volume to reserve for Heartbeat ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ [TOP] ¦
+--------------------------------------------------------------------------
+
Note that the same physical disk should appear on all nodes in the cluster.
Pick one instance of the physical disk you would like to use for
MNDHB. You will then be presented with an options page:
[Entry Fields]
Volume Group Name [mndhb_vg_06] +
Volume group MAJOR NUMBER [49] +#
PVID for Logical Volume 00c1890c57284929
Logical Volume Name [mndhb_lv_010] +
Physical partition SIZE in megabytes 4 +
Node Names racc1,racc2,racc3>
Resource Group Name [rg_diskhbmulti_04] +
Network Name [net_diskhbmulti_02] +
Warning:
Changing the volume group major number may result
in the command being unable to execute
successfully on a node that does not have the
major number currently available. Please check
for a commonly available major number on all nodes
before changing this setting.
In most cases you may keep the default parameters. The Resource Group,
Volume Group and Logical Volume names generated are unique across the
cluster, however you can change them to conform to your naming convention.
Volume Group, Logical Volume, and network names must be unique and
not already in use. You can use an existing concurrent resource group
or let HACMP create a new one for you.
This operation will create a new concurrent Volume Group (and resource
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-25
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
group, if you did not select an existing group). The Volume Group will
be imported on all nodes and a new Logical Volume will be created within
that Volume Group.
[Entry Fields]
Resource Group Name rg_diskhbmulti_04
Participating Nodes (Default Node Priority) racc1 racc2 racc3>
Note: After making these changes you must perform a verification and
synchronization operation to activate the new MNDHB.
This menu option is used to add a logical volume and MNDHB network to an
existing Volume Group. The new MNDHB network can be added to a new
physical volume (which is added to the existing Volume Group), or can
be created on an existing physical volume in the Volume Group as long
as no other MNDHB networks are using that physical volume.
A picklist will let you select the Volume Group you wish to use:
+--------------------------------------------------------------------------
+
+--------------------------------------------------------------------------
+
Note that there can be multiple Volume Groups associated with a single
Resource Group. After selecting the Volume Group you will be presented
with the following menu:
[Entry Fields]
*Physical Volume name +
Logical Volume Name [] +
Logical Volume Label [] +
Volume Group name mndhb_vg_01
Resource Group Name rg_diskhbmulti_02 +
Network Name [net_diskhbmulti_02] +
+--------------------------------------------------------------------------
+
¦ Physical Volume name ¦
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-27
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
¦ ¦
¦ Move cursor to desired item and press Enter. Use arrow keys to scroll.
¦
¦ ¦
¦ 00c1890c57284929 ( hdisk18 on all cluster nodes ) ¦
¦ 00c1890c57285a08 ( hdisk19 on all cluster nodes ) ¦
¦ 00c1890c5728836e ( hdisk20 on all cluster nodes ) ¦
¦ ¦
+--------------------------------------------------------------------------
+
Select the physical volume you wish to use, then choose a name for the new
Logical Volume. In this example assume we selected hdisk18, and the new
logical volume name is mndhb_lv_05:
[Entry Fields]
*Physical Volume name 00c1890c57284929 +
Logical Volume Name [mndhb_lv_05] +
Logical Volume Label [] +
Volume Group name mndhb_vg_01
Resource Group Name rg_diskhbmulti_02 +
Network Name [net_diskhbmulti_02] +
Note: After adding the new Logical Volume for MNDHB you must perform
a verification and synchronization operation to activate the new MNDHB.
This menu option is used to stop using a Logical Volume for MNDHB.
+--------------------------------------------------------------------------
+
¦ Select Multi-Node Disk Heartbeat Network to Remove ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ net_diskhbmulti_01 (using mndhb_vg_01 in group rg_diskhbmulti_02)
¦
¦ net_diskhbmulti_03 (using mndhb_vg_01 in group rg_diskhbmulti_02)
¦
¦ net_diskhbmulti_05 (using mndhb_vg_04 in group rg_diskhbmulti_03)
¦
¦ net_diskhbmulti_06 (using mndhb_vg_04 in group rg_diskhbmulti_03)
¦
¦ net_diskhbmulti_07 (using mndhb_vg_02 in group rg_diskhbmulti_02)
¦
¦ net_diskhbmulti_08 (using mndhb_vg_03 in group rg_diskhbmulti_02)
¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ Esc+8=Image Esc+0=Exit Enter=Do ¦
¦ /=Find n=Find Next ¦
+--------------------------------------------------------------------------
+
Select the MNDHB network you wish deactivate and you will be presented
with a confirmation screen:
[Entry Fields]
Node Names racc1 racc2 racc3>
VOLUME GROUP name mndhb_vg_03
Resource Group Name rg_diskhbmulti_02
Network Name net_diskhbmulti_08
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-29
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
------------------------------------------------------------------------
** IMPORTANT **
AP ------------------------------------------------------------------------
The recommended failure action in an Oracle RAC environment is
“Halt the node”. This is not the default action, and must be
explicitly set. See the section on monitoring and displaying
the MNDHB configuration to view the current failure action setting
for each Volume Group.
------------------------------------------------------------------------
When changing the failure action, a picklist of all Volume Groups that
contain Logical Volumes used for MNDHB will be presented:
+--------------------------------------------------------------------------
+
¦ Select Multi-Node Disk Heartbeat Volume Group ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ mndhb_vg_01 ¦
¦ # (Used in network(s) net_diskhbmulti_01 net_diskhbmulti_03 )
¦
¦ mndhb_vg_02 ¦
¦ # (Used in network(s) net_diskhbmulti_07 ) ¦
¦ mndhb_vg_03 ¦
¦ # (Used in network(s) net_diskhbmulti_08 ) ¦
¦ mndhb_vg_04 ¦
¦ # (Used in network(s) net_diskhbmulti_05 net_diskhbmulti_06 )
¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ Esc+8=Image Esc+0=Exit Enter=Do ¦
¦ /=Find n=Find Next ¦
+--------------------------------------------------------------------------
+
[Entry Fields]
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-31
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Note: After changing the failure action for a Volume Group you must
perform a verification and synchronization operation to activate the
changes.
If cluster services are down the MNDHB will be activated the next
time the cluster is started.
SMIT
====
The primary interface for listing the MNDHB configuration is through the
SMIT path: cl_show_mndhb_dialog
AP Example output:
Network net_dhbmulti_05
Network net_dhbmulti_05 uses Logical Volume [/dev/mndhb_lv_06] in
Volume Group [mndhb_vg_04] for heartbeat.
Volume Group [mndhb_vg_04] is managed by Resource Group [rg_dhb_03]
The following nodes participate in this network:
racc1
racc2
racc3
On loss of access, HACMP will:
Halt the node
- clstat
- cldisp
- cltopinfo
========================
MNDHB Limitations
========================
o Number of networks
HACMP and RSCT have limits on the number of networks that can be defined.
The number of “networks” is the number of IP (e.g. ethernet) and non-IP
(e.g. rs232) networks as well as individual network “offsets” as calculated
by RSCT. This number is calculated and checked during the Verification
process.
Adding MNDHB networks will increase the total number of networks in use.
In general, the following table can be used to determine how many MNDHB
networks a given cluster configuration will support. This restriction
will be addressed in a future release of HACMP
====================================================================
Nodes *Existing Non-MNDHB interfaceMNDHB networks
====================================================================
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-33
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
* Includes boot addresses and any non-IP adapters per node. Does
not include service or persistent addresses.
==============================================
o APAR IZ49879: Changes in NFS mount default options
==============================================
Default options for NFS cross mounts are changed in HACMP 5.4.1 and later
releases.
Please refer to mount man page for a distinction between the two.
====================================================================
o Dealing with an Aynchronous I/O cache logical volume failure
====================================================================
An aio cache failure can be identified by the following error log entry
Additionally, the lsmp command will show “ASYNC CACHE VALID: no”
AP
To minimize the likelihood of an aio cache failure, IBM recommends using
either LVM mirroring of the aio cache logical volume, or placing that
logical volume on a disk subsystem on a device with RAID capabilities.
IBM further recommends that the aio cache logical volume not share the
same disk as data logical volumes.
In the event of a cache device failure, the recovery steps are as follows.
These must be completed prior to any other operations on the volume group.
- Correct the problem that caused the failure. This may involve replacing
a failed disk, relocating volume group contents to a different disk,
moving a disk to active state, or other operations, depending on the
precise failure.
- For each remote disk, use the chdev command to tell the RPV client to resume
communication with its RPV server. If ‘hdisk8’ is one of the local disks
in the asychronously mirrored volume group,
# chdev -l hdisk8 -a resume=yes
hdisk8 changed
- Use the varyonvg command to tell LVM to access the disks that it had
previously marked as missing. (Note that you run the varyonvg command
while the volume group is already varied online.)
# varyonvg datavg
This step runs the syncvg command in the background to synchronize the
logical partition copies on the remote disks.
- The lsmp command should show that the cache device is now valid. Look for
“ASYNC CACHE VALID: yes” in the output for the mirror pool.
© Copyright IBM Corp. 2007, 2009 Appendix B. HACMP 5.5 release_notes and README B-35
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
========================
Accessing IBM on the Web
========================
http://www.ibm.com
http://www.ibm.com/systems/p/ha/
========
Feedback
========
IBM welcomes your comments. You can send any comments via e-mail to:
hafeedbk@us.ibm.com
References
SC23-4867-05 HACMP for AIX: HACMP Master Glossary
SC23-4864-06 HACMP for AIX: Concepts and Facilities Guide
SC23-4861-06 HACMP for AIX: Planning and Installation Guide
SC23-4862-06 HACMP for AIX: Administration Guide
SC23-5177-00 HACMP for AIX: Troubleshooting Guide
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-1
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit Objectives
After completing this unit, you should be able to:
•Configure IP Address Takeover (IPAT) via IP replacement
Notes
Uempty
IPAT via IP Replacement Configuration
•Define each network’s boot IP addresses in the AIX ODM
– Each interface IP address on a given node must be in a different logical IP subnet* and there must
be a common subnet among the nodes
– Define these address in the /etc/hosts file and configure them in HACMP topology
•Define service IP addresses in /etc/hosts and HACMP resources
– The address must be in the SAME subnet as a common interface subnet
– HACMP configures them to AIX as required
* See earlier discussion of heartbeating and failure diagnosis for explanation of why
UNIX Software Service Enablement
Notes
Requirements
Keep the following items in mind when you configure a network for IPAT via IP
replacement:
- There must be at least one logical IP subnet which has a communication interface
(NIC) on each node. (In HACMP 4.5 terminology, these were called boot adapters.)
- Each service IP address must be in the same logical IP subnet as one of the
non-service addresses. Contrast with IPAT via IP aliasing, where service addresses
are required to NOT be in a boot subnet.
- If you have more than one service IP address, they must all be in the same subnet.
The reason for this will become clear when we discuss what happens during a
takeover, see “IPAT via IP Replacement After a Node Fails” on page C-8.
- None of the other non-service addresses may be in the same subnet as the service
IP address (this is true regardless of whether IPAT via IP replacement is being used
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-3
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
subnet IP labels
n1-if1, n2-if1, appA-svc,
192.168.10/24
appB-svc
192.168.11/24 n1-if2, n2-if2
Uempty
IPAT via IP Replacement in Operation
•When the resource group comes up on a node, HACMP
replaces a boot (ODM) IP label with the service IP label
–It replaces the boot IP label on the same subnet if the resource group is
on its startup node or if the distribution startup policy is used
–It replaces a boot IP label on a different subnet otherwise
Notes
Operation
When the resource group comes up on its home node, the resource group’s service IP
address replaces the interface IP address of the NIC (configured in the AIX ODM) which
is in the same subnet as the service IP label (that is, the boot adapter in HACMP 4.x
terminology). Note that this approach implies that there cannot be two resource groups
in the cluster which both use IPAT via IP replacement and use the same node as their
home node unless their respective service IP addresses are in different subnets (in
other words, associated with different physical networks).
Also, since the service IP address replaces the existing IP address on the NIC, it is not
possible to have two or more service IP addresses in the same resource group which
are in the same IP subnet (as there will not be an adapter to assign the second service
IP address to).
When the resource group comes up on any node other than its home node, the
resource group’s service IP address replaces the interface IP address of one of the
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-5
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
NICs which is not in the same subnet as the service IP address (this is primarily to allow
some other resource group to use the node as its home node).
Uempty
IPAT via IP Replacement After an I/F Fails
•If the communication interface being used for the service IP label
fails, HACMP swaps the service IP label with a boot (ODM) IP label
on one of the node's remaining available (that is, currently
functional) communication interfaces
•The IP labels remain swapped when the failed interface recovers
NIC A NIC B
9.47.11.1 (ODM) 9.47.10.22 (service) 9.47.10.2 (ODM) 9.47.11.2 (ODM)
Notes
Interface failure
If a communications interface (NIC A) which is currently assigned an IPAT via IP
replacement service IP address fails, then HACMP moves the service IP address to
one of the other communication interfaces (NIC B) on the same node (to one of the
standby adapters using HACMP 4.x terminology).
If there are no available (that is, functional) NICs left the relevant network then HACMP
initiates a fallover.
Interface swap
The failed communications interface (NIC A) is then reconfigured with the address of
the communication interface (NIC B) as this allows the heartbeat mechanism to watch
for when the failed communication interface (NIC A) recovers.
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-7
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Notes
Node failure
If the node currently responsible for an IPAT via IP replacement using resource group
fails then HACMP initiates a fallover. When the resource group comes up on the
takeover node, the service IP addresses are assigned to NICs on the fallover node:
- Home node or Startup policy of Online Using Distribution Policy (rotate in
HACMP 4.x terminology)
If the takeover node is the home node for the resource group or the resource group
has a Startup policy of Online Using Distribution Policy (rotate in HACMP 4.x
terminology), the Service IP addresses replace the IP addresses of a
communications interface (NIC) with an IP address in the same subnet as the
service IP address.
- Not the home node and not Online Using Distribution Policy
If the takeover node is not the home node for the resource group and the resource
group does not have a Startup policy of Online Using Distribution Policy, the
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-9
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Notes
Advantages
Probably the most significant advantage of IPAT via IP replacement is that it supports
hardware address takeover (HWAT), which will be discussed in a few pages.
Another advantage is that it requires fewer subnets. If you are limited in the number of
subnets available for your cluster, this may be important.
Note: Another alternative, if you are limited on the number of subnets you have
available, is to use heartbeating via IP aliases. See Heartbeating Over IP Aliases in
Chapter 3 of the HACMP for AIX 5L Planning Guide, Version 5.4.
Disadvantages
Probably the most significant disadvantages are that IPAT via IP replacement dictates
that there is only one service IP label per subnet per resource group on one
communications interface. IPAT via IP replacement also makes it rather expensive (and
Uempty complex) to support lots of resource groups in a small cluster. In other words, you need
more network adapters to support more applications.
In addition, IPAT via replacement usually takes more time than IPAT via aliasing.
Note that HACMP tries to keep the service IP Labels available by swapping IP
addresses with other communication interfaces (standby adapters in HACMP 4.x
terminology) even if there are no resource groups currently on the node which uses
IPAT via IP replacement.
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-11
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Notes
Review
When using IPAT via aliasing, you can use AIX’s gratuitous ARP features to update
client and router ARP caches after a takeover. However, there may be issues.
Uempty to be causing the cluster and the cluster administrator far more serious problems than
the ARP cache issue involves).
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-13
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
•Suggestion:
Do not get involved with using either clinfo or HWAT to deal
with ARP cache issues until you've verified that there actually
are ARP issues.
Notes
Uempty
Hardware Address Takeover
•HACMP can be configured to swap a service IP label's
hardware address between network adapters
•HWAT is incompatible with IPAT via IP aliasing because each
service IP address must have its own hardware address and a
NIC can support only one hardware address at any given time
•Cluster implementer designates a Locally Administered
Address (LAA) which HACMP assigns to the NIC which has
the service IP label
Notes
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-15
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
HWAT considerations
There are a few points which must be kept in mind when contemplating HWAT:
- The hardware address which is associated with the service IP address must be
unique within the physical network that the service IP address is configured for.
- HWAT is not supported by IPAT via IP aliasing because each NIC can have more
than one IP address but each NIC can only have one hardware address.
- HWAT is only supported for Ethernet, token ring and FDDI networks (MCA FDDI
network cards do not support HWAT). ATM networks do not support HWAT.
- HWAT increases the takeover time (usually by just a few seconds).
- HWAT is an optional capability which must be configured into the HACMP cluster
(we will see how to do that in a few minutes).
- Cluster nodes using HWAT on token ring networks must be configured to reboot
after a system crash as the token ring card will continue to intercept packets for its
hardware address until the node starts to reboot.
Uempty
Hardware Address Takeover (1 of 3)
node1-if1 node2-if1
9.47.9.1 9.47.9.2
tr1 255.255.255.0
00:04:ac:48:22:f4
tr1
255.255.255.0
00:04:ac:62:72:61
tr1
Before node2-if2
node1-if2
tr0 9.47.5.3
255.255.255.0
resource 9.47.5.2
255.255.255.0 tr0
00:04:ac:48:22:f6
00:04:ac:62:72:49
group is
started
node1 node2
node2-if1
node1-if1 9.47.9.2
tr1 9.47.9.1 255.255.255.0
255.255.255.0 00:04:ac:62:72:61 tr1
00:04:ac:48:22:f4
After node2-if2
xweb
tr0
9.47.5.1
255.255.255.0
resource 9.47.5.2
255.255.255.0 tr0
00:04:ac:48:22:f6
40:04:ac:62:72:49
group is
started
node1 node2
UNIX Software Service Enablement
Notes
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-17
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
xweb node2-if1
9.47.5.1 9.47.9.2
tr1 255.255.255.0
40:04:ac:62:72:49
tr1
255.255.255.0
00:04:ac:62:72:61
tr1
Interface node2-if2
xweb
tr0 9.47.5.1
255.255.255.0
failure 9.47.5.2
255.255.255.0 tr0
40:04:ac:62:72:49 00:04:ac:48:22:f6
node1 node2
xweb
node1-if1 9.47.5.1
tr1 9.47.9.1 255.255.255.0
255.255.255.0 40:04:ac:62:72:49 tr1
00:04:ac:48:22:f4
Node failure node2-if2
xweb 9.47.5.2
9.47.5.1 255.255.255.0
tr0 255.255.255.0 00:04:ac:48:22:f6
tr0
40:04:ac:62:72:49
node1 node2
UNIX Software Service Enablement
Notes
Uempty
Hardware Address Takeover (3 of 3)
node1-if1 xweb
9.47.9.1 9.47.5.1
tr1 255.255.255.0 255.255.255.0 tr1
40:04:ac:62:72:49
00:04:ac:48:22:f4 When a failed node
comes back to life, node2-if2
node1-if2 the burned in ROM 9.47.5.2
tr0 9.47.5.3 255.255.255.0 tr0
255.255.255.0 Address is used on 00:04:ac:48:22:f6
00:04:ac:62:72:49
the service network
adapter.
node1 node2
node2-if1
node1-if1 9.47.9.2
tr1 9.47.9.1 255.255.255.0 tr1
255.255.255.0 00:04:ac:62:72:61
00:04:ac:48:22:f4
After HACMP is
xweb
started the node node2-if2
reintegrates 9.47.5.2
tr0 9.47.5.1 255.255.255.0 tr0
255.255.255.0 according to its 00:04:ac:48:22:f6
40:04:ac:62:72:49
resource group
parameters
node1 node2
UNIX Software Service Enablement
Notes
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-19
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
node1 node2
D D
A A
Notes
Reality check
A side note is probably in order: although most TCP/IP-capable systems respect
gratuitous ARP, there are strange devices out there that do not. This scenario is phoney
but it presents a real if rather unlikely problem. For example, the ATM network does not
support gratuitous ARP and so could be a candidate for the use of HWAT.
Uempty
Our Plan for Implementing HWAT
1.Stop cluster services on both cluster nodes
íUse the graceful shutdown option to bring down the resource groups
and their applications
2.Remove the alias service labels from the Resources
–They are in the wrong subnet for replacement
–They are automatically removed from the RG
3.Convert the net_ether_01 Ethernet network to use IPAT via IP
replacement:
–Disable IPAT via IP aliasing on the Ethernet network.
–Update /etc/hosts on both cluster nodes to describe service IP labels
and addresses on the 192.168.15.0 subnet
–Use the procedure described in the networking to select the (Locally
Administered Addresses (LAA) addresses
–Configure new service IP labels with these LAA addresses in the
HACMP SMIT screens
4.Define resource groups to use the new service IP labels
5.Synchronize the changes
6.Restart HACMP on the two nodes
Notes
Implementing HWAT
In order to use HWAT, we must use IPAT via replacement.
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-21
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Uempty
Stopping HACMP
# smit clstop
Stop Cluster Services
Type or select values in entry fields.
Press Enter AFTER making all desired changes.
[Entry Fields]
* Stop now, on system restart or both now
+
Stop Cluster Services on these nodes [node1,node2]
+
BROADCAST cluster shutdown? true
+
* Shutdown mode graceful
+
Notes
Stop HACMP
Make sure that HACMP is shut down gracefully, as we can’t have the application
running while we are changing service IP addresses.
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-23
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
+--------------------------------------------------------------------------+
¦ Select Service IP Label(s)/Address(es) to Remove ¦
¦ ¦
¦ Move cursor to desired item and press F7. ¦
¦ ONE OR MORE items can be selected. ¦
¦ Press Enter AFTER making all selections. ¦
¦ ¦
¦ xweb ¦
¦ yweb ¦
¦ zweb ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F7=Select F8=Image F10=Exit ¦
F1¦ Enter=Do /=Find n=Find Next ¦
F9+--------------------------------------------------------------------------+
Notes
Uempty
Disable IPAT via Aliases
Set the "Enable IP Address Takeover via IP Aliases" setting to "No"
and press Enter.
Notes
Introduction
Here we change the net_ether_01 network to disable IPAT via aliasing.
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-25
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Notes
Uempty
Creating a
Locally Administered Address (LAA)
•Each service IP label using HWAT will need an LAA
•The LAA must be unique on the cluster's physical network
•The MAC address based technologies (Ethernet, Token ring
and FDDI) use six byte hardware addresses of the form:
xx.xx.xx.xx.xx.xx
•The factory-set MAC address of the NIC will start with 0, 1, 2
or 3
–A MAC address that starts with 0, 1, 2 or 3 is called a Globally
Administered Address (GAA) because it is assigned to the NIC's
vendor by a central authority
•Incrementing this first digit by 4 transforms the GAA into a
Locally Administered Address (LAA) which will be unique
worldwide (unless someone has already used the same GAA
to create an LAA which isn't likely since GAAs are unique
worldwide)
UNIX Software Service Enablement
Notes
Hardware addresses
Hardware addresses must be unique, at a minimum, on the local network to which they
are connected. The factory set hardware address for each network interface card (NIC)
is administered by a central authority and should be unique in the world. These
addresses are called Globally Administered Addresses (GAAs).
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-27
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Notes
Uempty
Hardware Address Takeover Issues
•Do not enable the ALTERNATE hardware address field
in the SMIT devices menu
–Causes the adapter to boot on your chosen LAA rather than the
burned in ROM address
–Causes serious communications problems and puts the cluster in
to an unstable state
–Correct method is to enter your chosen LAA in to the SMIT
HACMP menus (remove the periods or colons before entering it
into the field)
•The Token-Ring documentation states that the LAA must
start with 42
•The FDDI documentation states that the first nibble (digit) of
the first byte of the LAA must be 4, 5, 6 or 7 (which is
compatible with the method for creating LAAs described
earlier)
•Token-Ring adapters do not release the LAA if AIX crashes.
–AIX must be set to reboot automatically after a system crash
(see smitty chgsys)
UNIX Software Service Enablement
Notes
Issues
The main thing to remember is that you do NOT configure the ALTERNATE hardware
address field in the SMIT devices panel.
You must leave that blank and configure this using the SMIT HACMP menus.
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-29
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Don't forget to specify the second LAA for the second service IP label.
UNIX Software Service Enablement
Notes
Uempty
Synchronize Your Changes
Synchronize the changes and run through the test plan.
Notes
Synchronize
Don’t forget to synchronize.
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-31
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Review
1. For IPAT via replacement (select all that apply)
a. Each service IP address must be in the same subnet as one of the non-service
addresses
b. Each service IP address must be in the same subnet
c. Each service IP address cannot be in any non-service address subnet
2. True or False?
If the takeover node is not the home node for the resource group and the
resource group does not have a Startup policy of Online Using Distribution
Policy, the service IP address replaces the IP address of a NIC with an IP
address in the same subnet as the subnet of the service IP address.
3. True or False?
In order to use HWAT, you must enable and complete the ALTERNATE
ETHERNET address field in the SMIT devices menu.
4. True or False?
You must stop the cluster in order to change from IPAT via aliasing to IPAT via
replacement.
Notes
Uempty
Unit Summary
•IPAT via IP replacement:
–May require fewer subnets than IPAT via aliasing
–May require more NICs than IPAT via aliasing
–Supports hardware address takeover
•HACMP replaces non-service IP labels with service IP labels on the
same subnet as the service IP label when the resource group is started
on its home node or if the Startup Policy is distributed
•HACMP replaces non-service IP labels with service IP labels on a
different subnet from the service IP label when the resource group is
moved to any other node
•IPAT via IP replacement configuration issues
–Service IP address must be the same subnet as one of the non-service
subnets
–All service IP addresses must be in the same subnet
–You must have at least as many NICs on each node as service IP
addresses
•Hardware address takeover (HWAT) issues
–Alternate hardware address (Locally Administered Address or LAA)
must be configured in HACMP. Do NOT use standard SMIT field.
–Alternate hardware address must be unique.
Notes
© Copyright IBM Corp. 2007, 2009 Appendix C. IPAT via IP Replacement C-33
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
References
SC23-5209-00 HACMP for AIX, Version 5.4 Installation Guide
SC23-4864-09 HACMP for AIX, Version 5.4:
Concepts and Facilities Guide
SC23-4861-09 HACMP for AIX, Version 5.4 Planning Guide
SC23-4862-09 HACMP for AIX, Version 5.4 Administration Guide
SC23-5177-03 HACMP for AIX, Version 5.4 Troubleshooting Guide
SC23-4867-08 HACMP for AIX, Version 5.4 Master Glossary
http://www-03.ibm.com/systems/p/library/hacmp_docs.html
HACMP manuals
© Copyright IBM Corp. 2007, 2009 Appendix D. Configuring target mode SSA D-1
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Unit Objectives
After completing this unit, you should be able to:
Notes
Uempty
Implementing Target Mode SSA
•The serial cable being used to implement the RS232 non-IP
network has been borrowed by someone and nobody
noticed
•A decision has been made to implement a target mode SSA
(tmssa) non-IP network as it won't fail unless complete
access to the shared SSA disks is lost by one of the nodes
(and someone is likely to notice that)
node1 node2
D D
A A
Notes
© Copyright IBM Corp. 2007, 2009 Appendix D. Configuring target mode SSA D-3
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Use the smitty ssaa fastpath to get to AIX's SSA Adapters menu.
UNIX Software Service Enablement
Notes
Required software
Target mode SSA support requires that the devices.ssa.tm.rte file set be installed on
all cluster nodes.
© Copyright IBM Corp. 2007, 2009 Appendix D. Configuring target mode SSA D-5
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Notes
Introduction
Once each node has a unique SSA node number, the AIX configuration manager needs
to be used to define the tmssa devices. Each node must have tmssa devices which
refer to each of the other nodes that they can see via the SSA loops. When cfgmgr is
run on a node, it sets up the node to accept tmssa packets, and it then defines tmssa
devices referring to any other nodes which respond to tmssa packets. In order for this to
all work, the other nodes must all be set up to accept and respond to tmssa packets.
Procedure
The end result is that the following procedure gets all the required tmssa devices
defined:
Uempty 1. Run cfgmgr on each cluster node in turn. This sets up each node to handle tmssa
packets, and defines the tmssa devices on each node to refer to nodes which have
already been setup for tmssa.
2. Run cfgmgr on each node in turn again (depending upon exactly what order you do this
in, it is actually possible to skip running cfgmgr on one of the nodes, but it is probably
not worth the trouble of being sure that the last cfgmgr run wasn’t required).
3. Verify the tmssa devices exist:
Run
# lsdev -C | grep tmssa
on each node. There should be a tmssa device (which is actually a target mode SSA
router acting as a pseudo device) configured on each node.
4. Verify the tmssa devices exist:
Run
# ls /dev/tmssa*
on each node. Note that each node has target mode SSA devices called
/dev/tmssa#.im and /dev/tmssa#.tm where # refers to the other node’s node number.
5. Test the target mode connection:
Enter the following command on the node with id 1 (make sure you specify the tm suffix
and not the im suffix):
# cat < /dev/tmssa2.tm
(This command should hang)
On the node with ID 2, enter the following command (make sure that you specify the im
suffix and not the tm suffix):
# cat /etc/hosts > /dev/tmssa1.im
(The /etc/hosts file should be displayed on the first node)
This validates that the target mode serial network in functional. Please note that any
text file may be substituted for /etc/hosts and you have to specify different tmssa
device names if you configured different SSA node numbers for each node. This is
simply an example.
© Copyright IBM Corp. 2007, 2009 Appendix D. Configuring target mode SSA D-7
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Extended Configuration
Move cursor to desired item and press Enter.
Discover HACMP-related Information from Configured Nodes
Extended Topology Configuration
Extended Resource Configuration
Extended Event Configuration
Extended Performance Tuning Parameters Configuration
Security and Users Configuration
Snapshot Configuration
Extended Verification and Synchronization
Notes
HACMP discover
By discovering the new devices, they will appear in SMIT pick lists when we configure
the tmssa non-IP network. Strictly speaking, it is not necessary to rerun the HACMP
discovery as it is possible to configure tmssa networks by entering in the tmssa device
names explicitly. As this is a rather error-prone process, it is probably best to use the
HACMP discovery mechanism to discover the devices for us.
Uempty
Defining a Non-IP tmssa Network (1 of 3)
This should look very familiar as it is the same procedure that was
used to define the non-IP RS232 network earlier.
Configure HACMP Communication Interfaces/Devices
Move cursor to desired item and press Enter.
+--------------------------------------------------------------------------+
¦ Select a category ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ Add Discovered Communication Interface and Devices ¦
¦ Add Predefined Communication Interfaces and Devices ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F8=Image F10=Exit Enter=Do ¦
F1¦ /=Find n=Find Next ¦
F9+--------------------------------------------------------------------------+
Notes
© Copyright IBM Corp. 2007, 2009 Appendix D. Configuring target mode SSA D-9
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
+--------------------------------------------------------------------------+
¦ Select a category ¦
¦ ¦
¦ Move cursor to desired item and press Enter. ¦
¦ ¦
¦ # Discovery last performed: (Feb 12 18:20) ¦
¦ Communication Interfaces ¦
¦ Communication Devices ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F8=Image F10=Exit Enter=Do ¦
F1¦ /=Find n=Find Next ¦
F9+--------------------------------------------------------------------------+
Notes
Communication device
The tmssa devices are communication devices used for non-IP communication, so
select Communication Devices here.
Uempty
Defining a Non-IP tmssa Network (3 of 3)
Now, we need to define the tmssa network using a process
Configure HACMP Communication Interfaces/Devices
Move cursor to desired item and press Enter.
Add Communication Interfaces/Devices
+--------------------------------------------------------------------------+
¦ Select Point-to-Point Pair of Discovered Communication Devices to Add ¦
¦ ¦
¦ Move cursor to desired item and press F7. Use arrow keys to scroll. ¦
¦ ONE OR MORE items can be selected. ¦
¦ Press Enter AFTER making all selections. ¦
¦ ¦
¦ # Node Device Device Path Pvid ¦
¦ > node2 tmssa1 /dev/tmssa1 ¦
¦ > node1 tmssa2 /dev/tmssa2 ¦
¦ node1 tty0 /dev/tty0 ¦
¦ node2 tty0 /dev/tty0 ¦
¦ node1 tty1 /dev/tty1 ¦
¦ node2 tty1 /dev/tty1 ¦
¦ ¦
¦ F1=Help F2=Refresh F3=Cancel ¦
¦ F7=Select F8=Image F10=Exit ¦
F1¦ Enter=Do /=Find n=Find Next ¦
F9+--------------------------------------------------------------------------+
Notes
Final step
Select the tmssa devices on each node and press Enter to define the network.
Refer to Chapter 6 of the HACMP v5.4 Installation Guide for information on configuring
all supported types of non-IP networks.
© Copyright IBM Corp. 2007, 2009 Appendix D. Configuring target mode SSA D-11
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
Notes
Synchronize
As when you make any change to the cluster, verify and synchronize.
Uempty
Unit Summary
Notes
© Copyright IBM Corp. 2007, 2009 Appendix D. Configuring target mode SSA D-13
Course materials may not be reproduced in whole or in part
without the prior written permission of IBM.
Student Notebook
backpg
Back page
®