Powerha Systemmirror For Aix V7.1 Two-Node Quick Configuration Guide
Powerha Systemmirror For Aix V7.1 Two-Node Quick Configuration Guide
Powerha Systemmirror For Aix V7.1 Two-Node Quick Configuration Guide
Shawn Bodily
Advanced Technical Skills
November 2012
A. Introduction .................................................................3
B. Configuring the Cluster ..............................................4
Configuring VLAN 3358 for SAN heartbeat.................8
C. Testing the Cluster.....................................................19
E. References ...................................................................22
Appendix A – Cluster Test Tool Log.............................23
• One p260 and one p460 Power Compute Nodes with 8GB Fibre Channel Mezzanine Adapters
• Flex System Enterprise Chassis with two 10GB Ethernet Switches and two 8GB QLogic Fibre
Channel Switch Modules
• V7000 Storage
• AIX 7.1, TL 1, SP 3
• PowerHA Version 7.1.1 SP3
• RSCT 3.1.2
If installing PowerHA v7.1.2 or v7.1.3 the following matrix can be referenced for base AIX level
requirements.
http://w3-03.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/TD101347
Though the following steps are the same for both 7.1.2 and 7.1.3, there is one menu difference in v7.1.3. It is
is the addition of unicast for heartbeating. An updated screen shot has been added to reflect it.
Additional details on installing and configuring PowerHA SystemMirror can be found in the Installing
PowerHA SystemMirror guide.
This cluster was configured utilizing standard SMIT sysmirror menus. There is also the option of utilizing the
PowerHA Systems Director plug-in to configure, monitor and manage the cluster. More information on the
Systems Director plug-in option can be found in the PowerHA v7.1 redbook.
To start creating the cluster, enter smitty sysmirror→ Cluster Nodes and Networks→ Initial Cluster Setup
(Typical) →Setup a Cluster, Nodes and Networks
The next step is to define a cluster repository disk and multicast address. We can use fastpaths in SMIT to
bypass additional menus. Execute smitty cm_setup_menu→ Define Repository Disk and Cluster IP Address:
In PowerHA v7.1.0-v7.1.2 the use of multicasting is required. The Cluster IP address is the multicast address.
Its not required to enter one as PowerHA will choose one for you. It usually creates one by taking the last 3
octets of the hostname IP address from the node in which the cluster is being created on and replacing the first
octet with 228. In our cluster, we created it on the p460 with hostname address of 172.23.17.35. The multicast
address was created automatically of 228.23.17.35.
However, in v7.1.3 unicast was re-introduced making multicast optional. The v7.1.3 menu is shown below:
After these two steps it is recommended to synchronize the cluster. (Execute smitty sysmirror→Cluster Nodes
and Networks→ Verify and Synchronize Cluster Configuration and press Enter twice). The main reason
being, the first time the cluster is synced the CAA cluster is created automatically. That way if a problem is
encountered, it can be addressed before adding all the additional cluster components. The following shows the
disk and CAA volume group information after the synchronization and CAA cluster was created successfully.
Other than the PowerHA 7.1.1 redbook and APAR that added this support, these steps have never been clearly
documented in an officially publication. However, during the collaboration on another whitepaper involving
PowerHA v7.1 and v7000 (http://tinyurl.com/c3vksk7), we took the opportunity to document it. The
following is a snippet from that whitepaper.
Login to each VIOS partition and change the fibre channel attributes using the chdev command. Depending
on how disks are configured in the system, you may have to use the rmdev command to put the device in the
defined state. After changing the attributes, use the cfgdev command to configure the device or reboot the
partition.
# > rmdev -dev fcs0 –ucfg -recursive
fcnet0 Defined
sfwcomm1 Defined
fscsi0 Defined
fcs0 Defined
# > chdev -dev fcs0 -perm -attr tme=yes
fcs0 changed
# > chdev -dev fscsi0 -perm -attr dyntrk=yes
fscsi0 changed
# > chdev -dev fscsi0 -perm -attr fc_err_recov=fast_fail
fscsi0 changed
# > lsdev -dev fcs0 -attr |grep tme
tme yes Target Mode Enabled True
# > lsdev -dev fscsi0 -attr |grep dyntrk
dyntrk yes Dynamic Tracking of FC Devices True
# > lsdev -dev fscsi0 -attr |grep fc_err_recov
Another valuable source of documentation for this topic can be found at:
http://www.ibm.com/developerworks/aix/library/au-aix-powerha-heartbeat/index.html
We now need to create our resources (application controller, service address, and shared volume group) and a
resource group to add them into.
In our scenario we have no real application to utilize. So we created a dummy application controller by simply
having it execute a banner command. We can add it by executing smitty sysmirror→ Cluster Applications and
Resources→Resources→ Configure User Applications (Scripts and Monitors) → Application Controller
Scripts
After adding the service IP we can see it has been added to the cluster topology as shown from the cllsif output
as follows:
To add the resources to the resource group, execute the same fastpath of smitty
cm_resource_groups→Change/Show Resources and Attributes for a Resource Group and choose the
previously created resource group. Then for the fields of Service IP Labels/Addresses and Application
Controllers, press F4 and a pop-up will appear with the ones previous created. Choose them, and press Enter.
The last thing to configure is the shared data volume group, logical volume(s) and filsystem(s). This can be
accomplished by using the Cluster Single Point of Control facility (C-SPOC). Enter smitty
cspoc→Storage→Volume Groups→Create a Volume Group (choose both nodes). Then choose the desired
disk as shown below:
In the final menu, fill out the fields as desired. Also note that you can choose to added the volume group into
the previously created resource group as shown.
Even when using fileystems, we always recommend creating the underlying logical volume and log logical
volume in order to control the naming convention. That way it ensures the names are unique among the cluster
nodes.
To create the logical volumes and filesystems we continue to utilize C-SPOC to create a new logical volume
and log logical volume. Execute smitty cspoc→Storage→Logical Volumes→Add a Logical Volume
Choose the desired disk to create the logical volume on as shown above. Then on the final menu, specify the
desired size, name and type. In our case the type is “jfs2”
Lastly we will now create a JFS2 filesystem on top of our original previously created logical volume. We
execute smitty cspoc→Storage→File Systems→Add a File System (choose previously created volume group)
Notice: In the volume group pop-up list the CAA specific volume group of caavg_private may appear.
NEVER chose that volume group. It should be considered a bug that needs to be fixed.
Then choose the specific filesystem, in our case it is “Enhanced Journaled File System” as shown below.
The execute the cluster test tool, enter smitty hacmp_testtool_menu, then choose “Execute Automated Test
Procedure” as shown below:
Once pressing enter, the final menu is displayed as shown below. The detailed results of each test are
displayed in the SMIT window during execution and are also saved in /var/hacmp/log/cl_testtool.log. Our
actual test results were added into the Appendix.
While this testing does cover the core basic functionality of the cluster, additional granular level testing via the
Custom Test Procedure is often desired to include such common events as:
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Initializing Variable Table
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: Using Process Environment for Variable Table
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Reading Static Configuration Data
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: Cluster Name: PHAFlexCluster
08/08/2012_14:13:11: Cluster Version: 13
08/08/2012_14:13:11: Local Node Name: p260_PowerHA
08/08/2012_14:13:11: Cluster Nodes: p260_PowerHA p460_PowerHA
08/08/2012_14:13:11: Found 1 Cluster Networks
08/08/2012_14:13:11: Found 3 Cluster Interfaces/Device/Labels
08/08/2012_14:13:11: Found 1 Cluster Resource Groups
08/08/2012_14:13:11: Found 10 Cluster Resources
08/08/2012_14:13:11: Event Timeout Value: 720
08/08/2012_14:13:11: Maximum Timeout Value: 2880
08/08/2012_14:13:11: Found 0 Cluster Sites
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Building Test Queue
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: Test Plan: /usr/es/sbin/cluster/cl_testtool/auto_topology
08/08/2012_14:13:11: Event 1: NODE_UP: NODE_UP,ALL,Start cluster services on all
available nodes
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Validate NODE_UP
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: Event node: ALL
08/08/2012_14:13:11: Configured nodes: p260_PowerHA p460_PowerHA
08/08/2012_14:13:11: Event 2: NODE_DOWN_GRACEFUL: NODE_DOWN_GRACEFUL,node1,Stop
cluster services gracefully on a node
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Validate NODE_DOWN_GRACEFUL
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: Event node: p260_PowerHA
08/08/2012_14:13:11: Configured nodes: p260_PowerHA p460_PowerHA
08/08/2012_14:13:11: Event 3: NODE_UP: NODE_UP,node1,Restart cluster services on the
node that was stopped
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Validate NODE_UP
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: Event node: p260_PowerHA
08/08/2012_14:13:11: Configured nodes: p260_PowerHA p460_PowerHA
08/08/2012_14:13:11: Event 4: NODE_DOWN_TAKEOVER: NODE_DOWN_TAKEOVER,node2,Stop
cluster services with takeover on a node
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Validate NODE_DOWN_TAKEOVER
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: Event node: p460_PowerHA
08/08/2012_14:13:11: Configured nodes: p260_PowerHA p460_PowerHA
08/08/2012_14:13:11: Event 5: NODE_UP: NODE_UP,node2,Restart cluster services on the
node that was stopped
08/08/2012_14:13:11: -------------------------------------------------------
08/08/2012_14:13:11: | Validate NODE_UP
08/08/2012_14:13:11: -------------------------------------------------------