Sponsored Content
Operating Systems AIX Upgrading PowerHA from 7.1.0 to 7.1.3 Post 302961185 by linux.amrit on Wednesday 25th of November 2015 01:42:18 AM
Old 11-25-2015
Upgrading PowerHA from 7.1.0 to 7.1.3

Hi All,

As per the IBM upgrade/support matrix

IBM Techdocs Technote: PowerHA for AIX Version Compatibility Matrix

we can't do online upgrade or rolling migration from powerha v7.1.0 to v7.1.3. on AIX61_TL9_SP4, So we are following the steps as below ...

1 ) Bring down the cluster
2 ) CAA_FORCE_ENABLED=1
3 ) /usr/sbin/rmcluster -n cluster_name
4 ) upgrade PowerHA using smitty update_all ( updating PowerHA software from 7.1.0 to 7.1.3 ) to all the nodes.
5 ) after upgrade restart all the nodes .. shutdown -Fr
6 ) Create the CAA cluster by running the /usr/es/sbin/cluster/utilities/clmkcaa command
7 ) Verify that the CAA services are active by running the /usr/sbin/lscluster -m
8 ) If step 6 failed then .. and cluster don't form properly then follow the below steps ..
/usr/es/sbin/cluster/conversion/clconvert_snapshot -v version -s snapshotname ( version is the previous version before upgrade can be taken using halevel command )
E.g. /usr/es/sbin/cluster/conversion/clconvert_snapshot -v 7.1.0 -s snapshotname
From the command line, enter smit hacmp.
In SMIT, select Cluster Nodes and Networks > Manage the cluster > Snapshot Configuration > Restore the Cluster Configuration from a Snapshot, and press Enter.
Select the snapshot that you want to restore, complete the required fields, and press Enter.
9 ) Run the /usr/sbin/lscluster -m command to verify that the Cluster Aware AIX (CAA) services are active on each node in the cluster.

So far this worked well, but few servers we aren't getting downtime for powerha upgrade, is there any one done online or rolling migration ? Any one found any clever way to over this downtime difficulty ?

Thanks In Advance,
Amrit
 

10 More Discussions You Might Find Interesting

1. AIX

strange route issue in powerha 5.4?

hi all we upgraded hacmp(powerha) from 5.2 to 5.4 recently . during the failover testing, we found a strange network issue. after standby node took service ip address (172.15.100.8) online at standby NIC, we were able to log in the standby node by telnet 172.15.100.8 which stays at... (1 Reply)
Discussion started by: rs6000er
1 Replies

2. AIX

PowerHA - assistant can't find network

I've configured multiple clusters using PowerHA 5.5, but never using the 'Two-node cluster configuration assistant'. I removed all the clusters, clean setup, and gave the assistant a try. I've filled in all the necessary options: Communication Path to Takeover Node Application Server ... (11 Replies)
Discussion started by: Koen87
11 Replies

3. AIX

PowerHA, same subnet for persistent and service ip

I am new in AIX and please forgive my poor english. I know that AIX allow same subnet IPs for different interfaces, which result in multipath routing / route striping. My question is, is there any best practice for the persistent and service IP with same subnet to stay on same interface, or... (5 Replies)
Discussion started by: skeyeung
5 Replies

4. AIX

PowerHA Disk on VIO Server

Hello AIX GURU's Can anybody tell me the steps to crate shared VG (enhanced concurent) for my Lpars from VIO server? my questions are: 1. Should I crate Enhanced Concurent VG in VIO and map it using virtual Scsi to Lpar? or 2. Can I just create virtual SCSI in VIO and map to Lpar and... (1 Reply)
Discussion started by: Vit0_Corleone
1 Replies

5. AIX

PowerHA with 3rd node for communication

Looking to find PowerHA with 3rd node used for communication (voting node) only. Does anyone use such configuration with IBM products? (2 Replies)
Discussion started by: gogogo
2 Replies

6. AIX

PowerHA HACMP on VIOS servers

Few questions regarding Power HA ( previously known as HACMP) and VIOS POWERVM IVM ( IBM Virtualization I/O Server ) Is it possible to create HACMP cluster between two VIOS servers Physical Machine_1 VIOS_SERVER_1 LPAR_1 SHARED_DISK_XX VIOS_SERVER_2 Physical Machine_2 LPAR_2... (6 Replies)
Discussion started by: filosophizer
6 Replies

7. AIX

Powerha on AIX 6.1.9

hello ive installed powerha 7.1.3 on two servers aix 6.1.9 6100-09-03-1415 work with dmx4 EMC storgae. after sync cluster ( terminate with OK ) ive see that the repository disk upper only in one machine : hdiskpower60 00c7f6b59fc60d9d caavg_private active... (1 Reply)
Discussion started by: ariec
1 Replies

8. AIX

VXVM in POWERHA 6.1

I have created a VxVM disk group in AIX7.1. I have tried to added this VxVM disk group in powerHA 6.1. But in cluster VxVM DGs are not listing. Is there any other procedure to add vxvm diskgroup to hacmp. Please share me steps for adding vxvm diskgroup to hacmp. (6 Replies)
Discussion started by: sunnybee
6 Replies

9. AIX

PowerHA; Import flash copy of VG?

Hey guys, wondering if this is possible to accomplish. PowerHA SystemMirror7.1 active/passive cluster. Restored a clustered system in test and upgraded to the latest version of PowerHA. Things are working great. However in the interest of time during cut over to the new system, I was hoping to... (4 Replies)
Discussion started by: j_aix
4 Replies

10. AIX

PowerHa 7.1 Migration check failed

Hello Guys, I was trying to upgrade the PowerHA from 6.1.0.9 to 7.1.3 in prod cluster. Have snapshot migration and run climgcheck. 1. Run climgcheck on node1(kul1pqcmur01) and it success 2. Upgrade the cluster sw 3. Run climgcheck on node2(kul1pqcmur02) and it failed with below errors ... (6 Replies)
Discussion started by: novaothers
6 Replies
cmdeleteconf(1m)														  cmdeleteconf(1m)

NAME
cmdeleteconf - Delete either the cluster or the package configuration SYNOPSIS
cmdeleteconf [-f] [-v] [-c cluster_name] [[-p package_name]...] DESCRIPTION
cmdeleteconf deletes either the entire cluster configuration, including all its packages, or only the specified package configuration. If neither cluster_name nor package_name is specified, cmdeleteconf will delete the local cluster's configuration and all its packages. If the local node's cluster configuration is outdated, cmdeleteconf without any argument will only delete the local node's configuration. If only the package_name is specified, the configuration of package_name in the local cluster is deleted. If both cluster_name and pack- age_name are specified, the package must be configured in the cluster_name, and only the package package_name will be deleted. cmdelete- conf with only cluster_name specified will delete the entire cluster configuration on all the nodes in the cluster, regardless of the con- figuration version. The local cluster is the cluster that the node running the cmdeleteconf command belongs to. Only a superuser, whose effective user ID is zero (see id(1) and su(1)), can delete the configuration. To delete the cluster configuration, halt the cluster first. To delete a package configuration you must halt the package first, but you do not need to halt the cluster (it may remain up or be brought down). To delete the package VxVM-CVM-pkg (HP-UX only), you must first delete all packages with STORAGE_GROUP defined. While deleting the cluster, if any of the cluster nodes are powered down, the user can choose to continue deleting the configuration. In this case, the cluster configuration on the down node will remain in place and, therefore, be out of sync with the rest of the cluster. If the powered-down node ever comes up, the user should execute the cmdeleteconf command with no argument on that node to clean up the config- uration before doing any other Serviceguard command. Options cmdeleteconf supports the following options: -f Force the deletion of either the cluster configuration or the package configuration. -v Verbose output will be displayed. -c cluster_name Name of the cluster to delete. The cluster must be halted already, if intending to delete the cluster. -p package_name Name of an existing package to delete from the cluster. The package must be halted already. There should not be any packages in the cluster with STORAGE_GROUP defined before having a package_name of VxVM-CVM-pkg (HP-UX only). RETURN VALUE
Upon completion, cmdeleteconf returns one of the following values: 0 Successful completion. 1 Command failed. EXAMPLES
The high availability environment contains the cluster, clusterA , and a package, pkg1. To delete package pkg1 in clusterA, do the following: cmdeleteconf -f -c clusterA -p pkg1 To delete the cluster clusterA and all its packages, do the following: cmdeleteconf -f -c clusterA AUTHOR
cmdeleteconf was developed by HP. SEE ALSO
cmcheckconf(1m), cmapplyconf(1m), cmgetconf(1m), cmmakepkg(1m), cmquerycl(1m). Requires Optional Serviceguard Software cmdeleteconf(1m)
All times are GMT -4. The time now is 05:48 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy