Hey guys, wondering if this is possible to accomplish.
PowerHA SystemMirror7.1 active/passive cluster. Restored a clustered system in test and upgraded to the latest version of PowerHA. Things are working great. However in the interest of time during cut over to the new system, I was hoping to take a flash copy of the data lun (datavg) from the current production cluster and import that to the new system, however, I'm getting stuck removing hdisk1 on the restored system.
Steps to recreate on the new cluster:
Stop the cluster services on both nodes so datavg is varied off
on both nodes - success
on both nodes - success on node2(passive) fail on node1(active)
No matter what I try I can't remove the disk on node1, getting this error.
The PVID's are going to change with the flash copy, so my thought was to remove these from ODM on the new system and then drop that lun. Next, add the flashed lun and do:
Your ODM might need to be massaged a bit (on some rare occasions it gets into an inconsistent state). Try the "synclvodm" command to correct that.
You might find out why the disk is busy by using "fuser" on the disk device ("/dev/hdiskX"). Maybe some monitoring agent or other software is holding it open.
Thanks for the reply. I actually resolved this yesterday.
The only way to do this was to first remove the cluster definition, because when volumes are part of a cluster resource group, you need to use cluster commands to make changes so the cluster is aware.
After removing the cluster definition I varied off the vg and exported the vg on both nodes. I could see the physical volume was still busy when trying to remove it, so I located the parent via:
Turns out the fiber adapter was keeping the volume busy, this adapter is also used for the rootvg so disabling wasn't an option. A reboot was required.
Following the reboot I was able to remove the volume and add the newly flashed lun. Lastly I redefined the cluster, tested, success.
The only way to do this was to first remove the cluster definition, because when volumes are part of a cluster resource group, you need to use cluster commands to make changes so the cluster is aware.
Interesting. This should not be the case. I remove volumes and change volume groups actually all the time, in 6.x-clusters as well as in 7.x-clusters. To have one disk remain constant in a VG and do "learning import" (importvg -L) on the other node is usually all it takes.
I usually even replace disks completely by: adding the new disk, mirror, importvg on the passive node, then unmirror, reducevg, another importvg on the passive node. Not even a takeover or application touchdown is necessary that way.
Quote:
Originally Posted by j_aix
I could see the physical volume was still busy when trying to remove it, so I located the parent via:
Turns out the fiber adapter was keeping the volume busy, this adapter is also used for the rootvg so disabling wasn't an option. A reboot was required.
Following the reboot I was able to remove the volume and add the newly flashed lun. Lastly I redefined the cluster, tested, success.
My "gut feeling" rather says the FC adapter was the culprit, not the cluster software. Anyways, i am glad you resolved it.
Interesting. This should not be the case. I remove volumes and change volume groups actually all the time, in 6.x-clusters as well as in 7.x-clusters. To have one disk remain constant in a VG and do "learning import" (importvg -L) on the other node is usually all it takes.
I usually even replace disks completely by: adding the new disk, mirror, importvg on the passive node, then unmirror, reducevg, another importvg on the passive node. Not even a takeover or application touchdown is necessary that way.
You're probably right, I think removing the cluster definition wouldn't have been necessary had I known it was the Fc adapter holding onto the volume. Of course in the process of narrowing it down it just turned out to be the way I did it. Since it was the only physical volume in the VG (and concurrent) I felt it was the safest way to go about it.
For my scenario I will need to do this (redefine the cluster) anyway due to a complete IP schema change. I may even try out the cluster snapshot feature going forward. Thanks again for the follow-up!
Hi All,
As per the IBM upgrade/support matrix
IBM Techdocs Technote: PowerHA for AIX Version Compatibility Matrix
we can't do online upgrade or rolling migration from powerha v7.1.0 to v7.1.3. on AIX61_TL9_SP4, So we are following the steps as below ...
1 ) Bring down the cluster
2 )... (2 Replies)
I have created a VxVM disk group in AIX7.1. I have tried to added this VxVM disk group in powerHA 6.1. But in cluster VxVM DGs are not listing. Is there any other procedure to add vxvm diskgroup to hacmp.
Please share me steps for adding vxvm diskgroup to hacmp. (6 Replies)
hello
ive installed powerha 7.1.3 on two servers aix 6.1.9 6100-09-03-1415
work with dmx4 EMC storgae.
after sync cluster ( terminate with OK ) ive see that the repository disk upper only in one machine :
hdiskpower60 00c7f6b59fc60d9d caavg_private active... (1 Reply)
I have noticed that when I copy files to flash disk and in UI I see message copy completed in notification area on task bar, usually copy is not completed. So, if I eject the media I loose data.
It is very serious problem because may cause loss of valuable and even critical data.
Moreover, when... (2 Replies)
Hi,
At first I have to say that I know nothing about Linux :(
But I am looking for solution how to copy one file from internal Lowrnace HDS flash memory.
Lowrance HDS is Linux based device.
Here is the link for firmware update files for this unit.
At the end of main firmware file there... (62 Replies)
Around a month ago we suffered a prolonged power outage, due to circumstances the servers and storage arrays were still in use when the UPS dropped. One of the servers was running a flash copy on the database while batch processing had commenced. This is the only server that suffered any further... (2 Replies)
hi all
we upgraded hacmp(powerha) from 5.2 to 5.4 recently .
during the failover testing, we found a strange network issue. after standby node took service ip address (172.15.100.8) online at standby NIC, we were able to log in the standby node by telnet 172.15.100.8 which stays at... (1 Reply)