06-30-2011
create a concurrent capable vg on one side, then varyoff, and import it on the other node, of course you need to see the disk on both nodes
make sure you use the same vg major number
importvg -V majornumber -y yourvg hdiskx
I always want to make sure, that my disks have the same hdisk number on both nodes, but this is not absolutely necessary
then discover os changes in hacmp, add the vg to your resource group (make sure it's varied of before you start the cluster)
synchronise the cluster and you are done
I recommend adding lvs und filesystems before you import the vg on the other node, so you don't have to work with cspoc, but with standard lvm commands
this is a general description because I don't have the time to make it more detailed, but ask if you have troubles somewhere
This User Gave Thanks to funksen For This Post:
10 More Discussions You Might Find Interesting
1. UNIX for Advanced & Expert Users
Hello,
I have logical volume group of 50GB, in which I have 2 logical volumes, LogVol01 and LogVol02, both are of 10GB.
If I extend LogVol01 further by 10GB, then it keeps the extended copy after logical volume 2. I want to know where it keeps this information
Regards
Himanshu (3 Replies)
Discussion started by: ghimanshu
3 Replies
2. AIX
Does anyone have any simple methods for moving a current logical volume from one volume group to another? I do not wish to move the data from one physical volume to another. Basically, I want to "relink" the logical volume to exist in a different volume group. Any ideas? (2 Replies)
Discussion started by: krisw
2 Replies
3. AIX
Hello,
I am a french computer technician, and i speak English just a little.
On Aix 5.3, I encounter a name conflict logical volume on two volume group.
The first volume lvnode01 is OK in rootvg and mounted. It is also consistent in the ODM
root # lsvg -l rootvg |grep lvnode01 ... (10 Replies)
Discussion started by: dantares
10 Replies
4. AIX
Hi,
I have a requirement to monitor the HACMP Resource Groups. At present in my environment, if the Resource Groups fail over from preferred node to Secondary node we dont get notification.
Can some one help me in creating a scrript. I have more than one RG online. (Max 4 Resource Groups in... (2 Replies)
Discussion started by: srnagu
2 Replies
5. AIX
Hi,
Not sure if this is the correct forum to post this on but maybe a mod could move it if not.
When trying to move a HACMP resource group between lpars on AIX I receive the following.
State not STABLE/RP_RUNNING or ibcasts Join for node 2 rejected,
Clearing in join protocol flag... (11 Replies)
Discussion started by: elmesy
11 Replies
6. AIX
Hi,
I'm new to HACMP. Currently I setup a cluster with nfs cross-mount follow this guide:
kristijan.org NFS cross-mounts in PowerHA/HACMPMy cluster has two nodes: erp01 and erp02.
I'm using nfs4 with filesystem for nfs is: /sapnfs
Cluster start without problems. But I cannnot move RG (with... (3 Replies)
Discussion started by: giobuon
3 Replies
7. AIX
Hello. I am Running AIX 6.1 and PowerHA 6.1
I have an active/active cluster (Prod/Dev) cluster. Each side will failover to the other.
I have on my prod side an active volume group with a file system. The VG is imported on both nodes and active (varried on, file system mounted) on the prod... (3 Replies)
Discussion started by: mhenryj
3 Replies
8. Shell Programming and Scripting
Hi Experts
I need an script to add an disk in to the veritas volume manager disk group.
For example:
# cd /tmp
# view disk
c6t5d2
c6t2d1
c6t3d7
c6t11d2
c7t11d2
c6t11d6
Normally we add the disk like this:
# vxdg -g freedg freedisk01=c6t5d2
# vxdg -g freedg freedisk02=c6t2d1
#... (3 Replies)
Discussion started by: indrajit_preet
3 Replies
9. UNIX for Dummies Questions & Answers
hi,
I want to create a volume group of 200 GB and then create different file systems on that.
please help me out. Its becomes confusing when the PP calculating PP.
I don't understand this concept. (2 Replies)
Discussion started by: kamaldev
2 Replies
10. Red Hat
Hello Guys,
I want to create a file system dedicated for an application installation. But there is no space in volume group to create a new logical volume. There is enough space in other logical volume which is being mounted on /var.
I know we can use that logical volume and create a virtual... (2 Replies)
Discussion started by: vamshigvk475
2 Replies
LEARN ABOUT HPUX
cmrunnode
cmrunnode(1m) cmrunnode(1m)
NAME
cmrunnode - run a node in a high availability cluster
SYNOPSIS
cmrunnode [-v] [node_name...] [-t | -w none]
DESCRIPTION
cmrunnode causes a node to start its cluster daemon to join the existing cluster. This command verifies the network configuration before
causing the node to start its cluster daemon.
To start a cluster on one of its nodes, a user must either be superuser(UID=0), or have an access policy of FULL_ADMIN allowed in the clus-
ter configuration file. See access policy in cmquerycl(1m).
Starting a node will not cause any active packages to be moved to the new node. However, if a package is DOWN, has its switching enabled,
and is able to run on the new node, that package will automatically run there.
If node_name is not specified, the cluster daemon will be started on the local node and will join the existing cluster.
Options
cmrunnode supports the following options:
-v Verbose output will be displayed.
-t Test only. Provide an assessment of the package placement without affecting the current state of the nodes or packages. The -w
option is not required with the -t option as -t does not validate network connectivity, but assumes that all the nodes can meet
any external dependencies such as EMS resources, package subnets, and storage.
node_name...
Start the cluster daemon on the specified node(s).
-w none
By default network probing is performed to check that the network connectivity is the same as when the cluster was configured.
Any anomalies are reported before the cluster daemons are started. The -w none option disables this probing. The option should
only be used if this network configuration is known to be correct from a recent check.
RETURN VALUE
cmrunnode returns the following value:
0 Successful completion.
1 Command failed.
EXAMPLES
Run the cluster daemon on the current node:
cmrunnode
Run the cluster daemons on node1 and node2:
cmrunnode node1 node2
AUTHOR
cmrunnode was developed by HP.
SEE ALSO
cmquerycl(1m), cmhaltcl(1m), cmhaltnode(1m), cmruncl(1m), cmviewcl(1m), cmeval(1m).
Requires Optional Serviceguard Software cmrunnode(1m)