06-30-2011
Guys Thanks a ton for the replies.
The issue has been resolved!
I did the following.
(Didnt create the VG using smitty vg)
1. smitty hacmp>(System Management)CSPOC>Storage>Volume Group>Create a VG
Selected both nodes
>node1
>node2
Selected the participating disks (one on node1 and the other on node2 same PVID)
Then I filled in the info > Resource Group, VG name and thats it..
My VG was created on both the nodes!
Then I went to the same options in point 1 and this time i selected Create a File system
Created my file system on the VG..
and now it is mounted on the active node and it moves with the resource group!
Thank you all
10 More Discussions You Might Find Interesting
1. UNIX for Advanced & Expert Users
Hello,
I have logical volume group of 50GB, in which I have 2 logical volumes, LogVol01 and LogVol02, both are of 10GB.
If I extend LogVol01 further by 10GB, then it keeps the extended copy after logical volume 2. I want to know where it keeps this information
Regards
Himanshu (3 Replies)
Discussion started by: ghimanshu
3 Replies
2. AIX
Does anyone have any simple methods for moving a current logical volume from one volume group to another? I do not wish to move the data from one physical volume to another. Basically, I want to "relink" the logical volume to exist in a different volume group. Any ideas? (2 Replies)
Discussion started by: krisw
2 Replies
3. AIX
Hello,
I am a french computer technician, and i speak English just a little.
On Aix 5.3, I encounter a name conflict logical volume on two volume group.
The first volume lvnode01 is OK in rootvg and mounted. It is also consistent in the ODM
root # lsvg -l rootvg |grep lvnode01 ... (10 Replies)
Discussion started by: dantares
10 Replies
4. AIX
Hi,
I have a requirement to monitor the HACMP Resource Groups. At present in my environment, if the Resource Groups fail over from preferred node to Secondary node we dont get notification.
Can some one help me in creating a scrript. I have more than one RG online. (Max 4 Resource Groups in... (2 Replies)
Discussion started by: srnagu
2 Replies
5. AIX
Hi,
Not sure if this is the correct forum to post this on but maybe a mod could move it if not.
When trying to move a HACMP resource group between lpars on AIX I receive the following.
State not STABLE/RP_RUNNING or ibcasts Join for node 2 rejected,
Clearing in join protocol flag... (11 Replies)
Discussion started by: elmesy
11 Replies
6. AIX
Hi,
I'm new to HACMP. Currently I setup a cluster with nfs cross-mount follow this guide:
kristijan.org NFS cross-mounts in PowerHA/HACMPMy cluster has two nodes: erp01 and erp02.
I'm using nfs4 with filesystem for nfs is: /sapnfs
Cluster start without problems. But I cannnot move RG (with... (3 Replies)
Discussion started by: giobuon
3 Replies
7. AIX
Hello. I am Running AIX 6.1 and PowerHA 6.1
I have an active/active cluster (Prod/Dev) cluster. Each side will failover to the other.
I have on my prod side an active volume group with a file system. The VG is imported on both nodes and active (varried on, file system mounted) on the prod... (3 Replies)
Discussion started by: mhenryj
3 Replies
8. Shell Programming and Scripting
Hi Experts
I need an script to add an disk in to the veritas volume manager disk group.
For example:
# cd /tmp
# view disk
c6t5d2
c6t2d1
c6t3d7
c6t11d2
c7t11d2
c6t11d6
Normally we add the disk like this:
# vxdg -g freedg freedisk01=c6t5d2
# vxdg -g freedg freedisk02=c6t2d1
#... (3 Replies)
Discussion started by: indrajit_preet
3 Replies
9. UNIX for Dummies Questions & Answers
hi,
I want to create a volume group of 200 GB and then create different file systems on that.
please help me out. Its becomes confusing when the PP calculating PP.
I don't understand this concept. (2 Replies)
Discussion started by: kamaldev
2 Replies
10. Red Hat
Hello Guys,
I want to create a file system dedicated for an application installation. But there is no space in volume group to create a new logical volume. There is enough space in other logical volume which is being mounted on /var.
I know we can use that logical volume and create a virtual... (2 Replies)
Discussion started by: vamshigvk475
2 Replies
LEARN ABOUT HPUX
cmrunnode
cmrunnode(1m) cmrunnode(1m)
NAME
cmrunnode - run a node in a high availability cluster
SYNOPSIS
cmrunnode [-v] [node_name...] [-t | -w none]
DESCRIPTION
cmrunnode causes a node to start its cluster daemon to join the existing cluster. This command verifies the network configuration before
causing the node to start its cluster daemon.
To start a cluster on one of its nodes, a user must either be superuser(UID=0), or have an access policy of FULL_ADMIN allowed in the clus-
ter configuration file. See access policy in cmquerycl(1m).
Starting a node will not cause any active packages to be moved to the new node. However, if a package is DOWN, has its switching enabled,
and is able to run on the new node, that package will automatically run there.
If node_name is not specified, the cluster daemon will be started on the local node and will join the existing cluster.
Options
cmrunnode supports the following options:
-v Verbose output will be displayed.
-t Test only. Provide an assessment of the package placement without affecting the current state of the nodes or packages. The -w
option is not required with the -t option as -t does not validate network connectivity, but assumes that all the nodes can meet
any external dependencies such as EMS resources, package subnets, and storage.
node_name...
Start the cluster daemon on the specified node(s).
-w none
By default network probing is performed to check that the network connectivity is the same as when the cluster was configured.
Any anomalies are reported before the cluster daemons are started. The -w none option disables this probing. The option should
only be used if this network configuration is known to be correct from a recent check.
RETURN VALUE
cmrunnode returns the following value:
0 Successful completion.
1 Command failed.
EXAMPLES
Run the cluster daemon on the current node:
cmrunnode
Run the cluster daemons on node1 and node2:
cmrunnode node1 node2
AUTHOR
cmrunnode was developed by HP.
SEE ALSO
cmquerycl(1m), cmhaltcl(1m), cmhaltnode(1m), cmruncl(1m), cmviewcl(1m), cmeval(1m).
Requires Optional Serviceguard Software cmrunnode(1m)