Sponsored Content
Full Discussion: VXVM in POWERHA 6.1
Operating Systems AIX VXVM in POWERHA 6.1 Post 302910967 by bakunin on Tuesday 29th of July 2014 10:41:44 AM
Old 07-29-2014
HACMP does not deal with disks (disk devices), it only deals with volume groups. Basically, a "resource group" in HACMP is:

- a bundle of filesystems, organized in certain volume groups
- a number of processes to be started/stopped
- an (several) IP addresses to serve up data/services

When a resource group starts: the VGs are activated (varyonvg), the respective filesystems are mounted, the processes are started (via start-scripts) and the IP-address is activated on some interface.

When the resource group stops this process is reversed: the IP address is deleted from the interface, the processes are stopped (via stop-scripts) and the filesystems are unmounted, then the VGs are deactivated.

How and from what devices these VGs are built is transparent for HACMP. The only requirement is that HACMP is able to activate the VGs from every cluster node.

I hope this helps.

bakunin
 

10 More Discussions You Might Find Interesting

1. AIX

strange route issue in powerha 5.4?

hi all we upgraded hacmp(powerha) from 5.2 to 5.4 recently . during the failover testing, we found a strange network issue. after standby node took service ip address (172.15.100.8) online at standby NIC, we were able to log in the standby node by telnet 172.15.100.8 which stays at... (1 Reply)
Discussion started by: rs6000er
1 Replies

2. AIX

PowerHA - assistant can't find network

I've configured multiple clusters using PowerHA 5.5, but never using the 'Two-node cluster configuration assistant'. I removed all the clusters, clean setup, and gave the assistant a try. I've filled in all the necessary options: Communication Path to Takeover Node Application Server ... (11 Replies)
Discussion started by: Koen87
11 Replies

3. AIX

PowerHA, same subnet for persistent and service ip

I am new in AIX and please forgive my poor english. I know that AIX allow same subnet IPs for different interfaces, which result in multipath routing / route striping. My question is, is there any best practice for the persistent and service IP with same subnet to stay on same interface, or... (5 Replies)
Discussion started by: skeyeung
5 Replies

4. AIX

PowerHA Disk on VIO Server

Hello AIX GURU's Can anybody tell me the steps to crate shared VG (enhanced concurent) for my Lpars from VIO server? my questions are: 1. Should I crate Enhanced Concurent VG in VIO and map it using virtual Scsi to Lpar? or 2. Can I just create virtual SCSI in VIO and map to Lpar and... (1 Reply)
Discussion started by: Vit0_Corleone
1 Replies

5. AIX

PowerHA with 3rd node for communication

Looking to find PowerHA with 3rd node used for communication (voting node) only. Does anyone use such configuration with IBM products? (2 Replies)
Discussion started by: gogogo
2 Replies

6. AIX

PowerHA HACMP on VIOS servers

Few questions regarding Power HA ( previously known as HACMP) and VIOS POWERVM IVM ( IBM Virtualization I/O Server ) Is it possible to create HACMP cluster between two VIOS servers Physical Machine_1 VIOS_SERVER_1 LPAR_1 SHARED_DISK_XX VIOS_SERVER_2 Physical Machine_2 LPAR_2... (6 Replies)
Discussion started by: filosophizer
6 Replies

7. AIX

Powerha on AIX 6.1.9

hello ive installed powerha 7.1.3 on two servers aix 6.1.9 6100-09-03-1415 work with dmx4 EMC storgae. after sync cluster ( terminate with OK ) ive see that the repository disk upper only in one machine : hdiskpower60 00c7f6b59fc60d9d caavg_private active... (1 Reply)
Discussion started by: ariec
1 Replies

8. AIX

PowerHA; Import flash copy of VG?

Hey guys, wondering if this is possible to accomplish. PowerHA SystemMirror7.1 active/passive cluster. Restored a clustered system in test and upgraded to the latest version of PowerHA. Things are working great. However in the interest of time during cut over to the new system, I was hoping to... (4 Replies)
Discussion started by: j_aix
4 Replies

9. AIX

PowerHa 7.1 Migration check failed

Hello Guys, I was trying to upgrade the PowerHA from 6.1.0.9 to 7.1.3 in prod cluster. Have snapshot migration and run climgcheck. 1. Run climgcheck on node1(kul1pqcmur01) and it success 2. Upgrade the cluster sw 3. Run climgcheck on node2(kul1pqcmur02) and it failed with below errors ... (6 Replies)
Discussion started by: novaothers
6 Replies

10. AIX

Upgrading PowerHA from 7.1.0 to 7.1.3

Hi All, As per the IBM upgrade/support matrix IBM Techdocs Technote: PowerHA for AIX Version Compatibility Matrix we can't do online upgrade or rolling migration from powerha v7.1.0 to v7.1.3. on AIX61_TL9_SP4, So we are following the steps as below ... 1 ) Bring down the cluster 2 )... (2 Replies)
Discussion started by: linux.amrit
2 Replies
scconf_dg_vxvm(1M)					  System Administration Commands					scconf_dg_vxvm(1M)

NAME
scconf_dg_vxvm - add, change, or update VxVM device group configuration. SYNOPSIS
scconf -a -D type=vxvm,devicegroup-options[,localonly=true|false] scconf -c -D devicegroup-options[,sync] scconf -r -D name=devicegroupname DESCRIPTION
Note - Beginning with the Sun Cluster 3.2 release, Sun Cluster software includes an object-oriented command set. Although Sun Cluster software still supports the original command set, Sun Cluster procedural documentation uses only the object-oriented command set. For more infor- mation about the object-oriented command set, see the Intro(1CL) man page. The following information is specific to the scconf command. To use the equivalent object-oriented commands, see the cldevicegroup(1CL) man page. The scconf_dg_vxvm command is used to add, change, and remove the VERITAS Volume Manager (VxVM) device groups to the Sun Cluster device- groups configuration. OPTIONS
See the scconf(1M) man page for the list of supported generic device-group options. The following action options describe the actions that the command performs. Only one action option is allowed in the command. -a Add a VxVM device group to the cluster configuration. The -a (add) option adds a new VxVM device group to the Sun Cluster device-groups configuration. With this option you define a name for the new device group, specify the nodes on which this group can be accessed, and specify a set of properties used to control actions. For VxVM device groups, you can only assign one VxVM disk group to a device group, and the device-group name must always match the name of the VxVM disk group. You cannot create a VxVM device group unless you first import the corresponding VxVM disk group on one of the nodes in that device's node list. Before you can add a node to a VxVM device group, every physical disk in the disk group must be physically ported to that node. After you register the disk group as a VxVM device group, you must first deport the disk group from the current node owner and turn off the auto-import flag for the disk group. To create a VxVM device group for a disk group, you must run the scconf command from the same node where the disk group was created. -c Change the ordering of the node preference list, change preference and failback policy, and change the desired number of sec- ondaries. The scconf -c (change) option changes the order of the potential primary node preference, to enable or disable failback, to add more global devices to the device group, and to change the desired number of secondaries. The sync suboption is used to synchronize the clustering software with VxVM disk-group volume information. The sync suboption is only valid with the change form of the command. Use the sync suboption whenever you add or remove a VxVM volume from a VxVM device group or change any volume attribute, such as owner, group, or access permissions. Also use the sync suboption to change a device-group configuration to a replicated or non-replicated configuration. For device groups that contain disks that use Hitachi TrueCopy data replication, this sync suboption synchronizes the device- group configuration and the replication configuration. This synchronization makes Sun Cluster software aware of disks that are configured for data replication and enables the software to handle failover or switchover as necessary. After you create a Solaris Volume Manager disk set that contain disks that are configured for replication, you must run the sync suboption for the corresponding svm or sds device group. A Solaris Volume Manager disk set is automatically registered with Sun Cluster software as an svm or sds device group, but replication information is not synchronized at that time. For newly created vxvm and rawdisk device-group types, you do not need to manually synchronize replication information for the disks. When you register a VxVM disk group or a raw-disk device group with Sun Cluster software, the software automatically discovers any replication information on the disks. To change the order-of-node preference list from false to true, you must specify in the nodelist all the nodes that currently exist in the device group. You must also set the preferenced suboption to true. If you do not specify the preferenced suboption with the change form of the command, the already established true or false setting is used. If a disk group should be accessed by only one node, it should be configured with the localonly property set to true. This property setting puts the disk group outside the control of Sun Cluster software. Only one node can be specified in the node list to create a localonly disk group. To change a local-only disk group to a regular VxVM disk group, set the localonly property to false. -r Remove the specified VxVM device group from the cluster. The -r (remove) option removes a VxVM device group from the Sun Cluster device-groups configuration. You can also use this form of command to remove the nodes from the VxVM device group configuration. EXAMPLES
Example 1 Using scconf Commands The following scconf commands create a VxVM device group, change the order of the potential primary nodes, change the preference and fail- back policy for the device group, change the desired number of secondaries, and remove the VxVM device group from the cluster configura- tion. host1# scconf -a -D type=vxvm,name=diskgrp1,nodelist=host1:host2:host3,preferenced=false,failback=enabled host1# scconf -c -D name=diskgrp1,nodelist=host2:host1:host3,preferenced=true,failback=disabled,numsecondaries=2 sync host1# scconf -r -D name=diskgrp1,nodelist=node1 ATTRIBUTES
See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Availability |SUNWscu | +-----------------------------+-----------------------------+ |Interface Stability |Evolving | +-----------------------------+-----------------------------+ SEE ALSO
Intro(1CL), cldevicegroup(1CL), scconf(1M), attributes(5) Sun Cluster 3.2 2 Aug 2006 scconf_dg_vxvm(1M)
All times are GMT -4. The time now is 07:57 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy