Sponsored Content
Full Discussion: SAN Migration
Operating Systems AIX SAN Migration Post 302900524 by bakunin on Wednesday 7th of May 2014 08:56:48 AM
Old 05-07-2014
Quote:
Originally Posted by ElizabethPJ
Thank you very much.
Yes I can get downtime for cluster
You do not even need a downtime for the cluster, provided that your SAN disks allow for concurrent access (check with "lsvg <vgname>" if it is "enhanced concurrent capable" and if it is opened in concurrent mode).

If this is indeed the case:

- attach the new disks to both nodes, run "cfgmgr" on both nodes to create the new devices. Check with "lspv".

- add the new volumes to your VGs on the active node:
Code:
extendvg <volumegroup> <hdisk-device>

- mirror the old disks to the new disks. Switch off automatic sync and do a "syncvg" afterwards, it is usually much faster. Use the "-P" switch to let it run in parallel:

Code:
mirrorvg -s <volumegroup> <hdisk-device>
syncvg -P <##> -v <volumegroup>

- after the syncing is finished do a "learning import" on the passive node. The passive node knows only the old disk to be in the VG so far, so use this to reimport the VG definition from:
Code:
importvg -L <volumegroup> <old hdisk-device>

- on the active node unmirror the VG, removing the mirror on the old disk, then move that old disk out of the VG:
Code:
unmirrorvg <volumegroup> <old hdisk-device>
reducevg <volumegroup> <old hdisk-device>

- again do a "learning import" on the passive node to sync the VG definition across cluster nodes, this time from the new disk:
Code:
importvg -L <volumegroup> <new hdisk-device>

Remove the zones from the fabric and delete the old hdisk-devices. When running "cfgmgr" again they should not be discovered any more.

I hope this helps
bakunin
 

10 More Discussions You Might Find Interesting

1. HP-UX

SAN migration

Hi, I'm going to be involved in a migration of SAN islands to one big SAN. I've not worked with SANs before and I'm not sure how to approach this. I suspect the disk devices on the HP servers are going to change, when the EVA's and servers are plugged into this new Cisco 9509 switch. Any... (0 Replies)
Discussion started by: Hils
0 Replies

2. Solaris

Thoughts/experiences of SAN attaching V880 to EMC SAN

Hi everyone, I wonder if I can canvas any opinions or thoughts (good or bad) on SAN attaching a SUN V880/490 to an EMC Clarion SAN? At the moment the 880 is using 12 internal FC-AL disks as a db server and seems to be doing a pretty good job. It is not I/O, CPU or Memory constrained and the... (2 Replies)
Discussion started by: si_linux
2 Replies

3. AIX

IBM SAN TO SAN Mirroring

Has anyone tried SAN to SAN mirroring on IBM DS SAN Storage. DS5020 mentions Enhanced Remote Mirror to multi-LUN applications I wonder if Oracle High availibility can be setup using Remote Mirror option of SAN ? (1 Reply)
Discussion started by: filosophizer
1 Replies

4. HP-UX

SAN Migration of HP-UX hosts

Hello gurus, I am a SAN Admin - not very familiar with the HPUX administration - so need help with the steps in regards to the migration I need to do at my client place. Environment: Migrating from CX4 to VMAX - using OR/Hot Pull. Here are the steps I have put together - HPUX gurus please... (5 Replies)
Discussion started by: jps460
5 Replies

5. Solaris

Need Help On Solaris Cluster Steps For San Migration

I am going to do a SAN Array migration and need the sequence of steps required on the Solaris cluster before moving the old array luns to new array luns. Here are the steps and I need info on the bold points:( I might even be wrong on the sequence of steps please correct me if I am wrong) 1.... (0 Replies)
Discussion started by: sunshinedeepu
0 Replies

6. HP-UX

SAN Migration question

Hi, I am very new to HP-UX, and we're going to be doing a SAN migration. We're going to take down the machine, and zone it to the new SAN. My question is, will the device names change and will that interfere with the LVM? If the new disks come in with different device names, how would I... (3 Replies)
Discussion started by: BG_JrAdmin
3 Replies

7. Red Hat

Volume group not activated at boot after SAN migration

I have an IBM blade running RHEL 5.4 server, connected to two Hitachi SANs using common fibre cards & Brocade switches. It has two volume groups made from old SAN LUNs. The old SAN needs to be retired so we allocated LUNs from the new SAN, discovered the LUNs as multipath disks (4 paths) and grew... (4 Replies)
Discussion started by: rbatte1
4 Replies

8. AIX

AIX - FC Switch migration, SAN Migration question!

I'm New to AIX / VIOS We're doing a FC switch cutover on an ibm device, connected via SAN. How do I tell if one path to my remote disk is lost? (aix lvm) How do I tell when my link is down on my HBA port? Appreciate your help, very much! (4 Replies)
Discussion started by: BG_JrAdmin
4 Replies

9. Solaris

How to identify if disk is attached to SAN and assist in migration.?

I am working on VM host and collecting data to identify the type of storage attached to the server which will be migrated to VNX. it has one ldom created on it luxadm probe output --- No Network Array enclosures found in /dev/es Found Fibre Channel device(s): Node... (7 Replies)
Discussion started by: kpatel786
7 Replies

10. Filesystems, Disks and Memory

Faster way: SAN hd to SAN hd copying

hi! i got a rhel 6.3 host that already have an xfs filesystem mounted from a SAN (let's call it SAN-1) whose size is 9TB. i will be receiving another SAN (let's call it SAN-2) storage of 15TB size. this new addition is physically on another SAN storage. SAN-1 is on a Pillar storage while the new... (6 Replies)
Discussion started by: rino19ny
6 Replies
scconf_dg_vxvm(1M)					  System Administration Commands					scconf_dg_vxvm(1M)

NAME
scconf_dg_vxvm - add, change, or update VxVM device group configuration. SYNOPSIS
scconf -a -D type=vxvm,devicegroup-options[,localonly=true|false] scconf -c -D devicegroup-options[,sync] scconf -r -D name=devicegroupname DESCRIPTION
Note - Beginning with the Sun Cluster 3.2 release, Sun Cluster software includes an object-oriented command set. Although Sun Cluster software still supports the original command set, Sun Cluster procedural documentation uses only the object-oriented command set. For more infor- mation about the object-oriented command set, see the Intro(1CL) man page. The following information is specific to the scconf command. To use the equivalent object-oriented commands, see the cldevicegroup(1CL) man page. The scconf_dg_vxvm command is used to add, change, and remove the VERITAS Volume Manager (VxVM) device groups to the Sun Cluster device- groups configuration. OPTIONS
See the scconf(1M) man page for the list of supported generic device-group options. The following action options describe the actions that the command performs. Only one action option is allowed in the command. -a Add a VxVM device group to the cluster configuration. The -a (add) option adds a new VxVM device group to the Sun Cluster device-groups configuration. With this option you define a name for the new device group, specify the nodes on which this group can be accessed, and specify a set of properties used to control actions. For VxVM device groups, you can only assign one VxVM disk group to a device group, and the device-group name must always match the name of the VxVM disk group. You cannot create a VxVM device group unless you first import the corresponding VxVM disk group on one of the nodes in that device's node list. Before you can add a node to a VxVM device group, every physical disk in the disk group must be physically ported to that node. After you register the disk group as a VxVM device group, you must first deport the disk group from the current node owner and turn off the auto-import flag for the disk group. To create a VxVM device group for a disk group, you must run the scconf command from the same node where the disk group was created. -c Change the ordering of the node preference list, change preference and failback policy, and change the desired number of sec- ondaries. The scconf -c (change) option changes the order of the potential primary node preference, to enable or disable failback, to add more global devices to the device group, and to change the desired number of secondaries. The sync suboption is used to synchronize the clustering software with VxVM disk-group volume information. The sync suboption is only valid with the change form of the command. Use the sync suboption whenever you add or remove a VxVM volume from a VxVM device group or change any volume attribute, such as owner, group, or access permissions. Also use the sync suboption to change a device-group configuration to a replicated or non-replicated configuration. For device groups that contain disks that use Hitachi TrueCopy data replication, this sync suboption synchronizes the device- group configuration and the replication configuration. This synchronization makes Sun Cluster software aware of disks that are configured for data replication and enables the software to handle failover or switchover as necessary. After you create a Solaris Volume Manager disk set that contain disks that are configured for replication, you must run the sync suboption for the corresponding svm or sds device group. A Solaris Volume Manager disk set is automatically registered with Sun Cluster software as an svm or sds device group, but replication information is not synchronized at that time. For newly created vxvm and rawdisk device-group types, you do not need to manually synchronize replication information for the disks. When you register a VxVM disk group or a raw-disk device group with Sun Cluster software, the software automatically discovers any replication information on the disks. To change the order-of-node preference list from false to true, you must specify in the nodelist all the nodes that currently exist in the device group. You must also set the preferenced suboption to true. If you do not specify the preferenced suboption with the change form of the command, the already established true or false setting is used. If a disk group should be accessed by only one node, it should be configured with the localonly property set to true. This property setting puts the disk group outside the control of Sun Cluster software. Only one node can be specified in the node list to create a localonly disk group. To change a local-only disk group to a regular VxVM disk group, set the localonly property to false. -r Remove the specified VxVM device group from the cluster. The -r (remove) option removes a VxVM device group from the Sun Cluster device-groups configuration. You can also use this form of command to remove the nodes from the VxVM device group configuration. EXAMPLES
Example 1 Using scconf Commands The following scconf commands create a VxVM device group, change the order of the potential primary nodes, change the preference and fail- back policy for the device group, change the desired number of secondaries, and remove the VxVM device group from the cluster configura- tion. host1# scconf -a -D type=vxvm,name=diskgrp1,nodelist=host1:host2:host3,preferenced=false,failback=enabled host1# scconf -c -D name=diskgrp1,nodelist=host2:host1:host3,preferenced=true,failback=disabled,numsecondaries=2 sync host1# scconf -r -D name=diskgrp1,nodelist=node1 ATTRIBUTES
See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Availability |SUNWscu | +-----------------------------+-----------------------------+ |Interface Stability |Evolving | +-----------------------------+-----------------------------+ SEE ALSO
Intro(1CL), cldevicegroup(1CL), scconf(1M), attributes(5) Sun Cluster 3.2 2 Aug 2006 scconf_dg_vxvm(1M)
All times are GMT -4. The time now is 10:07 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy