Sponsored Content
Full Discussion: EMC Failover
Operating Systems Solaris EMC Failover Post 302366128 by glenioborges on Thursday 29th of October 2009 12:53:48 AM
Old 10-29-2009
EMC Failover

Hi guys,

I'm running vxdmp and powerpath at the same time. Vxdmp for internal disks and powerpath for external.

The problem is that, on the failover tests which a fiber cable should be removed, the system cannot recognize the disks.

Any hints on how to configure powerpath in order to correctly handle the failover?

Thank you
 

10 More Discussions You Might Find Interesting

1. Solaris

Interfaces do not failover

Would anyone know why a V440 Solaris 9 machine's interfaces (ce0 and ce1) will not failover? Both interfaces are on the same subnet, but different IPs. I can pull the plug on once interface (ce0) and ce1 will be accessible. I can reverse it and pull ce1, and ce0 will NOT be accessible. ... (5 Replies)
Discussion started by: StorageGuy
5 Replies

2. HP-UX

is anybody using EMC symmetrix?

Hello, I'm looking for reference sites using HP-UX and EMC symmetrix disk. Then, May I ask you questions? (6 Replies)
Discussion started by: cooldugong
6 Replies

3. Solaris

Emc

Dear gentelmen kindly please update me me how can i know disks on EMC and get size for all disks on EMC? (1 Reply)
Discussion started by: magasem
1 Replies

4. HP-UX

Emc

Dear gentelmen kindly please update me how can i know disks on EMC and get size for all disks on EMC? (0 Replies)
Discussion started by: magasem
0 Replies

5. UNIX for Advanced & Expert Users

AIX MPIO and EMC

We are looking at running MPIO for it's redundancy and load balancing benefits. Does anyone know what pieces of software or modules are needed on the VIO server to get load balancing to work. Remember we are using EMC's DMX3500 storage system. We no longer want to use Powerpath. :rolleyes: ... (2 Replies)
Discussion started by: vxg0wa3
2 Replies

6. UNIX for Advanced & Expert Users

directories failover

Hi, I have 3 directories (/1 /2 and /3) with same contents and want to achieve fail over among them. The Solution i thought of: 1. link /data to /1 (and provide users with /data to access) 2. CRON a script (to run every minute) to verify if /1 exist or not, if it does not exist (or its... (3 Replies)
Discussion started by: prvnrk
3 Replies

7. AIX

failover on 4.5 hacmp

Hi All, How do I trigger the failover on the second hacmp server? Please give me idea and I will do the rest. Thanks, itik (2 Replies)
Discussion started by: itik
2 Replies

8. IP Networking

IP failover needed

All, I am looking to setup a simple IP failover setup, but it seems every place i look has some difficult setup like pacemaker or LVS. I only want to handle the IP piece to failover a VIP. Any suggestions? I checked out keepalived and that my work for me. (4 Replies)
Discussion started by: markdjones82
4 Replies

9. Gentoo

How to failover the cluster ?

How to failover the cluster ? GNU/Linux By which command, My Linux version 2008 x86_64 x86_64 x86_64 GNU/Linux What are the prerequisites we need to take while failover ? if any Regards (3 Replies)
Discussion started by: sidharthmellam
3 Replies

10. AIX

Help with EMC BCV device

I'm trying to auto-mount EMC Symmetrix BCV device at boot. but having problem making BCV available. I put script called mkbcv to the inittab and engineer suggested to add 120 sec sleep between cfgmgr so I did that also. My mkbcv script seems to be working fine, it says "hdisk4 Available" ... (1 Reply)
Discussion started by: shuhei365
1 Replies
vgmove(1M)																vgmove(1M)

NAME
vgmove - move data from an old set of disks in a volume group to a new set of disks SYNOPSIS
autobackup] diskmapfile vg_name autobackup] diskfile diskmapfile vg_name DESCRIPTION
The command migrates data from the existing set of disks in a volume group to a new set of disks. After the command completes successfully, the new set of disks will belong to the same volume group. The command is intended to migrate data on a volume group from old storage to new storage. The diskmapfile specifies the list of source disks to move data from, and the list of destination disks to move data to. The user may choose to list only a subset of the existing physical volumes in the volume group that need to be migrated to a new set of disks. The format of the diskmapfile file is shown below: source_pv_1 destination_pv_1_1 destination_pv_1_2 .... source_pv_2 destination_pv_2_1 destination_pv_2_2 .... .... source_pv_n destination_pv_n_1 destination_pv_n_2 .... If a destination disk is not already part of the volume group, it will be added using see vgextend(1M). Upon successful completion of the source disk will be automatically removed from the volume group using see vgreduce(1M). After successful migration, the destination disks are added to the LVM configuration files; namely, or The source disks along with their alternate links are removed from the LVM configuration files. A sample diskmapfile is shown below: /dev/disk/disk1 /dev/disk/disk51 /dev/disk/disk52 /dev/disk/disk2 /dev/disk/disk51 /dev/disk/disk3 /dev/disk/disk53 The diskmapfile can be manually created, or it can be automatically generated using the diskfile and diskmapfile options. The argument diskfile contains a list of destination disks, one per line such as the sample file below: /dev/disk/disk51 /dev/disk/disk52 /dev/disk/disk53 When the option is given, reads a list of destination disks from diskfile, generates the source to destination mapping, and saves it to diskmapfile. The volume group must be activated before running the command. If the command is interrupted before it completes, the volume group is in the same state it was at the beginning of the command. The migration can be continued by running the command with the same options and disk mapping file. Options and Arguments The command recognizes the following options and arguments: vg_name The path name of the volume group. Set automatic backup for this invocation of autobackup can have one of the following values: Automatically back up configuration changes made to the volume group. This is the default. After this command executes, the command is executed for the volume group; see vgcfgbackup(1M). Do not back up configuration changes this time. Specify the name of the file containing the source to destination disk mapping. If the option is also given, will generate the disk mapping and save it to this filename. (Note that if the diskmapfile already exists, the file will be overwritten). Otherwise, will perform the data migration using this diskmapfile. Specify the name of the file containing the list of destination disks. This option is used with the option to generate the diskmapfile. When the option is used, no volume group data is moved. Preview the actions to be taken but do not move any volume group data. Shared Volume Group Considerations For volume group version 1.0 and 2.0, cannot be used if the volume group is activated in shared mode. For volume groups version 2.1 (or higher), can be performed when activated in either shared, exclusive, or standalone mode. Note that the daemon must be running on all the nodes sharing a volume group activated in shared mode. See lvmpud(1M). When a node wants to share the volume group, the user must first execute a if physical volumes were moved in or out of the volume group at the time the volume group was not activated on that node. LVM shared mode is currently only available in Serviceguard clusters. EXTERNAL INFLUENCES
Environment Variables determines the language in which messages are displayed. If is not specified or is null, it defaults to "C" (see lang(5)). If any internationalization variable contains an invalid setting, all internationalization variables default to "C" (see environ(5)). EXAMPLES
Move data in volume group from to After the migration, remove from the volume group: Generate a source to destination disk map file for where the destination disks are and SEE ALSO
lvmpud(1M), pvmove(1M), vgcfgbackup(1M), vgcfgrestore(1M), vgextend(1M), vgreduce(1M), intro(7), lvm(7). vgmove(1M)
All times are GMT -4. The time now is 11:15 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy