Sponsored Content
Full Discussion: Failed mpio path on AIX5.3
Operating Systems AIX Failed mpio path on AIX5.3 Post 302875029 by -=XrAy=- on Monday 18th of November 2013 10:00:08 AM
Old 11-18-2013
Hi,

it looks like that the specific path is not (longer) available.
Maybe your storage admin has removed the mapping for this volume or the zoning for this path has been removed.

An lspath and lscfg -vl hdisk10 would be helpfull.

Regards

Last edited by -=XrAy=-; 11-18-2013 at 11:06 AM..
 

10 More Discussions You Might Find Interesting

1. UNIX for Advanced & Expert Users

AIX MPIO and EMC

We are looking at running MPIO for it's redundancy and load balancing benefits. Does anyone know what pieces of software or modules are needed on the VIO server to get load balancing to work. Remember we are using EMC's DMX3500 storage system. We no longer want to use Powerpath. :rolleyes: ... (2 Replies)
Discussion started by: vxg0wa3
2 Replies

2. Solaris

path 108528-29 failed

Hi All, The path requires me to go Single User mode. On single user mode, the patch still won't continue: dhcpagent supports a new keyword in /etc/default/dhcpagent. Please merge the existing /etc/default/dhcpagent with the patched version of /etc/default/dhcpagent shipped with this... (2 Replies)
Discussion started by: itik
2 Replies

3. AIX

AIX native MPIO

Hi folks, does anybody have a link to a documentation how to implement native MPIO on AIX? We are using EMC PowerPath and Datacore SanSymphony/Cambex for this so far and I wasn't able to find a good description on that topic. All I know so far is that mkpath, chpath and lspath are used to... (3 Replies)
Discussion started by: zaxxon
3 Replies

4. AIX

How to upgrade AIX5.3TL6-07 to AIX5.3TL8-04?

Hello All, After creating lpp_source/spot from AIX 5.3TL6-07 DVD on the NIM(AIX6.1), I went online and download the fix pack for reaching TL8-04. (it was less than 200Mgegs total) then I did define a newer lpp_source530TL8-04 via a copy from the older lpp_source530TL6-07 (I diduse a command... (2 Replies)
Discussion started by: sangers
2 Replies

5. AIX

MPIO Driver

On a particular LPAR, I was running AIX 5.3 TL 3. On Monday I did an update of the LPAR to 5.3 TL 9 SP2. The install was smooth, but then I ran into a problem. The MPIO driver does not work with LSI's StoreAge (SVM4). I did some looking, and looks like 5.3 TL3 = IBM.MPIO 5.3.0.30 5.3... (0 Replies)
Discussion started by: clking
0 Replies

6. AIX

MPIO reliability

Hi, we have a vew boxes using MPIO and they are connected to some virtualization software managing some disk subsystems, offering volumes to the AIX boxes. Sometimes when a cable has been plugged out for a test or when a real problem occurs, using lspath to show the state of the paths shows... (8 Replies)
Discussion started by: zaxxon
8 Replies

7. AIX

MPIO - list of supported arrays

Hello, we are planning to deploy some of our databases on AIX/LPAR based servers (we didn't bought it yet ...). IBM's engineers says that if we want to boot them from SAN the hardware array has to be compatible with MPIO but they don't want to deliver any document with list of arrays supported... (5 Replies)
Discussion started by: andy55
5 Replies

8. AIX

Need Help with SDD / SDDPCM / MPIO

This is getting very confusing for me, and appreciate if someone can help. Platform: Power VM ( Virtual I/O Server) ioslevel 2.1.3.10-FP23 # oslevel -s 6100-05-00-0000 Storage: IBM DS4300 Two HBAs - Dual Port Fibre Adapter Channels Each card has two ports , so a total of 4 ports going... (3 Replies)
Discussion started by: filosophizer
3 Replies

9. AIX

DISK and MPIO

Hello, I have some concerns over the disk management of my AIX system. For example server1 / > lspv hdisk0 00fa6d1288c820aa rootvg active hdisk1 00fa6d1288c8213c vg_2 active hdisk2 00c1cc14d6de272b ... (6 Replies)
Discussion started by: Phat
6 Replies

10. AIX

issue AIX MPIO path closed

Hello, I have encountered the issue on my AIX 6.1. As you can see below, the hdisk1 & hdisk8 are not recognized by the system / > lspv hdisk0 00fa6d1288c820aa rootvg active hdisk1 00fa6d1288c8213c None hdisk2 ... (9 Replies)
Discussion started by: Phat
9 Replies
vgreduce(1M)															      vgreduce(1M)

NAME
vgreduce - remove physical volumes from an LVM volume group SYNOPSIS
autobackup] vg_name pv_path ... autobackup] vg_name pv_path autobackup] vg_name DESCRIPTION
The command removes each physical volume specified by a pv_path argument from volume group vg_name. The command with option removes all missing physical volumes from the volume group. All but one physical volume can be removed. The last physical volume must remain in the volume group so that the logical volume driver can continue to operate. The last physical volume in the volume group can be removed with the command (see vgremove(1M)). Before executing remove all logical volumes residing on each physical volume represented by a pv_path by executing the command (see lvre- move(1M)). Any physical volume in the pv_path list that is also a member of a physical volume group (as defined in is also removed from the physical volume group. If the physical volume happens to be the last one in the physical volume group, the physical volume group is also removed from the volume group. When a physical volume in the pv_path list has multiple the physical volume is not removed from the volume group until all the links to the volume are removed. When a physical volume in the pv_path list is the (in use) to a physical volume, removing the link forces LVM to switch to the (For information on alternate links, see lvm(7)). When the pv_path removed is an to the device, only the link is removed; the volume group and physical volume are otherwise unchanged. When a physical volume is added into volume groups version 2.0 or higher, the number of physical extents available on this device will be added into of the volume group (see vgdisplay(1M)). If the size of physical volume is such that would exceed the VG Max Extents of the volume group, the number of physical extents available on this device will be reduced accordingly (see vgextend(1M)). If a subsequent on a different physical volume, frees up space, then the number of extents allocated to the partial physical volume is automatically increased. Under these circumstances the command will display the old and the new values for the count for this partial physical volume (see pvdis- play(1M)). Options and Arguments recognizes the following options and arguments: Set automatic backup for this invocation of this command. autobackup can have one of the following values: Automatically back up configuration changes made to the volume group. This is the default. After this command executes, the command (see vgcfgbackup(1M)) is executed for the volume group. Do not back up configuration changes this time. Force the reduction of missing physical volumes from the specified volume group. obtains the name of each physical volume (PV) belonging to the volume group. It then reads the kernel PV structures to work out which PVs are missing. PVs which are missing will be candidates for removal. If all the physical extents on the missing PV are free then it will be removed from the volume group. Otherwise will report the physical to logical extent mapping. For missing PVs which have extents in use, you must free up all the extents by using the or commands (see lvreduce(1M) and lvremove(1M)) and re-run with the option. This option is most commonly used when the command (see vgdisplay(1M)) shows "Cur PV" higher than "Act PV" and all of the PVs belonging to the volume group are attached. This option only works on PVs and not on links. See the option for details on handling missing links. Removes the specified pv_path/s from the or file. This task will only be performed if the pv_path is currently marked as missing from the volume group. This option was mainly designed for the problem where the or the command place too many links beyond the max limit allowed in the or file. Currently the max limit is 8 paths to a PV (seven alternates and one primary). In this situation, invoking the command without the option will not resolve the condition because the path is not attached to the volume group. Similarly, this condition can not be overcome by invoking with the option, as the same works on physical volumes rather than the links. pv_path The block device path name of a physical volume. vg_name The path name of the volume group. Shared Volume Group Considerations For volume group version 1.0 and 2.0, cannot be used if the volume group is activated in shared mode. For volume groups version 2.1 (or higher), can be performed when activated in either shared, exclusive, or standalone mode. Note that the daemon must be running on all the nodes sharing a volume group activated in shared mode. See lvmpud(1M). In the case a PV alternate link is reduced, is updated only on the server node. In the case a PV is reduced, all the PV device special files are removed from on all the nodes sharing the volume group. When a node wants to share the volume group, the user must first execute a if physical volumes were removed at the time the volume group was not activated on that node. LVM shared mode is currently only available in Serviceguard clusters. EXTERNAL INFLUENCES
Environment Variables determines the language in which messages are displayed. If is not specified or is null, it defaults to "C" (see lang(5)). If any internationalization variable contains an invalid setting, all internationalization variables default to "C" (see environ(5)). EXAMPLES
Remove physical volume from volume group Force, reduction of missing PVs from volume group: The following messages will appear after missing PVs has been removed successfully: PV with key 0 successfully deleted from vg Repair done, please do the following steps.....: 1. Save or to another file. 2. Remove or 3. Use to recreate or 4. NOW use to save the LVM setup. Remove the physical volume from the volume group when the physical volume is still in the or file but is currently marked as missing from the volume group: SEE ALSO
lvmpud(1M), vgchange(1M), vgcreate(1M), vgdisplay(1M), vgextend(1M), intro(7), lvm(7). vgreduce(1M)
All times are GMT -4. The time now is 04:00 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy