Sponsored Content
Operating Systems AIX Possible to connect a LPAR to two different VIO's? Post 302505067 by funksen on Wednesday 16th of March 2011 05:51:21 AM
Old 03-16-2011
if it's the same storage type that uses the same drivers, it's also possible to run two subsystems over the same fibre channel adapter
to make sure check out the documentation of your storage

of course the san zoning has to be adapted

if it's really the same storage with just two different arrays, just map the new luns to the existing vio server, map it to the lpar, mirror or migratepv the hdisks
 

10 More Discussions You Might Find Interesting

1. AIX

Restoring mksysb to vio lpar?

I have a standard template I deploy for each of my AIX servers. It's in a mksysb format which I pulled via the nim server after I originally set it up. I'm trying to restore this mksysb over the wire to a remote system. The remote system is an lpar running under vio (no HMC attached). The disk... (1 Reply)
Discussion started by: scottsl
1 Replies

2. UNIX for Dummies Questions & Answers

Unix VIO vs Dynamic LPAR

Can you have a Dynamic LPAR without VIO and if yes is there a benefit to having both? I am attempting to implement VIO and Resource pooling at my company and want to understand the limitations. Also, I have been told the managing and allocation of the Logical Volumes is a time consuming task as... (0 Replies)
Discussion started by: bombers73
0 Replies

3. AIX

Lost root password in LPAR (VIO Server).

Hi, I've lost the root password in a virtual partition created in a IBM Vio Server. May I recover that somehow? Thanks (2 Replies)
Discussion started by: r4mp4g3
2 Replies

4. AIX

LPAR and vio disk mapping

We have a frame the uses 2 vios that assign disk storage to LPAR's. We have a LPAr with multiple disk and I want to know how do I tell which vio is serving the disk. For example the LPAr has hdisk 0, 1, 2, 3 all the same size. I want to know which vio is serving hdisk0, 1. (4 Replies)
Discussion started by: daveisme
4 Replies

5. AIX

VIO Backing Disk LPAR how to find which one ?

hello Folks, my vio: $ lsmap -all SVSA Physloc Client Partition ID --------------- -------------------------------------------- ------------------ vhost0 U9117.MMA.6534BE4-V2-C11 0x00000003 VTD ... (0 Replies)
Discussion started by: filosophizer
0 Replies

6. AIX

CDROM Issue on LPAR HMC VIO

Whenever I assign/unassign a CDROM ( RAID CTRL ) to Active LPAR from HMC, I have to reboot the LPAR. How can I do it without rebooting the LPARS. POWER6 with HMC LPARS = AIX 6.1 Any info developerWorks : AIX and UNIX : PowerVM Forum : Moving CD-ROM/DVD-ROM dynamically ... But... (5 Replies)
Discussion started by: filosophizer
5 Replies

7. AIX

LPAR - What is my VIO (hostname, IP, ...)

I am wondering if there is a command I can run from **with-in** LPAR server (server01) to find out the hostname or IP address or other parameters for VIO server in charge of server01? (1 Reply)
Discussion started by: ms1232
1 Replies

8. AIX

VIO server does not show new LPAR

Hi All, I've been trying to create a new LPAR for some days now but I can't get the VIO server to recognize the new LPAR. This is what I did: On HMC create new LPAR On HMC create new virtual scsi adapter for LPAR in VIO profile On HMC add new virtual scsi adapter for LPAR with DLPAR... (6 Replies)
Discussion started by: petervg
6 Replies

9. AIX

cdrom confusion on the vio client lpar

Hi In my vio server I have the below output $ lsvopt | grep -i SAPSITGS sapsitgs_cdrom TL12UP.iso 3182 In my vio client lpar I have the below output root@sapsitgs:/ # lsdev -Cc cdromcd0 Available Virtual SCSI Optical Served by VIO Server cd1... (1 Reply)
Discussion started by: newtoaixos
1 Replies

10. AIX

VIO and LPAR disk/fcs communications

:wall::wall::wall: 1. I have created an LPAR in the HMC. 2. I have allocated the storage from an Hitachi AMS2500 and assigned it to the host group. 3. I have zoned the LPAR and Storage on a Brocade 5100. (The zone sees the AMS) Next I activated the LPAR in the HMC, SMS mode for the mksysb... (3 Replies)
Discussion started by: Dallasguy7
3 Replies
stmsboot(1M)															      stmsboot(1M)

NAME
stmsboot - administration program for the Solaris I/O multipathing feature SYNOPSIS
/usr/sbin/stmsboot [-d | -e | -u | -L | -l controller_number] The Solaris I/O multipathing feature is a multipathing solution for storage devices that is part of the Solaris operating environment. This feature was formerly known as Sun StorEdge Traffic Manager (STMS) or MPxIO. The stmsboot program is an administrative command to manage enumeration of fibre channel devices under Solaris I/O multipathing. Solaris I/O multipathing-enabled devices are enumerated under scsi_vhci(7D), providing multipathing capabilities. Solaris I/O multipathing-disabled devices are enumerated under the physical controller. In the /dev and /devices trees, Solaris I/O multipathing-enabled devices receive new names that indicate that they are under Solaris I/O multipathing control. This means a device will have a different name from its original name (following installation) when it is under Solaris I/O multipathing control. The stmsboot command automatically updates /etc/vfstab and dump configuration to reflect the device names changes when enabling or disabling Solaris I/O multipathing. A reboot is required for changes to take effect. The following options are supported: -e Enables Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports. Following this enabling, you are prompted to reboot. During the reboot, vfstab and the dump configuration will be updated to reflect the device name changes. -d Disables Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports. Following this disabling, you are prompted to reboot. During the reboot, vfstab and the dump configuration will be updated to reflect the device name changes. -u Updates vfstab and the dump configuration after you have manually modified the configuration to have Solaris I/O multipathing enabled or disabled on specific fp(7D) controller ports. This option prompts you to reboot. During the reboot, vfstab and the dump configura- tion will be updated to reflect the device name changes. -L Display the device name changes from non-Solaris I/O multipathing device names to Solaris I/O multipathing device names. -l controller_number Display the device name changes from non-Solaris I/O multipathing device names to Solaris I/O multipathing device names for the speci- fied controller. Along with its primary function of enabling or disabling Solaris I/O multipathing, the stmsboot command is used to update vfstab and the dump configuration to reflect device name changes. For a system to function properly, you must configure the applications that consume the devices by old names to use the new names. The -L and -l options display the mapping between the old and new device names. These options work after the changes made to the Solaris I/O multipathing configuration have taken effect. For example, you can use these options following the reboot after invoking stmsboot -e. The old device names must exist in order to display the mappings. Example 1: Enabling Solaris I/O Multipathing Following OS Upgrade To enable Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports run: # stmsboot -e Example 2: Disabling Solaris I/O Multipathing To disable Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports, run: # stmsboot -d Example 3: Enabling Solaris I/O Multipathing on Selected Ports You want to enable Solaris I/O multipathing on some fibre channel controller ports and disable the feature on the rest. You edit the fp.conf file (see fp(7D)) to enable or disable Solaris I/O multipathing on specific controller ports. You then run the following command to have vfstab and the dump configuration updated to reflect the new device names: # stmsboot -u See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Architecture |SPARC | +-----------------------------+-----------------------------+ |Availability |SUNWcsu, SUNWcslr | +-----------------------------+-----------------------------+ |Interface Stability |Obsolete | +-----------------------------+-----------------------------+ SEE ALSO
dumpadm(1M), ufsdump(1M), dumpdates(4), vfstab(4), fcp(7D), fctl(7D), fp(7D), qlc(7D), scsi_vhci(7D) Consult the Sun StorEdge Disk Tray [or Subsystem] Administrator's Guide for the T3, 3910, 3960, 6120, and 6320 storage subsystems. Sun StorEdge Traffic Manager Installation and Configuration Guide Solaris I/O multipathing is not supported on all devices. After enabling Solaris I/O multipathing, only those devices that Solaris I/O mul- tipathing supports are placed under Solaris I/O multipathing control. Non-supported devices remain as before. For Solaris releases prior to the current release, the -e and -d options remove the mpxio-disable property entries from fp.conf file (see fp(7D)) and add a global mpxio-disable entry to fp.conf. The current release of the Solaris operating system does not support the mpxio-disable property. Solaris I/O multipathing is always enabled. If you want to disable multipathing, you must use the mechanisms provided by the HBA drivers. See fp(7D). Enabling Solaris I/O Multipathing on a Sun StorEdge Disk Array The following applies to Sun StoreEdge T3, 3910, 3960, 6120, and 6320 storage subsystems. To place your Sun StorEdge disk subsystem under Solaris I/O multipathing control, in addition to enabling Solaris I/O multipathing, the mp_support of the subsystem must be set to mpxio mode. The preferred sequence is to change the subsystem's mp_support to mpxio mode, then run stmsboot -e. If Solaris I/O multipathing is already enabled but the subsystem's mp_support is not in mpxio mode, then change the mp_support to mpxio mode and run stmsboot -u. Refer to the Sun StorEdge Administrator's Guide for your subsystem for more details. ufsdump Users The ufsdump command keeps records of the filesystem dumps in /etc/dumpdates (see dumpdates(4)). Among other items, the records contain device names. An effect of the "active" stmsboot options (-e, -d, and -u) is to change the device name of a storage device. The stmsboot command does not modify the dumpdates file. Because of this, the dumpdates records will refer to the old device names, that is, the device names that were in effect before you ran stmsboot. The effect of this device name-dumpdates disagreement is that, following use of stms- boot, ufsdump will be processed as if no previous dump had ever been made, thus dumping the entire filesystem (effectively, a level 0 dump). Procedure to Use stmsboot in Sun Cluster Environment If possible, use stmsboot -e before you start installing Sun Cluster software. After you run stmsboot, you install Sun Cluster software as you normally would. If you install Sun Cluster software before running stmsboot, you must use the following procedure. On each machine in the cluster on which you want to enable the Solaris multipathing feature, enter: # stmsboot -e ...and allow the system to reboot. When the system comes up, enter the following two commands: 1. # /usr/cluster/bin/scdidadm -C 2. # /usr/cluster/bin/scdidadm -r The preceding commands update did mappings with new device names while preserving did instance numbers for disks that are connected to multiple cluster nodes. did instance numbers of the local disks might not be preserved. For this reason, the did disk names for local disks might change. 3. Update /etc/vfstab to reflect any new did disk names for your local disks. 4. Reboot the system. To disable the Solaris multipathing feature, use stmsboot -d (instead of stmsboot -e), then follow the procedure above. To view mappings between the old and new device names, run stmsboot -L. To view did device name mappings, run /usr/cluster/bin/scdidadm -L. 3 Mar 2005 stmsboot(1M)
All times are GMT -4. The time now is 05:35 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy