Sponsored Content
Full Discussion: AIX vpath lun
Operating Systems AIX AIX vpath lun Post 302271014 by Stephan on Tuesday 23rd of December 2008 01:34:41 PM
Old 12-23-2008
You could also issue the command lsvpcfg under AIX with sdd installed...
 

10 More Discussions You Might Find Interesting

1. AIX

vpath to an hdisk

Is there a simply way for me to map a vpath to an hdisk on AIX 5.2? (5 Replies)
Discussion started by: 2dumb
5 Replies

2. AIX

vpath and sharks

Hello everyone: just want to know how to get the name of the shark in which the vpath is? thanks (2 Replies)
Discussion started by: jcpetela
2 Replies

3. AIX

AIX not detecting LUN

Hello All, Really hoping someone can help. We have an AIX server connected to a Quantum SDLC sitting in front of an S10K library. The library has 3 logical partitions (two LTO4, one LTO3), so it requires 3 distinct medium changers. The SDLC presents the 3 luns (which are the medium... (1 Reply)
Discussion started by: jwholey
1 Replies

4. Solaris

I have LUN ID, how to find disk relate to that LUN ID?

I have a list of LUN ID, my task is to find if disk has been added or not. How do I do that? I have been searching the forum and not able to find answer. thanks (4 Replies)
Discussion started by: uuontario
4 Replies

5. AIX

Mount a SAN LUN which contains clone copy - AIX 6.1

Hello Everyone, Can someone help me to mount a SAN hdisk which contains a clone data copy(san) of the remote server to the another machine. Both servers are running in AIX. Thanks in advance ! Regards, Gowtham.G (3 Replies)
Discussion started by: gowthamakanthan
3 Replies

6. AIX

Increase LUN size in AIX with VIOS and HACMP

Hello! I have this infraestructure: - 1 POWER7 with single VIOS on Site A. - 1 POWER6 with single VIOS on Site B. - 1 LPAR called NodeA as primary node for PowerHA 6.1 on Site A. - 1 LPAR called NodeB as secondary (cold) node for PowerHA 6.1 on SiteB. - 1 Storage DS4700 on Site A. - 1... (8 Replies)
Discussion started by: enzote
8 Replies

7. AIX

Add new LUN in AIX

Hello all! I have not so much experience with AIX We have test p5 server. On server was deployed VIOS + 3 LPAR + AIX6.1 There is a test HACMP cluster of two nodes, the cluster is identical to production cluster (all clusters and hosts have been configured by specialist of integrator company and... (3 Replies)
Discussion started by: Ravil Khalilov
3 Replies

8. AIX

AIX help -reconnect EMC lun

we have an old AIX system with important data on an EMC LUN. the AIX server crashed and we had to rebuild it with the old AIX os 5.8. how do we reconnect the EMC LUN without losing the data? (15 Replies)
Discussion started by: jhudson12
15 Replies

9. Post Here to Contact Site Administrators and Moderators

AIX server do not detect LUN from HP storage

ssssssssssss (0 Replies)
Discussion started by: sampathenjoy
0 Replies

10. AIX

Changing AIX LUN paths

Hello, Looking for some guidance to changing how an AIX 6.1 server sees its luns. Presently they're connected via 1GB iscsi but I want to change it to 4GB fiber channel. Anyone done this before and if so is it an easy step? I'm thinking this can be done in this order: unmount the... (5 Replies)
Discussion started by: bbbngowc
5 Replies
BIOCTL(8)						    BSD System Manager's Manual 						 BIOCTL(8)

NAME
bioctl -- RAID management interface SYNOPSIS
bioctl device command [arg [...]] DESCRIPTION
RAID device drivers which support management functionality can register their services with the bio(4) driver. bioctl then can be used to manage the RAID controller's properties. COMMANDS
The following commands are supported: show [disks | volumes] Without any argument by default bioctl will show information about all volumes and the logical disks used on them. If disks is specified, only information about physical disks will be shown. If volumes is specified, only information about the volumes will be shown. alarm [disable | enable | silence | test] Control the RAID card's alarm functionality, if supported. By default if no argument is specified, its current state will be shown. Optionally the disable, enable, silence, or test arguments may be specified to enable, disable, silence, or test the RAID card's alarm. blink start channel:target.lun | stop channel:target.lun Instruct the device at channel:target.lun to start or cease blinking, if there's ses(4) support in the enclosure. hotspare add channel:target.lun | remove channel:target.lun Create or remove a hot-spare drive at location channel:target.lun. passthru add DISKID channel:target.lun | remove channel:target.lun Create or remove a pass-through device. The DISKID argument specifies the disk that will be used for the new device, and it will be created at the location channel:target.lun. NOTE: Removing a pass-through device that has a mounted filesys- tem will lead to undefined behaviour. check start VOLID | stop VOLID Start or stop consistency volume check in the volume with index VOLID. NOTE: Not many RAID controllers support this fea- ture. create volume VOLID DISKIDs [SIZE] STRIPE RAID_LEVEL channel:target.lun Create a volume at index VOLID. The DISKIDs argument will specify the first and last disk, i.e.: 0-3 will use the disks 0, 1, 2, and 3. The SIZE argument is optional and may be specified if not all available disk space is wanted (also dependent of the RAID_LEVEL). The volume will have a stripe size defined in the STRIPE argument and it will be located at channel:target.lun. remove volume VOLID channel:target.lun Remove a volume at index VOLID and located at channel:target.lun. NOTE: Removing a RAID volume that has a mounted filesystem will lead to undefined behaviour. EXAMPLES
The following command, executed from the command line, shows the status of the volumes and its logical disks on the RAID controller: $ bioctl arcmsr0 show Volume Status Size Device/Label RAID Level Stripe ================================================================= 0 Building 468G sd0 ARC-1210-VOL#00 RAID 6 128KB 0% done 0:0 Online 234G 0:0.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:1 Online 234G 0:1.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:2 Online 234G 0:2.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:3 Online 234G 0:3.0 noencl <WDC WD2500YS-01SHB1 20.06C06> To create a RAID 5 volume on the SCSI 0:15.0 location on the disks 0, 1, 2, and 3, with stripe size of 64Kb on the first volume ID, using all available free space on the disks: $ bioctl arcmsr0 create volume 0 0-3 64 5 0:15.0 To remove the volume 0 previously created at the SCSI 0:15.0 location: $ bioctl arcmsr0 remove volume 0 0:15.0 SEE ALSO
arcmsr(4), bio(4), cac(4), ciss(4), mfi(4) HISTORY
The bioctl command first appeared in OpenBSD 3.8, it was rewritten for NetBSD 5.0. AUTHORS
The bioctl interface was written by Marco Peereboom <marco@openbsd.org> and was rewritten with multiple features by Juan Romero Pardines <xtraeme@NetBSD.org>. BSD
March 16, 2008 BSD
All times are GMT -4. The time now is 07:30 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy