Sponsored Content
Full Discussion: AIX, VIOS and DS4800 SAN
Operating Systems AIX AIX, VIOS and DS4800 SAN Post 302237150 by raba on Wednesday 17th of September 2008 05:45:17 AM
Old 09-17-2008
Hi,

you'll first have to zone the WWNs together on the san switches.
Look at this site and notice the mkvdev commands to map the disk to the vioclient POWER5 virtualization: How to set up the IBM Virtual I/O Server
 

10 More Discussions You Might Find Interesting

1. AIX

Problems AIX and SAN.

Sorry for my english. We have a IBM BLADES JS21. AIX 5.3 update to 6. Our JS21 has 2 FC (fcs0 and fcs1). We have one DS4072, one Disk system with 2 controllers and 2 FC by controllers. This means, all AIX FC see all Disk systems controllers by 2 FC switchs. (one fc two roads) FC AIX... (4 Replies)
Discussion started by: fjgonzalez
4 Replies

2. AIX

AIX disk less with SAN

Hi All, I have mirrored SAN volume on my B80 rootvg. Can I just remove the mirror and "Remove a P V from a V G" and it will be a diskless AIX? Is that going to boot on SAN rootvg volume? Thanks in advance, itik (3 Replies)
Discussion started by: itik
3 Replies

3. AIX

how to increase FS size on SAN DS4800 ?

Hello, I am using DS4800 SAN with AIX 5.3, and would like to increase the size of the DS4800 Luns, so that my FS increase also. How to do this please ? Modifying Lun risk to make all data of the FS being lost ! NO ? (8 Replies)
Discussion started by: astjen
8 Replies

4. AIX

AIX & DS4800 error

Hi people... i have a problem with a DS4800 san... it keeps me showing this message (on picture) and i dont know how to solve it! someone knows? thanks a lot (2 Replies)
Discussion started by: gfca
2 Replies

5. AIX

Increase LUN size in AIX with VIOS and HACMP

Hello! I have this infraestructure: - 1 POWER7 with single VIOS on Site A. - 1 POWER6 with single VIOS on Site B. - 1 LPAR called NodeA as primary node for PowerHA 6.1 on Site A. - 1 LPAR called NodeB as secondary (cold) node for PowerHA 6.1 on SiteB. - 1 Storage DS4700 on Site A. - 1... (8 Replies)
Discussion started by: enzote
8 Replies

6. AIX

VIOS IP address - separate vlan for vios servers ?

Hello, Lets say for simplicity that I do not use any vlan config inside my server - one lpar group use hea physical port1, another group hea physical port2. Physical port1 configured as vlan1 on external switch, physical port2 as vlan2. What is the common practice - should I isolate my vios... (0 Replies)
Discussion started by: vilius
0 Replies

7. AIX

AIX VIOS IVM Ethernet problem

Hi All, I have a p520 box running 3 LPARs, 1 x VIOS, 1 x AIX 6.1 and 1 x AIX 7.1 and no HMC, all configuration is performed via VIOS/IVM Until recently the box did not have the AIX7.1 LPAR and everything was fine with the box. We needed to use AIX 7.1 for a new project, so I created a... (11 Replies)
Discussion started by: KevB
11 Replies

8. AIX

Revisiting IBM PowerVM VIOS AIX

coming from this thread: https://www.unix.com/aix/211107-ibm-power-pseries-open-firmware-boot-vios-powervm-vet-code.html PowerVM, formerly known as Advanced Power Virtualization (APV), is a chargeable feature of IBM POWER5, POWER6, POWER7, POWER8, and POWER9 servers and is required for support... (0 Replies)
Discussion started by: filosophizer
0 Replies

9. UNIX for Advanced & Expert Users

Unable to install client AIX LPAR to vscsi hdisk provided from VIOS

Hi everybody, I have Power5 server with 4 internal hdisks each of 70Gb. VIOS server was installed via Virtual I/O Server Image Repository on the HMC. HMC release - 7.7.0 VIOS rootvg installed on 2 disk(these disks merged to one storage pool during VIOS install process),and 2 others hdisks... (2 Replies)
Discussion started by: Ravil Khalilov
2 Replies

10. AIX

IBM VIOS AIX HMC Error

hello P6 was managed by HMC, tried to install on it VIOS 2.2.6 and got this error message Hosting requires a hosting partition -- boot not permitted vios-4 - imgbb.com then after creating an lpar in HMC got this message B2001230 vios-3 - imgbb.com anyone know how to solve this... (3 Replies)
Discussion started by: filosophizer
3 Replies
raidtab(5)							File Formats Manual							raidtab(5)

NAME
raidtab - configuration file for md (RAID) devices DESCRIPTION
/etc/raidtab is the default configuration file for the raid tools (raidstart and company). It defines how RAID devices are configured on a system. FORMAT
/etc/raidtab has multiple sections, one for each md device which is being configured. Each section begins with the raiddev keyword. The order of items in the file is important. Later raiddev entries can use earlier ones (which allows RAID-10, for example), and the parsing code isn't overly bright, so be sure to follow the ordering in this man page for best results. Here's a sample md configuration file: # # sample raiddev configuration file # 'old' RAID0 array created with mdtools. # raiddev /dev/md0 raid-level 0 nr-raid-disks 2 persistent-superblock 0 chunk-size 8 device /dev/hda1 raid-disk 0 device /dev/hdb1 raid-disk 1 raiddev /dev/md1 raid-level 5 nr-raid-disks 3 nr-spare-disks 1 persistent-superblock 1 parity-algorithm left-symmetric device /dev/sda1 raid-disk 0 device /dev/sdb1 raid-disk 1 device /dev/sdc1 raid-disk 2 device /dev/sdd1 spare-disk 0 Here is more information on the directives which are in raid configuration files; the options are listen in this file in the same order they should appear in the actual configuration file. raiddev device This introduces the configuration section for the stated device. nr-raid-disks count Number of raid devices in the array; there should be count raid-disk entries later in the file. (current maximum limit for RAID devices -including spares- is 12 disks. This limit is already extended to 256 disks in experimental patches.) nr-spare-disks count Number of spare devices in the array; there should be count spare-disk entries later in the file. Spare disks may only be used with RAID4 and RAID5, and allow the kernel to automatically build new RAID disks as needed. It is also possible to add/remove spares run- time via raidhotadd/raidhotremove, care has to be taken that the /etc/raidtab configuration exactly follows the actual configuration of the array. (raidhotadd/raidhotremove does not change the configuration file) persistent-superblock 0/1 newly created RAID arrays should use a persistent superblock. A persistent superblock is a small disk area allocated at the end of each RAID device, this helps the kernel to safely detect RAID devices even if disks have been moved between SCSI controllers. It can be used for RAID0/LINEAR arrays too, to protect against accidental disk mixups. (the kernel will either correctly reorder disks, or will refuse to start up an array if something has happened to any member disk. Of course for the 'fail-safe' RAID variants (RAID1/RAID5) spares are activated if any disk fails.) Every member disk/partition/device has a superblock, which carries all information necessary to start up the whole array. (for autodetection to work all the 'member' RAID partitions should be marked type 0xfd via fdisk) The superblock is not visible in the final RAID array and cannot be destroyed accidentally through usage of the md device files, all RAID data content is available for filesystem use. parity-algorithm which The parity-algorithm to use with RAID5. It must be one of left-asymmetric, right-asymmetric, left-symmetric, or right-symmetric. left-symmetric is the one that offers maximum performance on typical disks with rotating platters. chunk-size size Sets the stripe size to size kilobytes. Has to be a power of 2 and has a compilation-time maximum of 4M. (MAX_CHUNK_SIZE in the ker- nel driver) typical values are anything from 4k to 128k, the best value should be determined by experimenting on a given array, alot depends on the SCSI and disk configuration. device devpath Adds the device devpath to the list of devices which comprise the raid system. Note that this command must be followed by one of raid-disk, spare-disk, or parity-disk. Also note that it's possible to recursively define RAID arrays, ie. to set up a RAID5 array of RAID5 arrays. (thus achieving two-disk failure protection, at the price of more disk space spent on RAID5 checksum blocks) raid-disk index The most recently defined device is inserted at position index in the raid array. spare-disk index The most recently defined device is inserted at position index in the spare disk array. parity-disk index The most recently defined device is moved to the end of the raid array, which forces it to be used for parity. failed-disk index The most recently defined device is inserted at position index in the raid array as a failed device. This allows you to create raid 1/4/5 devices in degraded mode - useful for installation. Don't use the smallest device in an array for this, put this after the raid-disk definitions! NOTES
The raidtools are derived from the md-tools and raidtools packages, which were originally written by Marc Zyngier, Miguel de Icaza, Gadi Oxman, Bradley Ward Allen, and Ingo Molnar. SEE ALSO
raidstart(8), raid0run(8), mkraid(8), raidstop(8) raidtab(5)
All times are GMT -4. The time now is 03:31 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy