Sponsored Content
Operating Systems AIX IBM SAN STORAGE HOT SPARE DISK Post 302604167 by filosophizer on Saturday 3rd of March 2012 12:10:25 PM
Old 03-03-2012
[SOLVED] IBM SAN STORAGE HOT SPARE DISK

Hello,

I have DS4000 IBM SAN Storage ( aka FastT Storage )

One of my disks has failed and I had a hot spare disk covering all the arrays. As the disk failed, immediately the hotspare disk took over the failed disk ( see the JPEG in the attachment )

My Question: How can I make the hotspare which now took over the failed disk, to become part of the array and it should be a permanent disk ( instead of a hotspare disk as you can see + sign below the disk ) ?


thanks
IBM SAN STORAGE HOT SPARE DISK-hotsparejpg

Last edited by filosophizer; 03-14-2012 at 08:44 AM..
 

9 More Discussions You Might Find Interesting

1. AIX

IBM FastT600 SAN - RAID 5 Storage Manager Client v08.33.G5.03 - Recovery?

To summarize the problem: The "IBM FastT Storage Manager Client v8" shows that our Disk Farm is arranged into 6 logical drives each in a RAID 5 configuration. This software also shows that 5 of the 6 logical drives (from Disk Farm) are in a error state: "Failed Logical Drive - Drive Failure".... (1 Reply)
Discussion started by: aix-olympics
1 Replies

2. AIX

IBM SAN Storage DS4700 - Copy Mirror Flashcopy

Hello, Does anyone know how to copy SAN Storage logical disks from IBM TotalStorage Software. I have a SAN Logical Disk of 200GB mounted on my AIX LPAR_1 via fibre channel fcs0 I would like to make an exact copy of the SAN Logical Disk from IBM totalStorage and mount it on AIX LPAR_2 ... (4 Replies)
Discussion started by: filosophizer
4 Replies

3. Solaris

Hot Spare pool

One more query in SVM :) Now with hot spare spool... I can understand adding/replacing a slice in particular hot spare pool with "-a / -r" option (or) adding a slice to all existing hot spare pool with "-all" option. Here my query is for deleting, we have only option "-d". 1) If the hot... (2 Replies)
Discussion started by: gowthamakanthan
2 Replies

4. AIX

MPIO RDAC IBM SAN STORAGE DS4700 ?

Hello, I have AIX 6.1 with TL 4 and it is connected to IBM SAN STORAGE DS4700 After assigning some disks from SAN to AIX, I can see the disks in my AIX as hdisk2 Available 05-00-02 MPIO Other DS4K Array Disk hdisk3 Available 05-00-02 MPIO Other DS4K Array Disk But it should... (0 Replies)
Discussion started by: filosophizer
0 Replies

5. Solaris

Cannot see the IBM SAN storage

HI all, I had recently change the Server storage from EMC to the IBM SAN. but after the configuration, the IBM success to see the server HBA port and successfully assign a LUN for the server. When i go to the server, and restarted it. i use the "format" command to check, but din see any... (1 Reply)
Discussion started by: SmartAntz
1 Replies

6. Solaris

Hot Spare replacement

Hi Guys, Can Someone pls let me know the thorough process for Hot spare replacement as current Hot spare slice has broken down . :mad: Thanks ---------- Post updated at 06:34 PM ---------- Previous update was at 05:21 PM ---------- Update : Its a solaris 10 box (1 Reply)
Discussion started by: Solarister
1 Replies

7. AIX

IBM SAN storage -- cache battery

Hello, I have IBM SAN STORAGE DS4100 and one of the cache battery for the controller is dead. Suddenly the performance has been degraded and access to SAN disks ( reading and writing ) became very slow ? My query: Replacing the battery will take 6 days, so in the mean time what are the ways... (1 Reply)
Discussion started by: filosophizer
1 Replies

8. AIX

IBM SAN TO SAN Mirroring

Has anyone tried SAN to SAN mirroring on IBM DS SAN Storage. DS5020 mentions Enhanced Remote Mirror to multi-LUN applications I wonder if Oracle High availibility can be setup using Remote Mirror option of SAN ? (1 Reply)
Discussion started by: filosophizer
1 Replies

9. AIX

IBM AIX - SAN Storage DS4300 issue

Hi, This is follow up to the post https://www.unix.com/aix/233361-san-disk-appearing-double-aix.html When I connected Pseries Machine HBA Card ( Dual Port ) directly to the SAN Storage DS4300 , I was able to see Host Port Adapter WWN numbers , although I was getting this message... (2 Replies)
Discussion started by: filosophizer
2 Replies
raidtab(5)							File Formats Manual							raidtab(5)

NAME
raidtab - configuration file for md (RAID) devices DESCRIPTION
/etc/raidtab is the default configuration file for the raid tools (raidstart and company). It defines how RAID devices are configured on a system. FORMAT
/etc/raidtab has multiple sections, one for each md device which is being configured. Each section begins with the raiddev keyword. The order of items in the file is important. Later raiddev entries can use earlier ones (which allows RAID-10, for example), and the parsing code isn't overly bright, so be sure to follow the ordering in this man page for best results. Here's a sample md configuration file: # # sample raiddev configuration file # 'old' RAID0 array created with mdtools. # raiddev /dev/md0 raid-level 0 nr-raid-disks 2 persistent-superblock 0 chunk-size 8 device /dev/hda1 raid-disk 0 device /dev/hdb1 raid-disk 1 raiddev /dev/md1 raid-level 5 nr-raid-disks 3 nr-spare-disks 1 persistent-superblock 1 parity-algorithm left-symmetric device /dev/sda1 raid-disk 0 device /dev/sdb1 raid-disk 1 device /dev/sdc1 raid-disk 2 device /dev/sdd1 spare-disk 0 Here is more information on the directives which are in raid configuration files; the options are listen in this file in the same order they should appear in the actual configuration file. raiddev device This introduces the configuration section for the stated device. nr-raid-disks count Number of raid devices in the array; there should be count raid-disk entries later in the file. (current maximum limit for RAID devices -including spares- is 12 disks. This limit is already extended to 256 disks in experimental patches.) nr-spare-disks count Number of spare devices in the array; there should be count spare-disk entries later in the file. Spare disks may only be used with RAID4 and RAID5, and allow the kernel to automatically build new RAID disks as needed. It is also possible to add/remove spares run- time via raidhotadd/raidhotremove, care has to be taken that the /etc/raidtab configuration exactly follows the actual configuration of the array. (raidhotadd/raidhotremove does not change the configuration file) persistent-superblock 0/1 newly created RAID arrays should use a persistent superblock. A persistent superblock is a small disk area allocated at the end of each RAID device, this helps the kernel to safely detect RAID devices even if disks have been moved between SCSI controllers. It can be used for RAID0/LINEAR arrays too, to protect against accidental disk mixups. (the kernel will either correctly reorder disks, or will refuse to start up an array if something has happened to any member disk. Of course for the 'fail-safe' RAID variants (RAID1/RAID5) spares are activated if any disk fails.) Every member disk/partition/device has a superblock, which carries all information necessary to start up the whole array. (for autodetection to work all the 'member' RAID partitions should be marked type 0xfd via fdisk) The superblock is not visible in the final RAID array and cannot be destroyed accidentally through usage of the md device files, all RAID data content is available for filesystem use. parity-algorithm which The parity-algorithm to use with RAID5. It must be one of left-asymmetric, right-asymmetric, left-symmetric, or right-symmetric. left-symmetric is the one that offers maximum performance on typical disks with rotating platters. chunk-size size Sets the stripe size to size kilobytes. Has to be a power of 2 and has a compilation-time maximum of 4M. (MAX_CHUNK_SIZE in the ker- nel driver) typical values are anything from 4k to 128k, the best value should be determined by experimenting on a given array, alot depends on the SCSI and disk configuration. device devpath Adds the device devpath to the list of devices which comprise the raid system. Note that this command must be followed by one of raid-disk, spare-disk, or parity-disk. Also note that it's possible to recursively define RAID arrays, ie. to set up a RAID5 array of RAID5 arrays. (thus achieving two-disk failure protection, at the price of more disk space spent on RAID5 checksum blocks) raid-disk index The most recently defined device is inserted at position index in the raid array. spare-disk index The most recently defined device is inserted at position index in the spare disk array. parity-disk index The most recently defined device is moved to the end of the raid array, which forces it to be used for parity. failed-disk index The most recently defined device is inserted at position index in the raid array as a failed device. This allows you to create raid 1/4/5 devices in degraded mode - useful for installation. Don't use the smallest device in an array for this, put this after the raid-disk definitions! NOTES
The raidtools are derived from the md-tools and raidtools packages, which were originally written by Marc Zyngier, Miguel de Icaza, Gadi Oxman, Bradley Ward Allen, and Ingo Molnar. SEE ALSO
raidstart(8), raid0run(8), mkraid(8), raidstop(8) raidtab(5)
All times are GMT -4. The time now is 08:29 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy