Sponsored Content
Full Discussion: AIX hard disk failure
Operating Systems AIX AIX hard disk failure Post 303031751 by bakunin on Tuesday 5th of March 2019 09:19:30 AM
Old 03-05-2019
Quote:
Originally Posted by Phat
After some checks, I can see the hdisk5 becomes bigger with new size but the VG still in the old size.
It seems that the "physical" layer of the RAID is already reconfigured. Perhaps you have re-read the configuration too with the cfgmgr command and hence hdisk5 (this is the "logical" representation of the whole RAID) has become bigger. Anyways, you can make sure that the "new" hdisk5 is identified correctly in all its aspects.

Unmount the filesystems of the VG, then do a varyoffvg <VG> . Then delete the hdisk device and rediscover it:

Code:
rmdev -Rl hdisk5
cfgmgr

Now you need to tell the volume manager that the VG has changed. Issue a

Code:
chvg -g <volume-groupname>

which should do the trick. I am not sure if the VG needs to be varied on or off for that, so try in varyoffvg mode and if you get an error do a varyonvg <VG> and try again.

Ah, a last thing:

Quote:
We have pulled out the failure disk pdisk5 and added a new hard disk.
DON'T DO THAT!

In this case you were lucky, but generally - as i wrote above - it is a bad idea to remove disks which are still known to the system. Always deconfigure them first and pull them only then.

I hope this helps.

bakunin

Last edited by bakunin; 03-05-2019 at 10:30 AM..
 

10 More Discussions You Might Find Interesting

1. UNIX for Advanced & Expert Users

Disk failure

is there anu way by which i can find out if all the disks on the system are working ? Milind Shauche. (2 Replies)
Discussion started by: shauche
2 Replies

2. AIX

hard disk information in AIX

Hi, Other than df -k, is there any command that will tell me all physical hard drives installed on the system as well as the size of each one? I'm using AIX 5.1 Thanks, (3 Replies)
Discussion started by: quickfirststep
3 Replies

3. HP-UX

Disk Failure

I am new to being a Unix admin and have a question about replacing some hardware. I have a K class box using HP-UX 10.20 with three disks. Two of the drives are in one logical volume. Every 3 or 4 days, the syslog is showing that one of these drives is experiencing "POWERFAILED" and then recovering... (6 Replies)
Discussion started by: SemiOfCol
6 Replies

4. Filesystems, Disks and Memory

Looking for a solution to disk failure!

Hi people, I have been using my disk for quite a long time but the other day I heard the drive making some noise and had to restart the system again. But when I did that the disk would not boot and I fear that the data might be deleted or lost. So, if you people have any know about the ways to get... (1 Reply)
Discussion started by: adam466
1 Replies

5. Filesystems, Disks and Memory

Looking for a solution to disk failure!

Hi people, I have been using my disk for quite a long time but the other day I heard the drive making some noise and had to restart the system again. But when I did that the disk would not boot and I fear that the data might be deleted or lost. So, if you people have any know about the ways to... (2 Replies)
Discussion started by: christopher4
2 Replies

6. Solaris

SAN disk failure

hi all, have a solaris 9 OS and a SAN disk which used to work fine is not getting picked up by my machine. can anyone point out things to check in order to troubleshoot this ?? thanks in advance. (3 Replies)
Discussion started by: cesarNZ
3 Replies

7. SCO

declare disk driver for IDE hard disk

hi I've a fresh installation of SCO 5.0.7 on the IDE hard disk. For SCSI hard disk I can declare, for example blc disk driver using: # mkdev hd 0 SCSI-0 0 blc 0but it works for IDE hard disk? (3 Replies)
Discussion started by: ccc
3 Replies

8. Red Hat

How to monitor HP server hard disk failure ?

in red hat 4, 5 any one know any commands or any scritps to monitor HP DL 380 G5/6 server and trigger alarm when hard disk failed. thanks for all support ---------- Post updated at 02:45 PM ---------- Previous update was at 12:00 PM ---------- does HP ProLiant Support Pack support is... (4 Replies)
Discussion started by: maxlee24
4 Replies

9. Linux

C++ Code to Access Linux Hard Disk Sectors (with a LoopBack Virtual Hard Disk)

Hi all, I'm kind of new to programming in Linux & c/c++. I'm currently writing a FileManager using Ubuntu Linux(10.10) for Learning Purposes. I've got started on this project by creating a loopback device to be used as my virtual hard disk. After creating the loop back hard disk and mounting it... (23 Replies)
Discussion started by: shen747
23 Replies

10. AIX

Clone 1 Hard disk fromIBM Intellipoint server with AIX 5.x

Hello to all, Im having a new task in a new world (AIX - IBM Servers) I have an IBM Server (Type - 9111-285 very old one) with one Hard disk (73 GB 10 K) with AIX 5.x, and I need to clone the existing disk to another with the same specifications. Could you please give me some advice in order... (7 Replies)
Discussion started by: trevian3969
7 Replies
BIOCTL(8)						    BSD System Manager's Manual 						 BIOCTL(8)

NAME
bioctl -- RAID management interface SYNOPSIS
bioctl device command [arg [...]] DESCRIPTION
RAID device drivers which support management functionality can register their services with the bio(4) driver. bioctl then can be used to manage the RAID controller's properties. COMMANDS
The following commands are supported: show [disks | volumes] Without any argument by default bioctl will show information about all volumes and the logical disks used on them. If disks is specified, only information about physical disks will be shown. If volumes is specified, only information about the volumes will be shown. alarm [disable | enable | silence | test] Control the RAID card's alarm functionality, if supported. By default if no argument is specified, its current state will be shown. Optionally the disable, enable, silence, or test arguments may be specified to enable, disable, silence, or test the RAID card's alarm. blink start channel:target.lun | stop channel:target.lun Instruct the device at channel:target.lun to start or cease blinking, if there's ses(4) support in the enclosure. hotspare add channel:target.lun | remove channel:target.lun Create or remove a hot-spare drive at location channel:target.lun. passthru add DISKID channel:target.lun | remove channel:target.lun Create or remove a pass-through device. The DISKID argument specifies the disk that will be used for the new device, and it will be created at the location channel:target.lun. NOTE: Removing a pass-through device that has a mounted filesys- tem will lead to undefined behaviour. check start VOLID | stop VOLID Start or stop consistency volume check in the volume with index VOLID. NOTE: Not many RAID controllers support this fea- ture. create volume VOLID DISKIDs [SIZE] STRIPE RAID_LEVEL channel:target.lun Create a volume at index VOLID. The DISKIDs argument will specify the first and last disk, i.e.: 0-3 will use the disks 0, 1, 2, and 3. The SIZE argument is optional and may be specified if not all available disk space is wanted (also dependent of the RAID_LEVEL). The volume will have a stripe size defined in the STRIPE argument and it will be located at channel:target.lun. remove volume VOLID channel:target.lun Remove a volume at index VOLID and located at channel:target.lun. NOTE: Removing a RAID volume that has a mounted filesystem will lead to undefined behaviour. EXAMPLES
The following command, executed from the command line, shows the status of the volumes and its logical disks on the RAID controller: $ bioctl arcmsr0 show Volume Status Size Device/Label RAID Level Stripe ================================================================= 0 Building 468G sd0 ARC-1210-VOL#00 RAID 6 128KB 0% done 0:0 Online 234G 0:0.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:1 Online 234G 0:1.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:2 Online 234G 0:2.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:3 Online 234G 0:3.0 noencl <WDC WD2500YS-01SHB1 20.06C06> To create a RAID 5 volume on the SCSI 0:15.0 location on the disks 0, 1, 2, and 3, with stripe size of 64Kb on the first volume ID, using all available free space on the disks: $ bioctl arcmsr0 create volume 0 0-3 64 5 0:15.0 To remove the volume 0 previously created at the SCSI 0:15.0 location: $ bioctl arcmsr0 remove volume 0 0:15.0 SEE ALSO
arcmsr(4), bio(4), cac(4), ciss(4), mfi(4) HISTORY
The bioctl command first appeared in OpenBSD 3.8, it was rewritten for NetBSD 5.0. AUTHORS
The bioctl interface was written by Marco Peereboom <marco@openbsd.org> and was rewritten with multiple features by Juan Romero Pardines <xtraeme@NetBSD.org>. BSD
March 16, 2008 BSD
All times are GMT -4. The time now is 05:36 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy