Sponsored Content
Full Discussion: Upgrade disk in RAID 1
Operating Systems Solaris Upgrade disk in RAID 1 Post 302213064 by networkfre@k on Wednesday 9th of July 2008 05:36:39 AM
Old 07-09-2008
Hi,

Is the replacement of the two 72GB drives w/ two 146GB drives just
because you can get the drives quicker for replacement?

Because you dont use the additional added space from the step to 146GB drives, I can write a little how-to of doing that if you want..

In short :

-) Break the 2x72GB mirror by removing the second disk (c0t1d0)
-) Remove the drive and insert a 146GB drive
-) Copy (for now) the VTOC from disk1 (72GB) to disk2 (146GB) OR
create the needed existing partitions by hand and create a big (or two/three..) slice/partition to hold the additional ~65GB space
-) Create metadb replicas on the 146GB disk
-) Start mirror sync
-) When mirrored, break the mirror and remove the first (72GB) disk, V240 now runs from the 146GB disk in slot 2
-) Insert the other 146GB disk and copy the VTOC from disk2 like stated from jörg via prtvtoc , I always add the additional '-sh' arg to prtvtoc
prtvtoc -sh | fmthard -s -
-sh removes the human readable index and stuff , MAN page states that those arguments should be used when duplicating the VTOC from disk to disk
-) Create metaDBs on disk1 and mirror the disk
-) Create a filesystem on the new slice to use the additional space (of course on the metadevice)
-) D O N E

greets,
Christopher
 

10 More Discussions You Might Find Interesting

1. Filesystems, Disks and Memory

Creating a Mirror RAID With Existing Disk

Hi there, I'm not sure if this is possible, but here is what I'd like to do.. I have an existing 160GB drive in my Redhat 9.0 server that I would like to add an additional 160GB drive to and create a mirrored RAID of the first disk to the new disk. I would like to do this without having to... (2 Replies)
Discussion started by: sysera
2 Replies

2. UNIX for Advanced & Expert Users

disk upgrade

Hi all, this s going to be my 1st time upgrading a disk.I need the ways or steps to migrate data from a 2gb disk to a 4gb disk which sits in the High Availability Disk Array Model 10.the os disk is the internal disk of the K200 system so i dont think i would be meddling around with that.what r... (1 Reply)
Discussion started by: giriplug
1 Replies

3. Filesystems, Disks and Memory

Problem replace disk with RAID-5 volumes

Good morning, I have a problem replacing a disk with raid-5 volumes. An hardware error was occurred from a disk c9t3 so all slices were in maintenace. Every slice is part of a raid-5 volume. Any replica is present. Following Volume manager manual for replacing a disk, I have: - phisically... (0 Replies)
Discussion started by: bonovox
0 Replies

4. Solaris

Configuring RAID using single disk

Hi All, I have a SUN ENTERPRISE 3500 server with solaris 10 on it. I have already mirrored root partition.Now i need to mirror two more partitions with 25GB space. But i have only one disk having 70GB space.Total i have 7 disks but each one is of 18 GB only except one. Please find the output of... (2 Replies)
Discussion started by: Renjesh
2 Replies

5. Ubuntu

Ubunutu 8.04.4 RAID 1 mirror replace disk

Hi, I have an Ubuntu system which I have an faulted mirror. I trying to replace the disk, but I'm stuck on that it boots and only showing GRUB GRUB ## ## End Default Options ## title Ubuntu 8.04.4 LTS, kernel 2.6.24-26-server root (hd0,0) kernel ... (0 Replies)
Discussion started by: jld
0 Replies

6. UNIX for Advanced & Expert Users

Identify failed disk in Linux RAID

Good Evening, 2 years ago, I set up an Ubuntu file-server for a friend, who is a photograph amateur. Basically, the server offers a software RAID-5 that can be accessed remotely from a MAC. Unfortunately, I didn't labeled the hard drives (i.e. which physical drive corresponds to the /dev/sdX... (2 Replies)
Discussion started by: Loic Domaigne
2 Replies

7. Solaris

Hardware RAID not recognize the new disk [Sun T6320]

We have hardware RAID configured on Sun-Blade-T6320 and one of the disk got failed. Hence we replaced the failed disk. But the hot swapped disk not recognized by RAID. Kindly help on fixing this issue. We have 2 LDOM configured on this server and this server running on single disk. #... (8 Replies)
Discussion started by: rock123
8 Replies

8. AIX

SAN DISK raid level in AIX

Hello All, Our servers having emc clarion for the data disks. Is that possible to see those san disks raid level from AIX ? Am having AIX 6.1 and EMC clarion 5.5. Regards, Gowtham.G (3 Replies)
Discussion started by: gowthamakanthan
3 Replies

9. SuSE

/proc/mdstat and cciss not available, how to know your disk is raid

Im issuing a cat /proc/mdstat, dmraid -r, and finding a cciss, to know if my server is software raid and hardware raid. But all of them are missing. What is the other way to know, your disk are raid, your disks is sync, your disk are out of sync, your disk is failed, ASIDE LOOKING AT THEM... (0 Replies)
Discussion started by: invinzin21
0 Replies

10. Solaris

Patching on Raid 0 Disk

Dear All , We need to do patching on one Solaris Server , where we have raid 0 configured. What is the process to patch a Server if RAID 0 (Concat/Stripe) is there. Below is the sample output. # metadb flags first blk block count a m pc luo 16 ... (1 Reply)
Discussion started by: jegaraman
1 Replies
vxsparecheck(1M)														  vxsparecheck(1M)

NAME
vxsparecheck - monitor Veritas Volume Manager for failure events and replace failed disks SYNOPSIS
/etc/vx/bin/vxsparecheck [mail-address...] DESCRIPTION
The vxsparecheck command monitors Veritas Volume Manager (VxVM) by analyzing the output of the vxnotify command, waiting for failures to occur. It then sends mail via mailx to the logins specified on the command line, or (by default) to root. It then replaces any failed disks. After an attempt at replacement is complete, mail will be sent indicating the status of each disk replacement. The mail notification that is sent when a failure is detected follows this format: Failures have been detected by the Veritas Volume Manager: failed disks: medianame ... failed plexes: plexname ... failed subdisks: subdiskname ... failed volumes: volumename ... The Volume Manager will attempt to find hot-spare disks to replace any failed disks and attempt to reconstruct any data in volumes that have storage on the failed disk. The medianame list specifies disks that appear to have completely failed. The plexname list show plexes of mirrored volumes that have been detached due to I/O failures experienced while attempting to do I/O to subdisks they contain. The subdiskname list specifies subdisks in RAID-5 volumes that have been detached due to I/O errors. The volumename list shows non-RAID-5 volumes that have become unusable because disks in all of their plexes have failed (and are listed in the ``failed disks'' list) and shows those RAID-5 volumes that have become unusable because of multiple failures. If any volumes appear to have failed, the following paragraph will be included in the mail: The data in the failed volumes listed above is no longer available. It will need to be restored from backup. Replacement Procedure After mail has been sent, vxsparecheck finds a hot spare replacement for any disks that appear to have failed (that is, those listed in the medianame list). This involves finding an appropriate replacement for those eligible hot spares in the same disk group as the failed disk. A disk is eligible as a replacement if it is a valid Veritas Volume Manager disk (VM disk), has been marked as a hot-spare disk and con- tains enough space to hold the data contained in all the subdisks on the failed disk. To determine which disk from among the eligible hot spares to use, vxsparecheck first checks the file /etc/vx/sparelist (see Sparelist File below). If this file does not exist or lists no eligible hot spares for the failed disk, the disk that is ``closest'' to the failed disk is chosen. The value of ``closeness'' depends on the controller, target and disk number of the failed disk. A disk on the same controller as the failed disk is closer than a disk on a different controller; and a disk under the same target as the failed disk is closer than one under a different target. If no hot spare disk can be found, the following mail is sent: No hot spare could be found for disk medianame in diskgroup. No replacement has been made and the disk is still unusable. The mail then explains the disposition of volumes that had storage on the failed disk. The following message lists disks that had storage on the failed disk, but are still usable: The following volumes have storage on medianame: volumename These volumes are still usable, but the redundancy of those volumes is reduced. Any RAID-5 volumes with storage on the failed disk may become unusable in the face of further failures. If any non-RAID-5 volumes were made unusable due to the failure of the disk, the following message is included: The following volumes: volumename have data on medianame but have no other usable mirrors on other disks. These volumes are now unusable and the data on them is unavailable. If any RAID-5 volumes were made unavailable due to the disk failure, the following message is included The following RAID-5 volumes: volumename had storage on medianame and have experienced other failures. These RAID-5 volumes are now unusable and data on them is unavailable. If a hot-spare disk was found, a hot-spare replacement is attempted. This involves associating the device marked as a hot spare with the media record that was associated with the failed disk. If this is successful, the vxrecover(1M) command is used in the background to recover the contents of any data in volumes that had storage on the disk. If the hot-spare replacement fails, the following message is sent: Replacement of disk medianame in group diskgroup failed. The error is: error message If any volumes (RAID-5 or otherwise) are rendered unusable due to the failure, the following message is included: The following volumes: volumename occupy space on the failed disk and have no other available mirrors or have experienced other failures. These volumes are unusable, and the data they contain is unavailable. If the hot-spare replacement procedure completed successfully and recovery is under way, a final mail message is sent: Replacement of disk medianame in group diskgroup with disk device sparedevice has successfully completed and recovery is under way. If any non-RAID-5 volumes were rendered unusable by the failure despite the successful hot-spare procedure, the following message is included in the mail: The following volumes: volumename occupy spare on the replaced disk, but have no other enabled mirrors on other disks from which to perform recovery. These volumes must have their data restored. If any RAID-5 volumes were rendered unusable by the failure despite the successful hot-spare procedure, the following message is included in the mail: The following RAID-5 volumes: volumename have subdisks on the replaced disk and have experienced other failures that prevent recovery. These RAID-5 volumes must have their data restored. If any volumes (RAID-5 or otherwise) were rendered unusable, the following message is also included: To restore the contents of any volumes listed above, the volume should be started with the command: vxvol -f start volumename and the data restored from backup. Sparelist File The sparelist file is a text file that specifies an ordered list of disks to be used as hot spares when a specific disk fails. The system- wide sparelist file is located in /etc/vx/sparelist. Each line in the sparelist file specifies a list of spares for one disk. Lines beginning with the pound (#) character and empty lines are ignored. The format for a line in the sparelist file is: [ diskgroup:] diskname : spare1 [ spare2 ... ] The diskgroup field, if present, specifies the disk group within which the disk and designated spares reside. If this field is not speci- fied, the default disk group is determined using the rules given in the vxdg(1M) manual page. The diskname specifies the disk for which spares are being designated. The spare list after the colon lists the disks to be used as hot spares. The list is order dependent; in case of failure of diskname, the spares are tried in order. A spare will be used only if it is a valid hot spare (see above). If the list is exhausted without finding any spares, the default policy of using the closest disk is used. FILES
/etc/vx/sparelist Specifies a list of disks to serve as hot spares for a disk. NOTES
The sparelist file is not checked in any way for correctness until a disk failure occurs. It is possible to inadvertently specify a non- existent disk or inappropriate disk or disk group. Malformed lines are also ignored. SEE ALSO
mailx(1), vxintro(1M), vxnotify(1M), vxrecover(1M), vxrelocd(1M), vxunreloc(1M) VxVM 5.0.31.1 24 Mar 2008 vxsparecheck(1M)
All times are GMT -4. The time now is 10:21 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy