Sponsored Content
Operating Systems Linux Raid0 recovery from external HD Post 302386478 by dangral on Tuesday 12th of January 2010 01:46:19 PM
Old 01-12-2010
Is there any way to extrapolate the RAID0 information?
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Password recovery

We recently terminated a developer at my place of employment who created scripts on a windows server (that i do not have access to) that invoke FTP sessions on my UnixWare 7.1.1 servers. I need to know the password that is being used. Does anyone know of a good password crack? (8 Replies)
Discussion started by: rm -r *
8 Replies

2. Solaris

Solaris RAID0 doubt...

friends, Suppose I am typing metastat command and it is showing: d100: Concat/Stripe Size: 369495 blocks (180 MB) Stripe 0: (interlace: 32 blocks) Device Start Block Dbase Reloc c1d0s0 16065 Yes Yes c1d0s1 0 No Yes... (4 Replies)
Discussion started by: saagar
4 Replies

3. Solaris

Solaris recovery

Some thing happened to our solaris 10 ( sparc ) box and it is not coming up now. These are some of the console messages : I assume it is not able to find very basic system libraries so i need to tell it some how to find it under /lib:/usr/lib. I booted it from the CD but now i... (4 Replies)
Discussion started by: rajwinder
4 Replies

4. UNIX for Dummies Questions & Answers

Why is RAID0 faster?

I have read anecdotes about people installing RAID0 (RAID - Wikipedia, the free encyclopedia) on some of their machines because it gives a performance boost. Because bandwidth on the motherboard is limited, can someone explain exactly why it should be faster? (7 Replies)
Discussion started by: figaro
7 Replies

5. UNIX for Advanced & Expert Users

live upgrade with raid0 soft partitions

Hi, I have this mirrored system with soft-partitions. I have a difficulty determining the lucreate cmd in this env. #metastat -p d0 -m d10 d20 1 d10 1 1 c1t2d0s0 d20 1 1 c1t3d0s0 d1 -m d11 d21 1 d11 1 1 c1t2d0s5 d21 1 1 c1t3d0s5 d100 -p d1 -o 58720384 -b 8388608 d200 -p d1 -o... (1 Reply)
Discussion started by: chaandana
1 Replies

6. Hardware

HP9000 Server - Stuck on RAID0

Hey all, I've got an old HP9000 L1000 server with HP-UX installed. The drives that the OS is running on are in RAID0. I am concerned for the reliability of the server. The four hard drives in the front of the server are LVD 18.2 drives. I know with RAID0, if one drive fails, they all fail. ... (2 Replies)
Discussion started by: mroselli
2 Replies

7. Solaris

Cloning RAID0 drives, Solaris 10u11

Hello all, this is my first time posting here. Where I work we have multiple servers (x3-2's) running Solaris 10u11 with 2 drives configured as RAID0, 300GB per. There are 4-6 open slots for drives to clone to. Past attempts to clone/backup these drives has failed. One of the machines is... (1 Reply)
Discussion started by: eprlsguy
1 Replies

8. Gentoo

Data recovery of formatted external HDD

accidentally formatted ext3 external hard disk .. im using EAse us tool in windows system to recover the data ... will this works?? if yes ... the another external hard disk have to be formatted in which file system ? is there any other option ..please help me out (1 Reply)
Discussion started by: rajeshz
1 Replies

9. Solaris

Solaris 11 recovery

Hi, I need to recover the Solaris 11 OS, and it backup via Netbackup 7.6 file level backup only. Does anyone know what are steps to recover it? Thanks. :confused::confused::confused: (3 Replies)
Discussion started by: freshmeat
3 Replies

10. UNIX for Dummies Questions & Answers

Raid0 array stresses only 1 disk out of 3

Hi there, I've setup a raid0 array of 3 identical disks using : mdadm --create --verbose /dev/md0 --level=stripe --raid-devices=3 /dev/sdb1 /dev/sdc1 /dev/sdd1I'm using dstat to monitor the disk activity : dstat --epoch -D sdb,sdc,sdd --disk-util 30The results show that the stress is not... (8 Replies)
Discussion started by: chebarbudo
8 Replies
raidtab(5)							File Formats Manual							raidtab(5)

NAME
raidtab - configuration file for md (RAID) devices DESCRIPTION
/etc/raidtab is the default configuration file for the raid tools (raidstart and company). It defines how RAID devices are configured on a system. FORMAT
/etc/raidtab has multiple sections, one for each md device which is being configured. Each section begins with the raiddev keyword. The order of items in the file is important. Later raiddev entries can use earlier ones (which allows RAID-10, for example), and the parsing code isn't overly bright, so be sure to follow the ordering in this man page for best results. Here's a sample md configuration file: # # sample raiddev configuration file # 'old' RAID0 array created with mdtools. # raiddev /dev/md0 raid-level 0 nr-raid-disks 2 persistent-superblock 0 chunk-size 8 device /dev/hda1 raid-disk 0 device /dev/hdb1 raid-disk 1 raiddev /dev/md1 raid-level 5 nr-raid-disks 3 nr-spare-disks 1 persistent-superblock 1 parity-algorithm left-symmetric device /dev/sda1 raid-disk 0 device /dev/sdb1 raid-disk 1 device /dev/sdc1 raid-disk 2 device /dev/sdd1 spare-disk 0 Here is more information on the directives which are in raid configuration files; the options are listen in this file in the same order they should appear in the actual configuration file. raiddev device This introduces the configuration section for the stated device. nr-raid-disks count Number of raid devices in the array; there should be count raid-disk entries later in the file. (current maximum limit for RAID devices -including spares- is 12 disks. This limit is already extended to 256 disks in experimental patches.) nr-spare-disks count Number of spare devices in the array; there should be count spare-disk entries later in the file. Spare disks may only be used with RAID4 and RAID5, and allow the kernel to automatically build new RAID disks as needed. It is also possible to add/remove spares run- time via raidhotadd/raidhotremove, care has to be taken that the /etc/raidtab configuration exactly follows the actual configuration of the array. (raidhotadd/raidhotremove does not change the configuration file) persistent-superblock 0/1 newly created RAID arrays should use a persistent superblock. A persistent superblock is a small disk area allocated at the end of each RAID device, this helps the kernel to safely detect RAID devices even if disks have been moved between SCSI controllers. It can be used for RAID0/LINEAR arrays too, to protect against accidental disk mixups. (the kernel will either correctly reorder disks, or will refuse to start up an array if something has happened to any member disk. Of course for the 'fail-safe' RAID variants (RAID1/RAID5) spares are activated if any disk fails.) Every member disk/partition/device has a superblock, which carries all information necessary to start up the whole array. (for autodetection to work all the 'member' RAID partitions should be marked type 0xfd via fdisk) The superblock is not visible in the final RAID array and cannot be destroyed accidentally through usage of the md device files, all RAID data content is available for filesystem use. parity-algorithm which The parity-algorithm to use with RAID5. It must be one of left-asymmetric, right-asymmetric, left-symmetric, or right-symmetric. left-symmetric is the one that offers maximum performance on typical disks with rotating platters. chunk-size size Sets the stripe size to size kilobytes. Has to be a power of 2 and has a compilation-time maximum of 4M. (MAX_CHUNK_SIZE in the ker- nel driver) typical values are anything from 4k to 128k, the best value should be determined by experimenting on a given array, alot depends on the SCSI and disk configuration. device devpath Adds the device devpath to the list of devices which comprise the raid system. Note that this command must be followed by one of raid-disk, spare-disk, or parity-disk. Also note that it's possible to recursively define RAID arrays, ie. to set up a RAID5 array of RAID5 arrays. (thus achieving two-disk failure protection, at the price of more disk space spent on RAID5 checksum blocks) raid-disk index The most recently defined device is inserted at position index in the raid array. spare-disk index The most recently defined device is inserted at position index in the spare disk array. parity-disk index The most recently defined device is moved to the end of the raid array, which forces it to be used for parity. failed-disk index The most recently defined device is inserted at position index in the raid array as a failed device. This allows you to create raid 1/4/5 devices in degraded mode - useful for installation. Don't use the smallest device in an array for this, put this after the raid-disk definitions! NOTES
The raidtools are derived from the md-tools and raidtools packages, which were originally written by Marc Zyngier, Miguel de Icaza, Gadi Oxman, Bradley Ward Allen, and Ingo Molnar. SEE ALSO
raidstart(8), raid0run(8), mkraid(8), raidstop(8) raidtab(5)
All times are GMT -4. The time now is 09:18 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy