9 More Discussions You Might Find Interesting
1. Solaris
Hi
We have an M3000 single physical processor and 8gb of memory running Solaris 10. This system runs two Oracle Databases one on Oracle 9i and One on Oracle 10g.
As soon as the Oracle 10g database starts we see an immediate drop in system performance, for example opening an ssh session can... (6 Replies)
Discussion started by: gregsih
6 Replies
2. AIX
Hi,
I am new registered user here in this UNIX forums.
I am a new system administrator for AIX 6.1. One of our servers performs poorly every time our application (FINACLE) runs many processes/instances. (see below for topas snapshot)
I use NMON or Topas to monitor the server utilization. I... (9 Replies)
Discussion started by: guzzelle
9 Replies
3. Solaris
Hello,
we have a machine with Solaris Express 11, 2 LSI 9211 8i SAS 2 controllers (multipath to disks), multiport backplane, 16 Seagate Cheetah 15K RPM disks.
Each disk has a sequential performance of 220/230 MB/s and in fact if I do a
dd if=/dev/zero of=/dev/rdsk/<diskID_1> bs=1024k... (1 Reply)
Discussion started by: golemico
1 Replies
4. UNIX for Advanced & Expert Users
Hello all
We just built a storage cluster for our new xenserver farm. Using 3ware 9650SE raid controllers with 8 x 1TB WD sata disks in a raid 5, 256KB stripe size.
While making first performance test on the local storage server using dd (which simulates the read/write access to the disk... (1 Reply)
Discussion started by: roli8200
1 Replies
5. Solaris
Hello Experts,
I am using E250 on that solais 10 5/08 installed. I am unable to see disks. I connected 2 disks in that storage of 18gb each. When I run format command it is showing that 2 disks one is operating system and another one is 6MB. I checked probe-scsi and probe-scsi-all at ok... (6 Replies)
Discussion started by: younus_syed
6 Replies
6. UNIX for Advanced & Expert Users
Dear Sun gurus,
I have Sun Fire V240 server with its StorEdge 3300 disk-array. Following are its disks appeared in format command. I have prepared its partitions thru format and metainit & metattach (may be i have made wrong steps, causing the errors below because I have done thru some document... (1 Reply)
Discussion started by: shafeeq
1 Replies
7. UNIX for Advanced & Expert Users
Hello everyone. I'm trying to setup two A1000s connected to a single host w/ a dual port adapter. The host is a V480.
Do I need to have thesame firmware version on both controllers for the A1000s? If so, where can I download the latest and greatest firmware? I tried to google for it and... (8 Replies)
Discussion started by: xnightcrawl
8 Replies
8. UNIX for Advanced & Expert Users
Hi you all, I have a BIG performance problem on an Sun E3500, the scenario is described below:
I have several users (30) accessing via samba to the E3500 using an application built on Visual Foxpro from their Windows PC , the problem is that the first guy that logs in demands 30% of the E3500... (2 Replies)
Discussion started by: alex blanco
2 Replies
9. UNIX for Dummies Questions & Answers
Hi,
I am a Dba and very new to filesystems and stuff.
I think that we have Veritas filesystems on my Sun SOlaris 5.8 box, how do I confirm this: all my filesystems are mounted like this:
/dev/vx/dsk...
Now we are also using disk arrays (storedge a1000) how do I access them from the system.... (1 Reply)
Discussion started by: knarayan
1 Replies
GRAID(8) BSD System Manager's Manual GRAID(8)
NAME
graid -- control utility for software RAID devices
SYNOPSIS
graid label [-f] [-o fmtopt] [-S size] [-s strip] format label level prov ...
graid add [-f] [-S size] [-s strip] name label level
graid delete [-f] name [label | num]
graid insert name prov ...
graid remove name prov ...
graid fail name prov ...
graid stop [-fv] name ...
graid list
graid status
graid load
graid unload
DESCRIPTION
The graid utility is used to manage software RAID configurations, supported by the GEOM RAID class. GEOM RAID class uses on-disk metadata to
provide access to software-RAID volumes defined by different RAID BIOSes. Depending on RAID BIOS type and its metadata format, different
subsets of configurations and features are supported. To allow booting from RAID volume, the metadata format should match the RAID BIOS type
and its capabilities. To guarantee that these match, it is recommended to create volumes via the RAID BIOS interface, while experienced
users are free to do it using this utility.
The first argument to graid indicates an action to be performed:
label Create an array with single volume. The format argument specifies the on-disk metadata format to use for this array, such as
"Intel". The label argument specifies the label of the created volume. The level argument specifies the RAID level of the created
volume, such as: "RAID0", "RAID1", etc. The subsequent list enumerates providers to use as array components. The special name
"NONE" can be used to reserve space for absent disks. The order of components can be important, depending on specific RAID level
and metadata format.
Additional options include:
-f Enforce specified configuration creation if it is officially unsupported, but technically can be created.
-o fmtopt
Specifies metadata format options.
-S size Use size bytes on each component for this volume. Should be used if several volumes per array are planned, or if smaller
components going to be inserted later. Defaults to size of the smallest component.
-s strip Specifies strip size in bytes. Defaults to 131072.
add Create another volume on the existing array. The name argument is the name of the existing array, reported by label command. The
rest of arguments are the same as for the label command.
delete Delete volume(s) from the existing array. When the last volume is deleted, the array is also deleted and its metadata erased. The
name argument is the name of existing array. Optional label or num arguments allow specifying volume for deletion.
Additional options include:
-f Delete volume(s) even if it is still open.
insert Insert specified provider(s) into specified array instead of the first missing or failed components. If there are no such compo-
nents, mark disk(s) as spare.
remove Remove the specified provider(s) from the specified array and erase metadata. If there are spare disks present, the removed disk(s)
will be replaced by spares.
fail Mark the given disks(s) as failed, removing from active use unless absolutely necessary due to exhausted redundancy. If there are
spare disks present - failed disk(s) will be replaced with one of them.
stop Stop the given array. The metadata will not be erased.
Additional options include:
-f Stop the given array even if some of its volumes are opened.
list See geom(8).
status See geom(8).
load See geom(8).
unload See geom(8).
Additional options include:
-v Be more verbose.
SUPPORTED METADATA FORMATS
The GEOM RAID class follows a modular design, allowing different metadata formats to be used. Support is currently implemented for the fol-
lowing formats:
DDF The format defined by the SNIA Common RAID Disk Data Format v2.0 specification. Used by some Adaptec RAID BIOSes and some hardware
RAID controllers. Because of high format flexibility different implementations support different set of features and have different
on-disk metadata layouts. To provide compatibility, the GEOM RAID class mimics capabilities of the first detected DDF array.
Respecting that, it may support different number of disks per volume, volumes per array, partitions per disk, etc. The following con-
figurations are supported: RAID0 (2+ disks), RAID1 (2+ disks), RAID1E (3+ disks), RAID3 (3+ disks), RAID4 (3+ disks), RAID5 (3+
disks), RAID5E (4+ disks), RAID5EE (4+ disks), RAID5R (3+ disks), RAID6 (4+ disks), RAIDMDF (4+ disks), RAID10 (4+ disks), SINGLE (1
disk), CONCAT (2+ disks).
Format supports two options "BE" and "LE", that mean big-endian byte order defined by specification (default) and little-endian used
by some Adaptec controllers.
Intel The format used by Intel RAID BIOS. Supports up to two volumes per array. Supports configurations: RAID0 (2+ disks), RAID1 (2
disks), RAID5 (3+ disks), RAID10 (4 disks). Configurations not supported by Intel RAID BIOS, but enforceable on your own risk: RAID1
(3+ disks), RAID1E (3+ disks), RAID10 (6+ disks).
JMicron
The format used by JMicron RAID BIOS. Supports one volume per array. Supports configurations: RAID0 (2+ disks), RAID1 (2 disks),
RAID10 (4 disks), CONCAT (2+ disks). Configurations not supported by JMicron RAID BIOS, but enforceable on your own risk: RAID1 (3+
disks), RAID1E (3+ disks), RAID10 (6+ disks), RAID5 (3+ disks).
NVIDIA
The format used by NVIDIA MediaShield RAID BIOS. Supports one volume per array. Supports configurations: RAID0 (2+ disks), RAID1 (2
disks), RAID5 (3+ disks), RAID10 (4+ disks), SINGLE (1 disk), CONCAT (2+ disks). Configurations not supported by NVIDIA MediaShield
RAID BIOS, but enforceable on your own risk: RAID1 (3+ disks).
Promise
The format used by Promise and AMD/ATI RAID BIOSes. Supports multiple volumes per array. Each disk can be split to be used by up to
two arbitrary volumes. Supports configurations: RAID0 (2+ disks), RAID1 (2 disks), RAID5 (3+ disks), RAID10 (4 disks), SINGLE (1
disk), CONCAT (2+ disks). Configurations not supported by RAID BIOSes, but enforceable on your own risk: RAID1 (3+ disks), RAID10 (6+
disks).
SiI The format used by SiliconImage RAID BIOS. Supports one volume per array. Supports configurations: RAID0 (2+ disks), RAID1 (2
disks), RAID5 (3+ disks), RAID10 (4 disks), SINGLE (1 disk), CONCAT (2+ disks). Configurations not supported by SiliconImage RAID
BIOS, but enforceable on your own risk: RAID1 (3+ disks), RAID10 (6+ disks).
SUPPORTED RAID LEVELS
The GEOM RAID class follows a modular design, allowing different RAID levels to be used. Full support for the following RAID levels is cur-
rently implemented: RAID0, RAID1, RAID1E, RAID10, SINGLE, CONCAT. The following RAID levels supported as read-only for volumes in optimal
state (without using redundancy): RAID4, RAID5, RAID5E, RAID5EE, RAID5R, RAID6, RAIDMDF.
RAID LEVEL MIGRATION
The GEOM RAID class has no support for RAID level migration, allowed by some metadata formats. If you started migration using BIOS or in
some other way, make sure to complete it there. Do not run GEOM RAID class on migrating volumes under pain of possible data corruption!
2TiB BARRIERS
NVIDIA metadata format does not support volumes above 2TiB.
SYSCTL VARIABLES
The following sysctl(8) variable can be used to control the behavior of the RAID GEOM class.
kern.geom.raid.aggressive_spare: 0
Use any disks without metadata connected to controllers of the vendor matching to volume metadata format as spare. Use it with much
care to not lose data if connecting unrelated disk!
kern.geom.raid.clean_time: 5
Mark volume as clean when idle for the specified number of seconds.
kern.geom.raid.debug: 0
Debug level of the RAID GEOM class.
kern.geom.raid.enable: 1
Enable on-disk metadata taste.
kern.geom.raid.idle_threshold: 1000000
Time in microseconds to consider a volume idle for rebuild purposes.
kern.geom.raid.name_format: 0
Providers name format: 0 -- raid/r{num}, 1 -- raid/{label}.
kern.geom.raid.read_err_thresh: 10
Number of read errors equated to disk failure. Write errors are always considered as disk failures.
kern.geom.raid.start_timeout: 30
Time to wait for missing array components on startup.
kern.geom.raid.X.enable: 1
Enable taste for specific metadata or transformation module.
kern.geom.raid.legacy_aliases: 0
Enable geom raid emulation of legacy /dev/ar%d devices. This should aid the upgrade of systems from legacy to modern releases.
EXIT STATUS
Exit status is 0 on success, and non-zero if the command fails.
SEE ALSO
geom(4), geom(8), gvinum(8)
HISTORY
The graid utility appeared in FreeBSD 9.0.
AUTHORS
Alexander Motin <mav@FreeBSD.org>
M. Warner Losh <imp@FreeBSD.org>
BSD
April 4, 2013 BSD