Sponsored Content
Full Discussion: Metadevices in mirroring ?
Operating Systems Solaris Metadevices in mirroring ? Post 302360302 by aggadtech08 on Thursday 8th of October 2009 02:35:59 PM
Old 10-08-2009
Metadevices in mirroring ?

Hi Guys.

I have the follow disk mappig....

My doubt is that the filesystem root is in mirroring. I can see this in the configuration but , I dont know exactly if this in mirroring mean disk in RAID.

In short: Watching the configuration...Can I said if the filesystem / is in Raid?

Thanks in Advanced,
AGAD
Code:
root@sap02 # df -h
Filesystem             size   used  avail capacity  Mounted on
/dev/md/dsk/d10        7.9G   7.2G   599M    93%     /
/proc                        0K     0K     0K         0%       /proc
mnttab                      0K     0K     0K         0%       /etc/mnttab
fd                             0K     0K     0K         0%       /dev/fd
/dev/md/dsk/d40         2.0G   1.6G   295M    85%     /var
swap                         2.4G   48K   2.4G      1%       /var/run
swap                         3.3G   49M   2.4G      26%     /tmp
/dev/dsk/c0t0d0s5       8.9G   6.1G   2.6G     70%     /oracle
/dev/dsk/c0t0d0s6       12G    10G   1.5G      88%     /data

 
root@sap02 # metastat
d40: Mirror
    Submirror 0: d41
      State: Okay
    Submirror 1: d42
      State: Okay
    Pass: 1
    Read option: roundrobin (default)
    Write option: parallel (default)
    Size: 4194828 blocks (2.0 GB)
d41: Submirror of d40
    State: Okay
    Size: 4194828 blocks (2.0 GB)
    Stripe 0:
        Device     Start Block  Dbase        State Reloc Hot Spare
        c0t0d0s4          0     No            Okay   Yes

d42: Submirror of d40
    State: Okay
    Size: 4194828 blocks (2.0 GB)
    Stripe 0:
        Device     Start Block  Dbase        State Reloc Hot Spare
        c0t1d0s4          0     No            Okay   Yes

d20: Mirror
    Submirror 0: d21
      State: Okay
    Submirror 1: d22
      State: Okay
    Pass: 1
    Read option: roundrobin (default)
    Write option: parallel (default)
    Size: 4194828 blocks (2.0 GB)
d21: Submirror of d20
    State: Okay
    Size: 4194828 blocks (2.0 GB)
    Stripe 0:
        Device     Start Block  Dbase        State Reloc Hot Spare
        c0t0d0s1          0     No            Okay   Yes

d22: Submirror of d20
    State: Okay
    Size: 4194828 blocks (2.0 GB)
    Stripe 0:
        Device     Start Block  Dbase        State Reloc Hot Spare
        c0t1d0s1          0     No            Okay   Yes

d10: Mirror
    Submirror 0: d11
      State: Okay
    Submirror 1: d12
      State: Okay
    Pass: 1
    Read option: roundrobin (default)
    Write option: parallel (default)
    Size: 16779312 blocks (8.0 GB)
d11: Submirror of d10
    State: Okay
    Size: 16779312 blocks (8.0 GB)
    Stripe 0:
        Device     Start Block  Dbase        State Reloc Hot Spare
        c0t0d0s0          0     No            Okay   Yes

d12: Submirror of d10
    State: Okay
    Size: 16779312 blocks (8.0 GB)
    Stripe 0:
        Device     Start Block  Dbase        State Reloc Hot Spare
        c0t1d0s0          0     No            Okay   Yes

Device Relocation Information:
Device   Reloc  Device ID
c0t1d0   Yes    id1,sd@SSEAGATE_ST336607LSUN36G_3JA73PMN000074299YRL
c0t0d0   Yes    id1,sd@SSEAGATE_ST336607LSUN36G_3JA73RKG00007430WFD2
root@sap02 #


Last edited by pludi; 10-08-2009 at 04:04 PM.. Reason: code tags, please
 

9 More Discussions You Might Find Interesting

1. Solaris

Mirroring

I am running Solaris 10 and i need to mirror a 73 gig HD. How do you mirror one in Solaris? (2 Replies)
Discussion started by: dewsdwarfs
2 Replies

2. SCO

Mirroring

How Can I Do Mirroring In Unix? (0 Replies)
Discussion started by: DIMITRIOSDOUMOS
0 Replies

3. Solaris

Mirroring

Hi All i wish to mirror the root disk, but i face the below error. root@saturn # metainit d11 1 1 c0t0d0s0 metainit: saturn: c0t0d0s0: is mounted on / kindly assist... (27 Replies)
Discussion started by: SmartAntz
27 Replies

4. Solaris

Problem with live upgrade creation: Telling me metadevices do not exist

Hi I am having a problem creating my live upgrade environment. Here is the error I get: root@server:/# lucreate -c SOL10_18May -n SOL10_19May -z /lu_excludelist -m /:dev/md/dsk/d0:ufs -m /var:/dev/md/dsk/d4:ufs -m /export/home:/dev/md/dsk/d6:ufs -m -:/dev/md/dsk/d3:swap -C... (0 Replies)
Discussion started by: notreallyhere
0 Replies

5. Solaris

metadevices: how to test metadb (how to corrupt replicas/understanding replicas)

Hi all, I recently started exploring Solaris 10. I am testing metadevices now. I have been reading about the state databases here: 6.State Database (Overview) (Solaris Volume Manager Administration Guide) - Sun Microsystems So I created 3 metadbs on 2 slices (6 in total; c1t1d0s3... (3 Replies)
Discussion started by: deadeyes
3 Replies

6. Solaris

Metadevices deleted

Hi Guys, We have an issue, all metadb's on system was deleted and the system was rebooted. the system is currented mounted in single-user mode . its a x86 server. the volumes under SVM is as follows. / /var and /usr please suggest Please use code tags <- click the... (1 Reply)
Discussion started by: karthick.sh
1 Replies

7. Solaris

SVM - Metadevices are offline after changing hostname solaris x86

Hi , We are facing an issue on one of our solaris x86 server, After changing the hostname and a orderly reboot , all metadevices shows offline. please let us know the steps to restore back all metadevices to working state with this new hostname (3 Replies)
Discussion started by: karthick.sh
3 Replies

8. Solaris

ZFS mirroring

Hello, I just build a Solaris 10 server on an x86 box. I forgot to mirror the two disks when I install the OS. Can I get some help with this? I have this # zpool list rpool 278G 5.77G 272G 2% ONLINE - # zpool status pool: online state: ONLINE scan: none requested config: ... (12 Replies)
Discussion started by: bitlord
12 Replies

9. Solaris

Creating metadevices

Hi I am new to SVM, but I would like to configure it in one of my servers. The setup is as follows: I have one server running solaris 10 , and is connected to a NetApp via iSCSI protocol. This is how I intend to do it: Ask the NetApp admin, to provide me with 3 LUNs of 100G each, them use the... (14 Replies)
Discussion started by: fretagi
14 Replies
metastat(1M)															      metastat(1M)

NAME
metastat - display status for metadevice or hot spare pool SYNOPSIS
/usr/sbin/metastat -h /usr/sbin/metastat [-a] [-B] [-c] [-i] [-p] [-q] [-s setname] [-t] [metadevice...] [hot_spare_pool...] /usr/sbin/metastat [-a] [-B] [-c] [-i] [-p] [-q] [-s setname] component... The metastat command displays the current status for each metadevice (including stripes, concatenations, concatenations of stripes, mir- rors, RAID5, soft partitions, and trans devices) or hot spare pool, or of specified metadevices, components, or hot spare pools. It is helpful to run the metastat command after using the metattach command to view the status of the metadevice. metastat displays the state of each Solaris Volume Manager RAID-1 volume on the system. The possible states include: Okay The device reports no errors. Needs maintenance A problem has been detected. This requires that the system administrator replace the failed physical device. Vol- umes displaying Needs maintenance have incurred no data loss, although additional failures could risk data loss. Take action as quickly as possible. Last erred A problem has been detected. Data loss is a possibility. This might occur if a component of a submirror fails and is not replaced by a hot spare, therefore going into Needs maintenance state. If the corresponding component also fails, it would go into Last erred state and, as there is no remaining valid data source, data loss could be a pos- sibility. Unavailable A device cannot be accessed, but has not incurred errors. This might occur if a physical device has been removed with Solaris Dynamic Reconfiguration (DR) features, thus leaving the Solaris Volume Manager volume unavailable. It could also occur if an array or disk is powered off at system initialization, or if a >1TB volume is present when the system is booted in 32-bit mode. After the storage has been made available, run the metastat command with the -i option to update the status of the metadevices. This clears the unavailable state for accessible devices. See the for instructions on replacing disks and handling volumes in Needs maintenance or Last erred states. The following options are supported: -a Display all disk sets. Only metadevices in disk sets that are owned by the current host are displayed. -B Display the current status of all of the 64-bit metadevices and hot spares. -c Display concise output. There is one line of output for each metadevice. The output shows the basic structure and the error status, if any, for each metadevice. The -c output format is distinct from the -p output format. The -p option does not display metadevice status and is not intended as human-readable output. -h Display usage message. -i Check the status of RAID-1 (mirror) volumes, RAID-5 volumes, and hot spares. The inquiry checks each metadevice for accessibility, starting at the top level metadevice. When problems are discovered, the metadevice state databases are updated as if an error had occurred. -p Display the list of active metadevices and hot spare pools in the same format as md.tab. See md.tab(4). The -p output is designed for snapshotting the configuration for later recovery or setup. -q Display the status for metadevices without the device relocation information. -s setname Specify the name of the disk set on which metastat works. Using the -s option causes the command to perform its administrative function within the specified disk set. Without this option, the command performs its function on metadevices and hot spare pools in the local disk set. -t Display the current status and timestamp for the specified metadevices and hot spare pools. The timestamp provides the date and time of the last state change. The following operands are supported: component Display the status of the component hosting a soft partition, including extents, starting blocks, and block count. hot_spare_pool Display the status of the specified hot spare pool(s). metadevice Display the status of the specified metadevice(s). If a trans metadevice is specified, the status of the master and log devices is also displayed. Trans metadevices have been replaced by UFS logging. See NOTES. Example 1: Output Showing Mirror with Two Submirrors The following example shows the partial output of the metastat command after creating a mirror, d0, consisting of two submirrors, d70 and d80. # metastat d0 d0: Mirror Submirror 0: d80 State: Okay Submirror 1: d70 State: Resyncing Resync in progress: 15 % done Pass: 1 Read option: roundrobin (default) Write option: parallel (default) Size: 2006130 blocks . . . Example 2: Soft Partition on Mirror with Submirror The following example shows the partial output of the metastat command after creating a soft partition, d3, on concat d2, which is built on a soft partition. # metastat d2: Concat/Stripe Size: 204800 blocks Stripe 0: Device Start Block Dbase State Hot Spare d0 0 No Okay d0: Soft Partition Component: c0t3d0s0 Status: Okay Size: 204800 blocks Extent Start Block Block count 0 129 204800 d3: Soft Partition Component: d2 Status: Okay Size: 202752 blocks Extent Start Block Block count 0 129 202752 Example 3: Trans Metadevice The following example shows the output of the metastat command after creating a trans metadevice. # metastat d2: Concat/Stripe Size: 204800 blocks Stripe 0: Device Start Block Dbase State Hot Spare d0 0 No Okay d0: Soft Partition Component: c0t3d0s0 Status: Okay Size: 204800 blocks Extent Start Block Block count 0 129 204800 d3: Soft Partition Component: d2 Status: Okay Size: 202752 blocks Extent Start Block Block count 0 129 202752 Example 4: Multi-owner disk set The following example shows the output of the metastat command with a multi-owner disk set and application-based mirror resynchronization option. Application-based resynchronization is set automatically if needed. # metastat -s oban oban/d100: Mirror Submirror 0: oban/d10 State: Okay Submirror 1: oban/d11 State: Okay Pass: 1 Read option: roundrobin (default) Write option: parallel (default) Resync option: application based Owner: None Size: 1027216 blocks (501 MB) oban/d10: Submirror of oban/d100 State: Okay Size: 1027216 blocks (501 MB) Stripe 0: Device Start Block Dbase State Reloc Hot Spare c1t3d0s0 0 No Okay oban/d11: Submirror of oban/d100 State: Okay Size: 1027216 blocks (501 MB) Stripe 0: Device Start Block Dbase State Reloc Hot Spare c1t4d0s0 0 No Okay WARNINGS
metastat displays states as of the time the command is entered. It is unwise to use the output of the metastat -p command to create a md.tab(4) file for a number of reasons: o The output of metastat -p might show hot spares being used. o It might show mirrors with multiple submirrors. See metainit(1M) for instructions for creating multi-way mirrors using metainit and metattach. o A slice may go into an error state after metastat -p is issued. The following exit values are returned: 0 Successful completion. >0 An error occurred. See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Availability |SUNWmdr | +-----------------------------+-----------------------------+ |Stability |Evolving | +-----------------------------+-----------------------------+ mdmonitord(1M), metaclear(1M), metadb(1M), metadetach(1M), metahs(1M), metainit(1M), metaoffline(1M), metaonline(1M), metaparam(1M), metarecover(1M), metarename(1M), metareplace(1M), metaroot(1M), metaset(1M), metassist(1M), metasync(1M), metattach(1M), md.tab(4), md.cf(4), mddb.cf(4), md.tab(4), attributes(5), md(7D) Trans metadevices have been replaced by UFS logging. Existing trans devices are not logging--they pass data directly through to the under- lying device. See mount_ufs(1M) for more information about UFS logging. 30 Mar 2005 metastat(1M)
All times are GMT -4. The time now is 02:38 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy