Sponsored Content
Full Discussion: Creating metadevices
Operating Systems Solaris Creating metadevices Post 303013939 by fretagi on Friday 2nd of March 2018 06:58:47 AM
Old 03-02-2018
So I cannot use the internal disk for state database replicas....
 

9 More Discussions You Might Find Interesting

1. Programming

creating so's

hi everyone i have a doubt about ".so" files. what is the need of ".so" files. why we use ".so" files. where we can use ".so" files. how can i create ".so" files. can u tell me a good example please thank you (6 Replies)
Discussion started by: ramesh.jella
6 Replies

2. Shell Programming and Scripting

help needed with creating challenging bash script with creating directories

Hi, Can someone help me with creating a bash shell script. I need to create a script that gets a positive number n as an argument. The script must create n directories in the current directory with names like map_1, map_2 etcetera. Each directory must be contained within its predecessor. So... (7 Replies)
Discussion started by: I-1
7 Replies

3. Solaris

Metadevices in mirroring ?

Hi Guys. I have the follow disk mappig.... My doubt is that the filesystem root is in mirroring. I can see this in the configuration but , I dont know exactly if this in mirroring mean disk in RAID. In short: Watching the configuration...Can I said if the filesystem / is in Raid? ... (3 Replies)
Discussion started by: aggadtech08
3 Replies

4. Solaris

Problem with live upgrade creation: Telling me metadevices do not exist

Hi I am having a problem creating my live upgrade environment. Here is the error I get: root@server:/# lucreate -c SOL10_18May -n SOL10_19May -z /lu_excludelist -m /:dev/md/dsk/d0:ufs -m /var:/dev/md/dsk/d4:ufs -m /export/home:/dev/md/dsk/d6:ufs -m -:/dev/md/dsk/d3:swap -C... (0 Replies)
Discussion started by: notreallyhere
0 Replies

5. Solaris

metadevices: how to test metadb (how to corrupt replicas/understanding replicas)

Hi all, I recently started exploring Solaris 10. I am testing metadevices now. I have been reading about the state databases here: 6.State Database (Overview) (Solaris Volume Manager Administration Guide) - Sun Microsystems So I created 3 metadbs on 2 slices (6 in total; c1t1d0s3... (3 Replies)
Discussion started by: deadeyes
3 Replies

6. Solaris

Metadevices deleted

Hi Guys, We have an issue, all metadb's on system was deleted and the system was rebooted. the system is currented mounted in single-user mode . its a x86 server. the volumes under SVM is as follows. / /var and /usr please suggest Please use code tags <- click the... (1 Reply)
Discussion started by: karthick.sh
1 Replies

7. Solaris

SVM - Metadevices are offline after changing hostname solaris x86

Hi , We are facing an issue on one of our solaris x86 server, After changing the hostname and a orderly reboot , all metadevices shows offline. please let us know the steps to restore back all metadevices to working state with this new hostname (3 Replies)
Discussion started by: karthick.sh
3 Replies

8. Homework & Coursework Questions

Creating a .profile, displaying system variables, and creating an alias

Use and complete the template provided. The entire template must be completed. If you don't, your post may be deleted! 1. The problem statement, all variables and given/known data: Here is what I am supposed to do, word for word from my assignment page: 1. Create/modify and print a... (2 Replies)
Discussion started by: Jagst3r21
2 Replies

9. SCO

Creating a VM

Also, in my tinkering around I was able to get SCO installed on a VM, using VM Workstation version 12.5 in a Windows 7 Pro-64 computer. I chose the VM Workstation 8.x compatibility model for my VM, an IDE hard drive at 0:0 and an IDE CD Rom at 1:0 using the physical drive of the computer. Using the... (3 Replies)
Discussion started by: jgt
3 Replies
mediator(7D)							      Devices							      mediator(7D)

NAME
mediator - support for HA configurations consisting of two strings of drives DESCRIPTION
Beginning with a prior version, Solaris Volume Manager provided support for high-availability (HA) configurations consisting of two hosts that share at least three strings of drives and that run software enabling exclusive access to the data on those drives from one host. (Note: Volume Manager, by itself, does not actually provide a high-availability environment. The diskset feature is an enabler for HA con- figurations.) Volume Manager provides support for a low-end HA solution consisting of two hosts that share only two strings of drives. The hosts in this type of configuration, referred to as mediators, run a special daemon, rpc.metamedd(1M). The mediator hosts take on additional responsibil- ities to ensure that data is available in the case of host or drive failures. In a mediator configuration, two hosts are physically connected to two strings of drives. This configuration can survive the failure of a single host or a single string of drives, without administrative intervention. If both a host and a string of drives fail (multiple fail- ures), the integrity of the data cannot be guaranteed. At this point, administrative intervention is required to make the data accessible. The following definitions pertain to a mediator configuration: diskset A set of drives containing metadevices and hot spares that can be shared exclusively (but not concurrently) by two hosts. Volume Manager state datAbreplicated database that stores metadevice configuration and state information. mediator host A host that runs the rpc.metamedd(1M) daemon and that has been added to a diskset. The mediator host participates in checking the state database and the mediator quorum. mediator quorum The condition achieved when the number of accessible mediator hosts is equal to half+1 the total number of config- ured mediator hosts. Because it is expected that there will be two mediator hosts, this number will normally be 2 ([(2/2) + 1] = 2.) replica A single copy of the Volume Manager metadevice state database. replica quorum The condition achieved when the number of accessible replicas is equal to half+1 the total number of configured replicas. For example, if a system is configured with ten replicas, the quorum is met when six are accessible ([(10/2) + 1 = 6]). A mediator host running the rpc.metamedd(1M) daemon keeps track of replica updates. As long as the following conditions are met, access to data occurs without any administrative intervention: o The replica quorum is not met. o Half of the replicas is still accessible. o The mediator quorum is met. The following conditions describe the operation of mediator hosts: 1. If the replica quorum is met, access to the diskset is granted. At this point no mediator host is involved. 2. If the replica quorum is not met, half of the replicas is accessible, the mediator quorum is met, and the replica and mediator data match, access to the diskset is granted. The mediator host contributes the deciding vote. 3. If the replica quorum is not met, half of the replicas is accessible, the mediator quorum is not met, half of the mediator hosts is accessible, and the replica and mediator data match, the system prompts you to grant or deny access to the diskset. 4. If the replica quorum is not met, half of the replicas is accessible, the mediator quorum is met, and the replica and mediator data do not match, access to the diskset is read-only. You can delete replicas, release the diskset, and retake the diskset to gain read-write access to the data in the diskset. 5. In all other cases, the diskset access is read-only. You can delete replicas, release the diskset, and retake the diskset to gain read-write access to the data in the diskset. The metaset(1M) command administers disksets and mediator hosts. The following options to the metaset command pertain only to administering mediator hosts. -a -m mediator_host_listAdds mediator hosts to the named set. A mediator_host_list is the nodename of the mediator host to be added and up to 2 other aliases for the mediator host. The nodename and aliases for each mediator host are separated by com- mas. Up to 2 mediator hosts can be specified for the named diskset. -d -m mediator_host_listDeletes mediator hosts from the named diskset. Mediator hosts are deleted from the diskset by specifying the node- name of mediator host to delete. -q Displays an enumerated list of tags pertaining to ``tagged data'' that may be encountered during a take of the own- ership of a diskset. -t [-f] -y Takes ownership of a diskset safely, unless -f is used, in which case the take is unconditional. If metaset finds that another host owns the set, this host will not be allowed to take ownership of the set. If the set is not owned by any other host, all the disks within the set will be owned by the host on which metaset was executed. The metadevice state database is read in and the shared metadevices contained in the set become accessible. The -t option will take a diskset that has stale databases. When the databases are stale, metaset will exit with code 66, and a message will be printed. At that point, the only operations permitted are the addition and deletion of repli- cas. Once the addition or deletion of the replicas has been completed, the diskset should be released and retaken to gain full access to the data. If mediator hosts have been configured, some additional exit codes are possible. If half of the replicas and half of the mediator hosts are operating properly, the take will exit with code 3. At this point, you can add or delete replicas, or use the -y option on a subsequent take. If the take operation encounters ``tagged data,'' the take operation will exit with code 2. You can then run the metaset command with the -q option to see an enumerated list of tags. -t [-f] -u tagnumber Once a tag has been selected, a subsequent take with -u tagnumber can be executed to select the data associated with the given tagnumber. SEE ALSO
metaset(1M), md(7D), rpc.metamedd(1M), rpc.metad(1M) Sun Cluster documentation, Solaris Volume Manager Administration Guide NOTES
Diskset administration, including the addition and deletion of hosts and drives, requires all hosts in the set to be accessible from the network. SunOS 5.10 20 September 2000 mediator(7D)
All times are GMT -4. The time now is 12:50 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy