SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node


 
Thread Tools Search this Thread
Operating Systems Solaris SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node
# 1  
Old 04-14-2011
SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node

Hi,

Is it possible to have a Solaris cluster of 2 nodes at SITE-A using SVM and creating metaset using say 2 LUNs (on SAN). Then replicating these 2 LUNs to remote site SITE-B via storage based replication and then using these LUNs by importing them as a metaset on a server at SITE-B which is not running Solaris cluster? Solaris OS versions on all 3 nodes (2 nodes at SITE-A and 1 node at SITE-B) is same.
This is to say can I have a solution having 2 nodes in a Solaris cluster at SITE-A and have DR by having a single non-clustered node at SITE-B using SVM metaset replicated via storage replication.

Thanks.
Login or Register to Ask a Question

Previous Thread | Next Thread

10 More Discussions You Might Find Interesting

1. HP-UX

Mount FIle systems from node-1 onto node-2

Hi, We have HP UX service guard cluster on OS 11.23. Recently 40+ LUNs presented to both nodes by SAN team but I was asked to mount them on only one node. I created required VGs/LVs, created VxFS and mounted all of them and they are working fine. Now client requested those FS on 2nd node as... (4 Replies)
Discussion started by: prvnrk
4 Replies

2. AIX

Cluster node not starting

Setting up HACMP 6.1 on a two node cluster. The other node works fine and can start properly on STABLE state (VGs varied, FS mounted, Service IP aliased). However, the other node is always stuck on ST_JOINING state. Its taking forever and you can't stop the cluster as well or recover from script... (2 Replies)
Discussion started by: depam
2 Replies

3. UNIX for Advanced & Expert Users

VCS triggerring panic on 1 node, root disk under SVM

We have two node cluster with OS disk mirrored under SVM. There is slight disk problem on one of the mirror disk causing cluster to panic. Failure of one mirror disk causing VCS to panic the node. Why VCS is not able to write /var filesystem, as one of the disk is healthy. ... (1 Reply)
Discussion started by: amlanroy
1 Replies

4. Solaris

Tracing node to a particular HBA in Solaris 9

I have one disk that is reporting I/O errors but the same LUN mounted on a different node is able to access it without issue, is there a way to identify which HBA is being used for the LUN without swapping each out at a time? (4 Replies)
Discussion started by: thmnetwork
4 Replies

5. Solaris

unable to mount metaset on cluster node

Dear all, I have created a shared metaset(500gb) having 3 hosts in which 2 hosts are in cluster and 1 is non cluster. I have taken the ownership in cluster node from non cluster node but the problem is i am unable to mount the file system it is giving error "/dev/md/eccdb-ds/d100 or /eccdb-ds... (1 Reply)
Discussion started by: spandhan
1 Replies

6. High Performance Computing

Setting up 2 node cluster using solaris 10

hi, i am trying to setup a 2 node cluster environment. following is what i have; 1. 2 x sun ultra60 - 450MHz procs, 1GB RAM, 9GB HDD, solaris 10 2. 2 x HBA cards 3. 2 x Connection leads to connect ultra60 with D1000 4. 1 x D1000 storage box. 5. 3 x 9GB HDD + 2 x 36GB HDD first of all,... (1 Reply)
Discussion started by: solman17
1 Replies

7. High Performance Computing

Removed crashed node from Solaris Cluster 3.0

All- I am new to these forums so please excuse me if this post is in the wrong place. I had a node crash in a 4 node cluster and mgmt has determined this node will not be part of the cluster when rebuilt. I am researching how to remove it from the cluster information on the other 3 nodes and... (2 Replies)
Discussion started by: bluescreen
2 Replies

8. HP-UX

Node can't join cluster

Need help guys! when running cmrunnode batch i'm getting this error cmrunnode : Waiting for cluster to... (1 Reply)
Discussion started by: Tris
1 Replies

9. Solaris

Not able to copy the tree node text in solaris, while easily done in window

I m not able to copy the text present on the tree's node to terminal or other text editor in solaris. I m using <Shift><control> C and V comaand for the same but the text is not being copied and pasted on the text pad or the terminal window. While the same is possible in windows OS using ctrl+c... (3 Replies)
Discussion started by: friendanoop
3 Replies

10. Shell Programming and Scripting

The other node name of a SUN cluster

Hello, Under ksh I have to run a script on one of the nodes of a Solaris 8 cluster which at some time must execute a command on the alternate node: # rsh <name> "command" I have to implement this script on all the clusters of my company (a lot of...). Fortunately, the names of the two nodes... (11 Replies)
Discussion started by: heartwork
11 Replies
Login or Register to Ask a Question
cmhaltnode(1m)															    cmhaltnode(1m)

NAME
cmhaltnode - halt a node in a high availability cluster SYNOPSIS
cmhaltnode [-f] [-v] [-t] [node_name...] DESCRIPTION
cmhaltnode causes a node to halt its cluster daemon and remove itself from the existing cluster. To halt cluster on the node, a user must either be superuser(UID=0), or have an access policy of FULL_ADMIN allowed in the cluster configu- ration file. See access policy in cmquerycl. When cmhaltnode is run on a node, the cluster daemon is halted and, optionally, all packages that were running on that node are moved to other nodes if possible. If node_name is not specified, the cluster daemon running on the local node will be halted and removed from the existing cluster. If you issue this command while a cluster is still in the process of forming, the command will fail with the message "Unable to connect to daemon." If this happens, wait for the cluster to form successfully, then issue the command again. Options cmhaltnode supports the following options: -f Force the node to halt even if there are packages or group members running on it. The group members on the node will be terminated. The halt scripts for all packages running on the node will be run; based on priority or dependency relation- ships, this may affect packages on other nodes. In other words, packages on other nodes may either start or halt based on this package halting. If the package configuration and current cluster membership permit, and if the package halt script succeeds, the packages will be started on other nodes. Without this option, if packages are running on the given node, the command will fail. If a package fails to halt, the node halt will also fail. -v Verbose output will be displayed. -t Test only. Provide an assessment of the package placement without affecting the current state of the nodes or packages. This option validates the node's eligibility with respect to the package dependencies as well as the external dependencies such as EMS resources, package subnets, and storage before predicting any package placement decisions. If there is a pack- age in maintenance mode running on the nodes being halted, the package will always be halted and not failover to another node; the report will not display an assessment for that package. node_name... The name of the node(s) to halt. RETURN VALUE
cmhaltnode returns the following value: 0 Successful completion. 1 Command failed. EXAMPLES
Halt the cluster daemon on two other nodes: cmhaltnode node2 node3 AUTHOR
cmhaltnode was developed by HP. SEE ALSO
cmquerycl(1m), cmhaltcl(1m), cmruncl(1m), cmrunnode(1m), cmviewcl(1m), cmeval(1m). Requires Optional Serviceguard Software cmhaltnode(1m)