Sponsored Content
Full Discussion: Linux Cluster failover issue
Operating Systems Linux Red Hat Linux Cluster failover issue Post 302895843 by munna529 on Wednesday 2nd of April 2014 10:07:08 PM
Old 04-02-2014
RedHat Linux Cluster failover issue

Hi Guys,

I am not much aware of clusters but i have few questions can someone provide the overview as it would be very helpful for me.

How can i perform cluster failover test to see all the services are failing back to other node ? If it is using veritas cluster then what kind of pre-prequisite i need to follow ?

If it is not using veritas cluster then how to check whether host is clustered or not ? if clustered how can i perform failover to other nodes?

If my question is not clear please forgive me and let me know what are the required steps i need to follow this procedure ?

Thanks
 

9 More Discussions You Might Find Interesting

1. High Performance Computing

sun Cluster resource group cant failover

I have rcently setup a 4 node cluster running sun cluster 3.2 and I have installed 4 zones on each node. when installing the zones I had to install the zone on all nodes the on the last node do a zlogin -C <zonename> this worked ok. theni I tried to siwitch the zone to node a thei work... (14 Replies)
Discussion started by: lesliek
14 Replies

2. HP-UX

ServiceGuard cluster & volume group failover

I have a 2-node ServiceGuard cluster. One of the cluster packages has a volume group assigned to it. When I fail the package over to the other node, the volume group does not come up automatically on the other node. I have to manually do a "vgchange -a y vgname" on the node before the package... (5 Replies)
Discussion started by: Wotan31
5 Replies

3. High Performance Computing

Veritas Cluster Server Management Console IP Failover

I have just completed a first RTFM of "Veritas Cluster Server Management Console Implementation Guide" 5.1, with a view to assessing it to possibly make our working lives easier. Unfortunately, at my organisation, getting a test installation would be worse than pulling teeth, so I can't just go... (2 Replies)
Discussion started by: Beast Of Bodmin
2 Replies

4. Solaris

Sun Cluster 3.1 failover

Hi, We have two sun SPARC server in Clustered (Sun Cluster 3.1). For some reason, System 1 failed over to System 2. Where can I find the logs which could tell me the reason for this failover? Thanks (5 Replies)
Discussion started by: Mack1982
5 Replies

5. Gentoo

How to failover the cluster ?

How to failover the cluster ? GNU/Linux By which command, My Linux version 2008 x86_64 x86_64 x86_64 GNU/Linux What are the prerequisites we need to take while failover ? if any Regards (3 Replies)
Discussion started by: sidharthmellam
3 Replies

6. Solaris

Sun cluster 4.0 - zone cluster failover doubt

Hello experts - I am planning to install a Sun cluster 4.0 zone cluster fail-over. few basic doubts. (1) Where should i install the cluster s/w binaries ?. ( global zone or the container zone where i am planning to install the zone fail-over) (2) Or should i perform the installation on... (0 Replies)
Discussion started by: NVA
0 Replies

7. Solaris

Sun Cluster 3.2 Issue

Hello everyone, I have two Solaris 10 servers that are on cluster. The cluster is a Sun Cluster 3.2 I have a script cronned that stop/start a ressource in a resource group everyday. Today I have checked the status of the ressources and I found that my ressource group have a "Error--stop... (1 Reply)
Discussion started by: adilyos
1 Replies

8. Solaris

Solaris Cluster Failover based on scan rate

Dear Experts, If there is a possible Solaris Cluster failover to second node based on scan rate? I need the documentation If solaris cluster can do this. Thank You in Advance Edy (3 Replies)
Discussion started by: edydsuranta
3 Replies

9. AIX

Cluster communication issue

Hi, I am using Power HA7.1.1 SP5 AIx 7.1 My both cluster nodes are independently working. RG informations are not updating each other. Node A shows that node B is down and vice versa. RG1 is running node A, RG2 running on node B. === clRGinfo From Node B === RG01 OFFLINE ... (2 Replies)
Discussion started by: sunnybee
2 Replies
clusvcadm(8)						       Red Hat Cluster Suite						      clusvcadm(8)

NAME
clusvcadm - Cluster User Service Administration Utility SYNOPSIS
clusvcadm [-d <service> ] [-e <service> ] [-l] [-u] [-S] [-m <member> ] [-r <service> ] [-R <service> ] [-M <service> ] [-Z <service> ] [-U <service> ] [-s <service> ] [-v] DESCRIPTION
The clusvcadm command allows an administrator to enable, disable, relocate, and restart user services in a cluster. In order to perform cluster service operations, the cluster daemons must be running (and have quorum) on the member system on which the command is invoked. OPTIONS
-d <service> Stops and disables the user service named service -e <service> Enables and starts the user service named service -l Lock services in preparation for cluster shutdown. This should only be used if the administrator intends to perform a global, clus- ter wide shutdown. This prevents services from starting (but not stopping, like -Z does). Once the cluster quorum is dissolved, this state is reset. If a new instance of rgmanager boots while others are locked, the behavior is undefined. -u Unlock resource group managers. This allows services to start again. -S Display whether the resource group managers are locked or not. This can be used to verify the correct operation of the -l and -u options, but is only useful for debugging. -m <member> When used in conjunction with either the -e or -r options, this specifies the preferred target member on which to start the service. -r <service> Relocates the user service named service to another cluster member. -R <service> Restarts the user service named service on the cluster member on which it is currently running. -M <service> Use a special "migration" operation to move the user service named service to another cluster member. Currently, this is only use- ful for virtual machines. Use of migration usually requires special configuration of the local virtual machine manager in order to work correctly. -Z <service> Freezes the service named service on the cluster member on which it is currently running. This will prevent status checks of the service as well as failover in the event the node fails or rgmanager is stopped. -U <service> Unfreezes the user service named service on the cluster member on which it is currently running. This will re-enable status checks. -s <service> Stops the service named service until a member transition or until it is enabled again. -v Display version information and exit. NOTES
Executing -U (unfreeze) on a service which was frozen in the started state while the service owner is offline results in an undefined (and possibly dangerous) condition. Manually ensure all resources are clear before doing this. SEE ALSO
clustat(8) Jan 2005 clusvcadm(8)
All times are GMT -4. The time now is 06:19 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy