Sponsored Content
Operating Systems Linux Red Hat Switching services from active node to failovernode in cluster Post 302969241 by vamshigvk475 on Sunday 20th of March 2016 10:18:17 AM
Old 03-20-2016
RedHat Switching services from active node to failovernode in cluster

Hi All,

I was asked to Switch the service from failover to active node as we need to reboot active node for regular maintenance. Unfortunately our senior admin left the organization. It would be helpful if some one answer this query. Detailed info with steps will be appreciated.


Here are the below details. I need to reboot lab1 and need to switch all services to lab2 or lab3.

I believe the /appfs is related to cluster. I'm bit rusty with cluster administration Smilie
Please help!!!Smilie

Code:
#hostname
lab2

#clustat

 Member Name                              Status
    ------ ----                              ------
    lab1-int                             Online, rgmanager
    lab2-int                             Online, Local, rgmanager
    lab3-int                             Online, rgmanager
   
    Service Name         Owner (Last)                   State
    ------- ----         ----- ------                   -----
    app_gfs_svc         lab2-int                   started
    app_prod            lab1-int                   started
    app_test            (lab1-int)                 disabled
    app_upgrade         (none)                         disabled


  # df -TPH /appfs
  Filesystem    Type     Size   Used  Avail Use% Mounted on
  /dev/dm-50     gfs      32G    24G   8.3G  75% /appfs


Last edited by Scrutinizer; 03-20-2016 at 12:05 PM.. Reason: code tags
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

The other node name of a SUN cluster

Hello, Under ksh I have to run a script on one of the nodes of a Solaris 8 cluster which at some time must execute a command on the alternate node: # rsh <name> "command" I have to implement this script on all the clusters of my company (a lot of...). Fortunately, the names of the two nodes... (11 Replies)
Discussion started by: heartwork
11 Replies

2. HP-UX

Node can't join cluster

Need help guys! when running cmrunnode batch i'm getting this error cmrunnode : Waiting for cluster to... (1 Reply)
Discussion started by: Tris
1 Replies

3. UNIX for Dummies Questions & Answers

Switching the active pts?

My apologies in advance if this question has already been asked or if I use incorrect terminology; I have tried searching for nearly an hour without any luck. Is there any way to change the active TTY for a session? I was monitoring a long (~5 hour) process through an SSH connection and the... (4 Replies)
Discussion started by: tspicher
4 Replies

4. HP-UX

MC/SG Fail to join cluster node

Hi, Please advise me whereas I have two node cluster server configured with MC/SG. Application and DB are running on Node 1, while Node 2 is standby. All the volume group devices are part of cluster environment. There is only one package running at node 1. Node 2 is having the problem to... (1 Reply)
Discussion started by: rauphelhunter
1 Replies

5. High Performance Computing

Setting up 2 node cluster using solaris 10

hi, i am trying to setup a 2 node cluster environment. following is what i have; 1. 2 x sun ultra60 - 450MHz procs, 1GB RAM, 9GB HDD, solaris 10 2. 2 x HBA cards 3. 2 x Connection leads to connect ultra60 with D1000 4. 1 x D1000 storage box. 5. 3 x 9GB HDD + 2 x 36GB HDD first of all,... (1 Reply)
Discussion started by: solman17
1 Replies

6. Solaris

Active Sun cluster node?

I now the logical name and Virtual IP of the cluster. How can I find the active sun cluster node without having root access? (3 Replies)
Discussion started by: sreeniatbp
3 Replies

7. HP-UX

Identify cluster active node

Hello, Is there any way to identify the active node in a HP-UX cluster without root privileges? (3 Replies)
Discussion started by: psimoes79
3 Replies

8. Solaris

SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node

Hi, Is it possible to have a Solaris cluster of 2 nodes at SITE-A using SVM and creating metaset using say 2 LUNs (on SAN). Then replicating these 2 LUNs to remote site SITE-B via storage based replication and then using these LUNs by importing them as a metaset on a server at SITE-B which is... (0 Replies)
Discussion started by: dn2011
0 Replies

9. Solaris

Sun Cluster switching issue

I have installed sun cluster 3.2 on two sprac servers. Configured a failover resource group. Added a LogicalHostname resource to it. LogicalHostname is also added to /etc/hosts with ip address. I am able to access cluster by share ip used for logical hostname but when i try to switch the resource... (0 Replies)
Discussion started by: ahmadnauman
0 Replies

10. AIX

Cluster node not starting

Setting up HACMP 6.1 on a two node cluster. The other node works fine and can start properly on STABLE state (VGs varied, FS mounted, Service IP aliased). However, the other node is always stuck on ST_JOINING state. Its taking forever and you can't stop the cluster as well or recover from script... (2 Replies)
Discussion started by: depam
2 Replies
CLUSTER.CONF(5) 						      cluster							   CLUSTER.CONF(5)

NAME
cluster.conf - configuration file for cman and related daemons SYNOPSIS
/etc/cluster/cluster.conf DESCRIPTION
When cman_tool(8) starts the corosync(8) daemon, the cluster.conf data is read into the corosync in-memory database (confdb). The configu- ration is used by corosync, cman and other related cluster daemons and programs. When cman configures corosync with cluster.conf, the corosync.conf(5) file is not used. A basic cluster configuration is described below. Configuration options for other daemons/programs are described in their own man pages. ccs_tool(8) can be used to do some basic cluster.conf editing. The cluster.rng schema is used to validate cluster.conf. Unrecognized items will produce a warning during cluster startup, and invalid xml structure will cause the cluster startup to fail. See ccs_config_validate(8) and ccs_config_dump(8). Cluster The top level cluster section contains all other sections and has two required attributes: name The name of the cluster can be up to 15 characters long (16 including terminating null). It is important that this name be unique among clusters on the same network. config_version The config_version specifies the revision level of the file and should be increased each time the file is updated. <cluster name="alpha" config_version="1"> </cluster> Cluster Nodes The set of nodes that make up the cluster are defined in the clusternodes section which contains multiple clusternode sections. A clus- ternode has two required attributes: name The node name should correspond to the hostname on the network interface to be used for cluster communication. nodeid The node id must be greater than zero and unique. <cluster name="alpha" config_version="1"> <clusternodes> <clusternode name="node-01" nodeid="1"> </clusternode> <clusternode name="node-02" nodeid="2"> </clusternode> <clusternode name="node-03" nodeid="3"> </clusternode> </clusternodes> </cluster> Logging Cluster daemons use a common logging section to configure their loggging behavior. <cluster name="alpha" config_version="1"> <logging/> </cluster> Global settings apply to all: <logging debug="on"/> Per-daemon logging_daemon subsections override the global settings. Daemon names that can be configured include: corosync, qdiskd, groupd, fenced, dlm_controld, gfs_controld, rgmanager. <logging> <logging_daemon name="qdiskd" debug="on"/> <logging_daemon name="fenced" debug="on"/> </logging> Corosync daemon settings apply to all corosync subsystems by default, but subsystems can also be configured individually. These include CLM, CPG, MAIN, SERV, CMAN, TOTEM, QUORUM, CONFDB, CKPT, EVT. <logging> <logging_daemon name="corosync" subsys="QUORUM" debug="on"/> <logging_daemon name="corosync" subsys="CONFDB" debug="on"/> </logging> The attributes available at global, daemon and subsystem levels are: to_syslog enable/disable messages to syslog (yes/no), default "yes" to_logfile enable/disable messages to log file (yes/no), default "yes" syslog_facility facility used for syslog messages, default "daemon" syslog_priority messages at this level and up will be sent to syslog, default "info" logfile_priority messages at this level and up will be written to log file, default "info" logfile the log file name, default /var/log/cluster/<daemon>.log debug="on" a shortcut for logfile_priority="debug" EXAMPLE
An explicit configuration for the default settings would be: <logging to_syslog="yes" to_logfile="yes" syslog_facility="daemon" syslog_priority="info" logfile_priority="info"> <logging_daemon name="qdiskd" logfile="/var/log/cluster/qdiskd.log"/> <logging_daemon name="fenced" logfile="/var/log/cluster/fenced.log"/> <logging_daemon name="dlm_controld" logfile="/var/log/cluster/dlm_controld.log"/> <logging_daemon name="gfs_controld" logfile="/var/log/cluster/gfs_controld.log"/> <logging_daemon name="rgmanager" logfile="/var/log/cluster/rgmanager.log"/> <logging_daemon name="corosync" logfile="/var/log/cluster/corosync.log"/> </logging> To include debug messages (and above) from all daemons in their default log files, either of the following which are equivalent: <logging debug="on"/> <logging logfile_priority="debug"/> To exclude all log messages from syslog: <logging to_syslog="no"/> To disable logging to all log files: <logging to_file="no"/> To include debug messages (and above) from all daemons in syslog: <logging syslog_priority="debug"/> To limit syslog messages to error (and above), keeping info (and above) in log files (this logfile_priority setting is the default so could be omitted): <logging syslog_priority="error" logfile_priority="info"/> FILES
/etc/cluster/cluster.conf standard location of cluster configuration file /usr/share/cluster/cluster.rng standard location of cluster.conf schema SEE ALSO
ccs_tool(8), ccs_config_dump(8), ccs_config_validate(8), cman_tool(8), cman(5), qdisk(5), fenced(8), fence_node(8), dlm_controld(8), gfs_controld(8), rgmanager(8) cluster 2010-01-12 CLUSTER.CONF(5)
All times are GMT -4. The time now is 04:26 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy