Sponsored Content
Operating Systems Solaris What is the procedure to reboot cluster nodes Post 302580838 by cubemonkey on Friday 9th of December 2011 11:49:58 PM
Old 12-10-2011
Every cluster is different .. I've seen some database clusters that the only thing the cluster controls is the filesystems. (Like that's not a disaster waiting to happen.. ). In that particular case, off-lining cluster resources without DBA involvement could make for a bad day Smilie Since it looks like you might not be familiar with the nuances of this cluster, here's what I consider the safe route for DB servers:

Do hastatus -sum, and note the group that controls the database. Then look at
Code:
/etc/VRTSvcs/conf/config/main.cf

and see what that group actually does (or look via the hagrp and hares commands). Assuming the database itself is controlled by the cluster, bring it down like this -

In one window:
Code:
tail /var/VRTSvcs/log/engine_A.log

In another:
Code:
hastop -g <database group> -sys <system its running on>

Watch hastatus, and/or the log file you're tailing. If things go down smoothly, then great. If it hangs up waiting on the DB, let the DBA do their thing. The log will usually tell you everything you need to know. Be patient, depending on the DB, it can take a long time to come down.

Once the cluster and the DBA are both satisfied that the DB is down, you can usually then do a hastop -all, and the cluster should pretty easily take care of the dependencies. Wait for it to complete, and help it along if necessary using the info from the log file you're tailing.

Personally, if I'm not 100% comfortable with the system I'm on, I'm paranoid. In that case I like to do everything with cluster nodes one at a time. So offline all resources on all nodes of the cluster and stop VCS, then shut down one node, then the next, etc. Same on the way up. Bring up nodes one at a time, if you want to be extra careful. Let VCS find its brain on one node before another tries.

I've brought down CFS nodes at the same time, and end up with goofy fencing issues. (in hindsight I should have fully closed out gab and llt). It's never happened when I bring them down one at a time, so if I have the time, I like to do it that way.

Anyway, hastop -all would probably work just fine on a properly configured cluster. The fun is when the cluster isn't properly configured. And unless you know for sure either way, it's best to play it safe.

Last edited by cubemonkey; 12-10-2011 at 12:55 AM..
 

10 More Discussions You Might Find Interesting

1. High Performance Computing

Bonding, IP alias, Virtual IP, 2 nodes cluster

Hi ! I have a simple setup of 2 PC (with linux Red-Hat) where the first PC is the primary machine and the second the backup. I use DRBD for data replication and Red-Hat cluster suite for HA (High Availability). I have tested both. Now I NEED a COMMON IP ADDRESS (or Master/unique IP address) for... (0 Replies)
Discussion started by: Danny Gilbert
0 Replies

2. UNIX for Dummies Questions & Answers

IP Alias, Bonding or Virtual IP, 2 nodes Cluster, which one to use ?

Hi ! I have a simple setup of 2 PC (with linux Red-Hat) where the first PC is the primary machine and the second the backup. I use DRBD for data replication and Red-Hat cluster suite for HA (High Availability). I have tested both. Now I NEED a COMMON IP ADDRESS (or Master/unique IP address) for... (3 Replies)
Discussion started by: Danny Gilbert
3 Replies

3. Emergency UNIX and Linux Support

Rebooting 3 to 1 Cluster nodes.

hello Gurus, My current set up is 3 to 1 Cluster (SUN Cluster 3.2) running oracle database. Task is to reboot the servers. My query is about the procedure to do the same. My understanding is suspend the databases to avoid switchover. Then execute the command scshutdown to down the cluster... (4 Replies)
Discussion started by: EmbedUX
4 Replies

4. Red Hat

Centos/rhel 5 cluster 3 nodes with out Quorum

Hi all, i have 3 nodes cluster (Centos 5 cluster suit) with out quorum disk, node vote = 1, the value of a quorum = 2, when 2 nodes going offline, cluster services are destoys. How i can save the cluster and all services(move all services to one alive node) with out quorum disk when other... (3 Replies)
Discussion started by: Flomaster
3 Replies

5. Solaris

Need advise on setting up solaris 10 2 nodes cluster

I am new to setting up sun solaris 10 cluster, I have 2 sun sparc t3-1 servers (identical), going to use them as web servers (sun one java web server 7), looking for data replication and real time fail over. My question is do I need external storage to configure the cluster? or I can just use... (3 Replies)
Discussion started by: spitfire2011
3 Replies

6. Red Hat

How to troubleshoot a 1000 nodes Apache cluster?

Hi all. May I get some expert advice on troubleshooting performance issues of a 1000 nodes Apache LB cluster. Users report slow loading/response of webpages. Different websites are hosted on this cluster for different clients. But all are reporting the same issue. Could you please let me know... (1 Reply)
Discussion started by: admin_xor
1 Replies

7. AIX

Re-cluster 2 HACMP 5.2 nodes

Hi, A customer I'm supporting once upon a time broke their 2 cluster node database servers so they could use the 2nd standby node for something else. Now sometime later they want to bring the 2nd node back into the cluster for resilance. Problem is there are now 3 VG's that have been set-up... (1 Reply)
Discussion started by: elcounto
1 Replies

8. UNIX for Advanced & Expert Users

Arbitrator for 2 nodes ocfs cluster

Is there any way to create a arbitrary node for ocfs2 on a virtual machine (others are physical servers) so it won't go panic when one of physical server goes down? This is for load balanced application servers. Any setting example or tips? Thanks. (0 Replies)
Discussion started by: malayo
0 Replies

9. Red Hat

RedHat Cluster: Nodes won't see each other

Hi All; I try to build a Redhat Cluster (CentOS 6) on vmware. But each node sees the other down like: # clustat Cluster Status for mycluster @ Wed Apr 8 11:01:38 2015 Member Status: Quorate Member Name ID Status ------ ---- ... (0 Replies)
Discussion started by: Meacham12
0 Replies

10. Red Hat

RedHat Cluster: Nodes won't see each other

Hi All; I try to build a Redhat Cluster (CentOS 6) on vmware. But each node sees the other down like: # clustat Cluster Status for mycluster @ Wed Apr 8 11:01:38 2015 Member Status: Quorate Member Name ID Status ------ ---- ... (1 Reply)
Discussion started by: Meacham12
1 Replies
CRM_MON(8)							  [FIXME: manual]							CRM_MON(8)

NAME
crm_mon - monitor the cluster's status SYNOPSIS
crm_mon [-V] -d -pfilename -h filename crm_mon [-V] [-1|-n|-r] -h filename crm_mon [-V] [-n|-r] -X filename crm_mon [-V] [-n|-r] -c|-1 crm_mon [-V] -i interval crm_mon -? DESCRIPTION
The crm_mon command allows you to monitor your cluster's status and configuration. Its output includes the number of nodes, uname, uuid, status, the resources configured in your cluster, and the current status of each. The output of crm_mon can be displayed at the console or printed into an HTML file. When provided with a cluster configuration file without the status section, crm_mon creates an overview of nodes and resources as specified in the file. OPTIONS
--help, -? Provide help. --verbose, -V Increase the debug output. --interval seconds, -i seconds Determine the update frequency. If -i is not specified, the default of 15 seconds is assumed. --group-by-node, -n Group resources by node. --inactive, -r Display inactive resources. --as-console, -c Display the cluster status on the console. --one-shot, -1 Display the cluster status once on the console then exit (does not use ncurses). --as-html filename, -h filename Write the cluster's status to the specified file. --daemonize, -d Run in the background as a daemon. --pid-file filename, -p filename Specify the daemon's pid file. --xml-file filename, -X filename Specify an XML file containing a cluster configuration and create an overview of the cluster's configuration. EXAMPLES
Display your cluster's status and get an updated listing every 15 seconds: crm_mon Display your cluster's status and get an updated listing after an interval specified by -i. If -i is not given, the default refresh interval of 15 seconds is assumed: crm_mon -i interval[s] Display your cluster's status on the console: crm_mon -c Display your cluster's status on the console just once then exit: crm_mon -1 Display your cluster's status and group resources by node: crm_mon -n Display your cluster's status, group resources by node, and include inactive resources in the list: crm_mon -n -r Write your cluster's status to an HTML file: crm_mon -h filename Run crm_mon as a daemon in the background, specify the daemon's pid file for easier control of the daemon process, and create HTML output. This option allows you to constantly create HTML output that can be easily processed by other monitoring applications: crm_mon -d -p filename -h filename Display the cluster configuration laid out in an existing cluster configuration file (filename), group the resources by node, and include inactive resources. This command can be used for dry-runs of a cluster configuration before rolling it out to a live cluster. crm_mon -r -n -X filename FILES
/var/lib/heartbeat/crm/cib.xml--the CIB (minus status section) on disk. Editing this file directly is strongly discouraged. AUTHOR
crm_mon was written by Andrew Beekhof. [FIXME: source] 07/05/2010 CRM_MON(8)
All times are GMT -4. The time now is 11:05 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy