Sponsored Content
Full Discussion: Sun Cluster 3.1 failover
Operating Systems Solaris Sun Cluster 3.1 failover Post 302457049 by Mack1982 on Monday 27th of September 2010 07:11:23 AM
Old 09-27-2010
Well apparently, the Sun Cluster was fine. The issue was only at RAC level....

Oracle:
Quote:
Most cluster error messages are logged by the syslog mechanism to files specified in /etc/syslog.conf - in this case of both nodes this is /var/adm/messages.

Other log files and data services are described in SunSolve Document ID: 1009248.1 (attached).

Since the release of Sun Cluster 3.1 Update 1, Sun Cluster provides a cluster event log daemon, cl_eventlogd, which runs on all the nodes of the cluster and logs cluster events to the /var/cluster/logs/eventlog file.

Whilst the eventlog is 0 bytes on both nodes the daemon is running on both nodes, so I see no reason to suspect they are not logging properly, just that there are no events to log.

ps-ef.out: /usr/cluster/lib/sc/cl_eventlogd


ps-ef.out: /usr/cluster/lib/sc/cl_eventlogd

To continue with the investigations I need you to tell me what the problem was that prompted you to log the call.
 

10 More Discussions You Might Find Interesting

1. High Performance Computing

sun Cluster resource group cant failover

I have rcently setup a 4 node cluster running sun cluster 3.2 and I have installed 4 zones on each node. when installing the zones I had to install the zone on all nodes the on the last node do a zlogin -C <zonename> this worked ok. theni I tried to siwitch the zone to node a thei work... (14 Replies)
Discussion started by: lesliek
14 Replies

2. HP-UX

ServiceGuard cluster & volume group failover

I have a 2-node ServiceGuard cluster. One of the cluster packages has a volume group assigned to it. When I fail the package over to the other node, the volume group does not come up automatically on the other node. I have to manually do a "vgchange -a y vgname" on the node before the package... (5 Replies)
Discussion started by: Wotan31
5 Replies

3. High Performance Computing

SUN Cluster Vs Veritas Cluster

Dear All, Can anyone explain about Pros and Cons of SUN and Veritas Cluster ? Any comparison chart is highly appreciated. Regards, RAA (4 Replies)
Discussion started by: RAA
4 Replies

4. High Performance Computing

Veritas Cluster Server Management Console IP Failover

I have just completed a first RTFM of "Veritas Cluster Server Management Console Implementation Guide" 5.1, with a view to assessing it to possibly make our working lives easier. Unfortunately, at my organisation, getting a test installation would be worse than pulling teeth, so I can't just go... (2 Replies)
Discussion started by: Beast Of Bodmin
2 Replies

5. Solaris

Sun cluster and Veritas cluster question.

Yesterday my customer told me to expect a vcs upgrade to happen in the future. He also plans to stop using HDS and move to EMC. Am thinking how to migrate to sun cluster setup instead. My plan as follows leave the existing vcs intact as a fallback plan. Then install and build suncluster on... (5 Replies)
Discussion started by: sparcguy
5 Replies

6. Gentoo

How to failover the cluster ?

How to failover the cluster ? GNU/Linux By which command, My Linux version 2008 x86_64 x86_64 x86_64 GNU/Linux What are the prerequisites we need to take while failover ? if any Regards (3 Replies)
Discussion started by: sidharthmellam
3 Replies

7. Solaris

Sun cluster 4.0 - zone cluster failover doubt

Hello experts - I am planning to install a Sun cluster 4.0 zone cluster fail-over. few basic doubts. (1) Where should i install the cluster s/w binaries ?. ( global zone or the container zone where i am planning to install the zone fail-over) (2) Or should i perform the installation on... (0 Replies)
Discussion started by: NVA
0 Replies

8. Solaris

Solaris Cluster Failover based on scan rate

Dear Experts, If there is a possible Solaris Cluster failover to second node based on scan rate? I need the documentation If solaris cluster can do this. Thank You in Advance Edy (3 Replies)
Discussion started by: edydsuranta
3 Replies

9. Red Hat

Linux Cluster failover issue

Hi Guys, I am not much aware of clusters but i have few questions can someone provide the overview as it would be very helpful for me. How can i perform cluster failover test to see all the services are failing back to other node ? If it is using veritas cluster then what kind of... (2 Replies)
Discussion started by: munna529
2 Replies

10. Solaris

Can we add Oracle ASM in sun cluster3.3 or 4.0 in failover mode

Hi I am new to this forum & oracle DBA also, I would like to know that can we add Oracle ASM in failover mode in sun cluster 3.3 or 4.0 means that if suppose oracle is running along with ASM on node1 & this node went down due to hardware issue then both oracle along with ASM must move to... (1 Reply)
Discussion started by: hb00
1 Replies
CLUSTER.CONF(5) 						      cluster							   CLUSTER.CONF(5)

NAME
cluster.conf - configuration file for cman and related daemons SYNOPSIS
/etc/cluster/cluster.conf DESCRIPTION
When cman_tool(8) starts the corosync(8) daemon, the cluster.conf data is read into the corosync in-memory database (confdb). The configu- ration is used by corosync, cman and other related cluster daemons and programs. When cman configures corosync with cluster.conf, the corosync.conf(5) file is not used. A basic cluster configuration is described below. Configuration options for other daemons/programs are described in their own man pages. ccs_tool(8) can be used to do some basic cluster.conf editing. The cluster.rng schema is used to validate cluster.conf. Unrecognized items will produce a warning during cluster startup, and invalid xml structure will cause the cluster startup to fail. See ccs_config_validate(8) and ccs_config_dump(8). Cluster The top level cluster section contains all other sections and has two required attributes: name The name of the cluster can be up to 15 characters long (16 including terminating null). It is important that this name be unique among clusters on the same network. config_version The config_version specifies the revision level of the file and should be increased each time the file is updated. <cluster name="alpha" config_version="1"> </cluster> Cluster Nodes The set of nodes that make up the cluster are defined in the clusternodes section which contains multiple clusternode sections. A clus- ternode has two required attributes: name The node name should correspond to the hostname on the network interface to be used for cluster communication. nodeid The node id must be greater than zero and unique. <cluster name="alpha" config_version="1"> <clusternodes> <clusternode name="node-01" nodeid="1"> </clusternode> <clusternode name="node-02" nodeid="2"> </clusternode> <clusternode name="node-03" nodeid="3"> </clusternode> </clusternodes> </cluster> Logging Cluster daemons use a common logging section to configure their loggging behavior. <cluster name="alpha" config_version="1"> <logging/> </cluster> Global settings apply to all: <logging debug="on"/> Per-daemon logging_daemon subsections override the global settings. Daemon names that can be configured include: corosync, qdiskd, groupd, fenced, dlm_controld, gfs_controld, rgmanager. <logging> <logging_daemon name="qdiskd" debug="on"/> <logging_daemon name="fenced" debug="on"/> </logging> Corosync daemon settings apply to all corosync subsystems by default, but subsystems can also be configured individually. These include CLM, CPG, MAIN, SERV, CMAN, TOTEM, QUORUM, CONFDB, CKPT, EVT. <logging> <logging_daemon name="corosync" subsys="QUORUM" debug="on"/> <logging_daemon name="corosync" subsys="CONFDB" debug="on"/> </logging> The attributes available at global, daemon and subsystem levels are: to_syslog enable/disable messages to syslog (yes/no), default "yes" to_logfile enable/disable messages to log file (yes/no), default "yes" syslog_facility facility used for syslog messages, default "daemon" syslog_priority messages at this level and up will be sent to syslog, default "info" logfile_priority messages at this level and up will be written to log file, default "info" logfile the log file name, default /var/log/cluster/<daemon>.log debug="on" a shortcut for logfile_priority="debug" EXAMPLE
An explicit configuration for the default settings would be: <logging to_syslog="yes" to_logfile="yes" syslog_facility="daemon" syslog_priority="info" logfile_priority="info"> <logging_daemon name="qdiskd" logfile="/var/log/cluster/qdiskd.log"/> <logging_daemon name="fenced" logfile="/var/log/cluster/fenced.log"/> <logging_daemon name="dlm_controld" logfile="/var/log/cluster/dlm_controld.log"/> <logging_daemon name="gfs_controld" logfile="/var/log/cluster/gfs_controld.log"/> <logging_daemon name="rgmanager" logfile="/var/log/cluster/rgmanager.log"/> <logging_daemon name="corosync" logfile="/var/log/cluster/corosync.log"/> </logging> To include debug messages (and above) from all daemons in their default log files, either of the following which are equivalent: <logging debug="on"/> <logging logfile_priority="debug"/> To exclude all log messages from syslog: <logging to_syslog="no"/> To disable logging to all log files: <logging to_file="no"/> To include debug messages (and above) from all daemons in syslog: <logging syslog_priority="debug"/> To limit syslog messages to error (and above), keeping info (and above) in log files (this logfile_priority setting is the default so could be omitted): <logging syslog_priority="error" logfile_priority="info"/> FILES
/etc/cluster/cluster.conf standard location of cluster configuration file /usr/share/cluster/cluster.rng standard location of cluster.conf schema SEE ALSO
ccs_tool(8), ccs_config_dump(8), ccs_config_validate(8), cman_tool(8), cman(5), qdisk(5), fenced(8), fence_node(8), dlm_controld(8), gfs_controld(8), rgmanager(8) cluster 2010-01-12 CLUSTER.CONF(5)
All times are GMT -4. The time now is 04:18 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy