Sponsored Content
Special Forums UNIX and Linux Applications High Performance Computing sun Cluster resource group cant failover Post 302212983 by incredible on Wednesday 9th of July 2008 01:43:29 AM
Old 07-09-2008
log files - /var/cluster/logs, /var/adm/messages
sccheck logs - /var/cluster/sccheck/report.<date>
CCR files - /etc/cluster/ccr
Cluster Infra file - /etc/cluster/ccr/infrastructure
 

10 More Discussions You Might Find Interesting

1. HP-UX

ServiceGuard cluster & volume group failover

I have a 2-node ServiceGuard cluster. One of the cluster packages has a volume group assigned to it. When I fail the package over to the other node, the volume group does not come up automatically on the other node. I have to manually do a "vgchange -a y vgname" on the node before the package... (5 Replies)
Discussion started by: Wotan31
5 Replies

2. High Performance Computing

Veritas Cluster Server Management Console IP Failover

I have just completed a first RTFM of "Veritas Cluster Server Management Console Implementation Guide" 5.1, with a view to assessing it to possibly make our working lives easier. Unfortunately, at my organisation, getting a test installation would be worse than pulling teeth, so I can't just go... (2 Replies)
Discussion started by: Beast Of Bodmin
2 Replies

3. Solaris

Sun Cluster 3.1 failover

Hi, We have two sun SPARC server in Clustered (Sun Cluster 3.1). For some reason, System 1 failed over to System 2. Where can I find the logs which could tell me the reason for this failover? Thanks (5 Replies)
Discussion started by: Mack1982
5 Replies

4. AIX

Resource Group Monitoring

Hi, I have a requirement to monitor the HACMP Resource Groups. At present in my environment, if the Resource Groups fail over from preferred node to Secondary node we dont get notification. Can some one help me in creating a scrript. I have more than one RG online. (Max 4 Resource Groups in... (2 Replies)
Discussion started by: srnagu
2 Replies

5. Gentoo

How to failover the cluster ?

How to failover the cluster ? GNU/Linux By which command, My Linux version 2008 x86_64 x86_64 x86_64 GNU/Linux What are the prerequisites we need to take while failover ? if any Regards (3 Replies)
Discussion started by: sidharthmellam
3 Replies

6. AIX

Adding a Volume Group to an HACMP Resource Group?

Hi, I have a 2 node Cluster. Which is working in active/passive mode (i.e Node#1 is running and when it goes down the Node#2 takes over) Now there's this requirement that we need a mount point say /test that should be available in active node #1 and when node #1 goes down and node#2 takes... (6 Replies)
Discussion started by: aixromeo
6 Replies

7. Solaris

Sun cluster 4.0 - zone cluster failover doubt

Hello experts - I am planning to install a Sun cluster 4.0 zone cluster fail-over. few basic doubts. (1) Where should i install the cluster s/w binaries ?. ( global zone or the container zone where i am planning to install the zone fail-over) (2) Or should i perform the installation on... (0 Replies)
Discussion started by: NVA
0 Replies

8. Solaris

Solaris Cluster Failover based on scan rate

Dear Experts, If there is a possible Solaris Cluster failover to second node based on scan rate? I need the documentation If solaris cluster can do this. Thank You in Advance Edy (3 Replies)
Discussion started by: edydsuranta
3 Replies

9. Red Hat

Linux Cluster failover issue

Hi Guys, I am not much aware of clusters but i have few questions can someone provide the overview as it would be very helpful for me. How can i perform cluster failover test to see all the services are failing back to other node ? If it is using veritas cluster then what kind of... (2 Replies)
Discussion started by: munna529
2 Replies

10. Solaris

Process to add mount point in Sun Cluster existing HAplus resource

Hi Well I would like to know step by step process of adding a mountpoint in HAPLUS resource in SUN cluster as I go the below command to add a mount point but not the step by step process of adding a mount point in existing HA Plus resource. clrs set -p FileSystemMountPoints+=<new_MP>... (3 Replies)
Discussion started by: amity
3 Replies
CLUSTER.CONF(5)                                                       cluster                                                      CLUSTER.CONF(5)

NAME
cluster.conf - configuration file for cman and related daemons SYNOPSIS
/etc/cluster/cluster.conf DESCRIPTION
When cman_tool(8) starts the corosync(8) daemon, the cluster.conf data is read into the corosync in-memory database (confdb). The configu- ration is used by corosync, cman and other related cluster daemons and programs. When cman configures corosync with cluster.conf, the corosync.conf(5) file is not used. A basic cluster configuration is described below. Configuration options for other daemons/programs are described in their own man pages. ccs_tool(8) can be used to do some basic cluster.conf editing. The cluster.rng schema is used to validate cluster.conf. Unrecognized items will produce a warning during cluster startup, and invalid xml structure will cause the cluster startup to fail. See ccs_config_validate(8) and ccs_config_dump(8). Cluster The top level cluster section contains all other sections and has two required attributes: name The name of the cluster can be up to 15 characters long (16 including terminating null). It is important that this name be unique among clusters on the same network. config_version The config_version specifies the revision level of the file and should be increased each time the file is updated. <cluster name="alpha" config_version="1"> </cluster> Cluster Nodes The set of nodes that make up the cluster are defined in the clusternodes section which contains multiple clusternode sections. A clus- ternode has two required attributes: name The node name should correspond to the hostname on the network interface to be used for cluster communication. nodeid The node id must be greater than zero and unique. <cluster name="alpha" config_version="1"> <clusternodes> <clusternode name="node-01" nodeid="1"> </clusternode> <clusternode name="node-02" nodeid="2"> </clusternode> <clusternode name="node-03" nodeid="3"> </clusternode> </clusternodes> </cluster> Logging Cluster daemons use a common logging section to configure their loggging behavior. <cluster name="alpha" config_version="1"> <logging/> </cluster> Global settings apply to all: <logging debug="on"/> Per-daemon logging_daemon subsections override the global settings. Daemon names that can be configured include: corosync, qdiskd, groupd, fenced, dlm_controld, gfs_controld, rgmanager. <logging> <logging_daemon name="qdiskd" debug="on"/> <logging_daemon name="fenced" debug="on"/> </logging> Corosync daemon settings apply to all corosync subsystems by default, but subsystems can also be configured individually. These include CLM, CPG, MAIN, SERV, CMAN, TOTEM, QUORUM, CONFDB, CKPT, EVT. <logging> <logging_daemon name="corosync" subsys="QUORUM" debug="on"/> <logging_daemon name="corosync" subsys="CONFDB" debug="on"/> </logging> The attributes available at global, daemon and subsystem levels are: to_syslog enable/disable messages to syslog (yes/no), default "yes" to_logfile enable/disable messages to log file (yes/no), default "yes" syslog_facility facility used for syslog messages, default "daemon" syslog_priority messages at this level and up will be sent to syslog, default "info" logfile_priority messages at this level and up will be written to log file, default "info" logfile the log file name, default /var/log/cluster/<daemon>.log debug="on" a shortcut for logfile_priority="debug" EXAMPLE
An explicit configuration for the default settings would be: <logging to_syslog="yes" to_logfile="yes" syslog_facility="daemon" syslog_priority="info" logfile_priority="info"> <logging_daemon name="qdiskd" logfile="/var/log/cluster/qdiskd.log"/> <logging_daemon name="fenced" logfile="/var/log/cluster/fenced.log"/> <logging_daemon name="dlm_controld" logfile="/var/log/cluster/dlm_controld.log"/> <logging_daemon name="gfs_controld" logfile="/var/log/cluster/gfs_controld.log"/> <logging_daemon name="rgmanager" logfile="/var/log/cluster/rgmanager.log"/> <logging_daemon name="corosync" logfile="/var/log/cluster/corosync.log"/> </logging> To include debug messages (and above) from all daemons in their default log files, either of the following which are equivalent: <logging debug="on"/> <logging logfile_priority="debug"/> To exclude all log messages from syslog: <logging to_syslog="no"/> To disable logging to all log files: <logging to_file="no"/> To include debug messages (and above) from all daemons in syslog: <logging syslog_priority="debug"/> To limit syslog messages to error (and above), keeping info (and above) in log files (this logfile_priority setting is the default so could be omitted): <logging syslog_priority="error" logfile_priority="info"/> FILES
/etc/cluster/cluster.conf standard location of cluster configuration file /usr/share/cluster/cluster.rng standard location of cluster.conf schema SEE ALSO
ccs_tool(8), ccs_config_dump(8), ccs_config_validate(8), cman_tool(8), cman(5), qdisk(5), fenced(8), fence_node(8), dlm_controld(8), gfs_controld(8), rgmanager(8) cluster 2010-01-12 CLUSTER.CONF(5)
All times are GMT -4. The time now is 10:07 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy