Sponsored Content
Full Discussion: Cluster configuration file
Operating Systems Linux Red Hat Cluster configuration file Post 302771279 by mastansaheb on Wednesday 20th of February 2013 06:47:21 AM
Old 02-20-2013
Cluster configuration file

Hi Everybody,


I would like to know meaning of following parameters in cluster.conf file in linux

Code:
<device name="blade_fence" blade="2" option="off"
<clusternode name="Blade1-int" votes="1" nodeid="1">

what do you mean by "option=off" and "votes=1"

Thanks

Last edited by zaxxon; 02-20-2013 at 08:39 AM..
 

10 More Discussions You Might Find Interesting

1. High Performance Computing

Building a Solaris Cluster Express cluster in a VirtualBox on OpenSolaris

Provides a description of how to set up a Solaris Cluster Express cluster in a VirtualBox on OpenSolaris. More... (0 Replies)
Discussion started by: Linux Bot
0 Replies

2. High Performance Computing

SUN Cluster Vs Veritas Cluster

Dear All, Can anyone explain about Pros and Cons of SUN and Veritas Cluster ? Any comparison chart is highly appreciated. Regards, RAA (4 Replies)
Discussion started by: RAA
4 Replies

3. Solaris

SC3.2 issue - cluster transport configuration not right - resulting fail

I am trying to set up a two host cluster. trouble is with the cluster transport configuration. i'm using e1000g2 and g3 for the cluster transport. global0 and global1 are my two nodes, and I am running the scinstall from global1. i think i should be expecting, is this: The following... (19 Replies)
Discussion started by: frustin
19 Replies

4. Solaris

Sun cluster and Veritas cluster question.

Yesterday my customer told me to expect a vcs upgrade to happen in the future. He also plans to stop using HDS and move to EMC. Am thinking how to migrate to sun cluster setup instead. My plan as follows leave the existing vcs intact as a fallback plan. Then install and build suncluster on... (5 Replies)
Discussion started by: sparcguy
5 Replies

5. AIX

HACMP 5.4.1 Two-Node-Cluster-Configuration-Assistant fails

This post just as a follow-up for thread https://www.unix.com/aix/115548-hacmp-5-4-aix-5300-10-not-working.html: there was a bug in the clcomdES that would cause the Two-Node-Cluster-Configuration-Assistant to fail even with a correct TCP/IP adapter setup. That affected HACMP 5.4.1 in combinatin... (0 Replies)
Discussion started by: shockneck
0 Replies

6. Solaris

Cluster interconnect, adapter configuration

Hello I have problem with scinstall. I found information that i should`t configure public network before using scinstall. Each node configuration: 4 node, 3 network adapter on each node. 1 for public adapter and 2 for cluster interconnect. Two switches. After scinstall first node is... (1 Reply)
Discussion started by: time0ut
1 Replies

7. Solaris

In cluster configuration ora* VGs are not controlled by VCS

Need some one to explain "In cluster configuration ora* VGs are not controlled by VCS". Actually, I am not aware of this. Thanks, Rama (0 Replies)
Discussion started by: ramareddi16
0 Replies

8. Solaris

Sun Cluster configuration issue

I am using VMware Workstation-7 on Windows-XP host . I am trying to configure Solaris 10-X86 guest os based 2 nodes Sun Cluster . I have added one extra Virtual Lan adapter on my VMware with another subnet (that I would like to put for SUN Cluster private communication). I have... (0 Replies)
Discussion started by: sanjee
0 Replies

9. Solaris

Sun cluster 4.0 - zone cluster failover doubt

Hello experts - I am planning to install a Sun cluster 4.0 zone cluster fail-over. few basic doubts. (1) Where should i install the cluster s/w binaries ?. ( global zone or the container zone where i am planning to install the zone fail-over) (2) Or should i perform the installation on... (0 Replies)
Discussion started by: NVA
0 Replies

10. UNIX and Linux Applications

Configuration of Linux cluster managment on Red Hat 5.x server

Hi Experts, I have question regarding linux cluster managment on Red Hat 5.x server. When I try to install 'luci' or 'ricci' in one of our linux servers it is giving me below error:- yum install luci Loaded plugins: katello, product-id, rhnplugin, security, subscription-manager Updating... (0 Replies)
Discussion started by: Amey Joshi
0 Replies
o2cb(7)                                                         OCFS2 Manual Pages                                                         o2cb(7)

NAME
o2cb - Default cluster stack for the OCFS2 file system. DESCRIPTION
o2cb is the default cluster stack for the OCFS2 file system. It includes a node manager (o2nm) to keep track of the nodes in the cluster, a heartbeat agent (o2hb) to detect live nodes, a network agent (o2net) for intra-cluster node communication and a distributed lock manager (o2dlm) to keep track of lock resources. All these components are in-kernel. It also includes an in-memory file system, dlmfs, to allow userspace to access the in-kernel dlm. This cluster stack has two configuration files, namely, /etc/ocfs2/cluster.conf and /etc/sysconfig/o2cb. Whereas the former keeps track of the cluster layout, the latter keeps track of the cluster timeouts. Both files are only read when the cluster is brought online. Values in use by the online cluster can be perused in the /sys/kernel/config/cluster directory structure. CONFIGURATION
The cluster layout is specified in /etc/ocfs2/cluster.conf. While it is easier to populate and propagate this configuration file using ocfs2console(8), one can also do it by manually as long as care is taken to format the file correctly. While the console utility is intuitive to use, there are few points to keep in mind. 1. The node name needs to match the hostname. It does not need to include the domain name. For example, appserver.oracle.com can be appserver. 2. The IP address need not be the one associated with that hostname. As in, any valid IP address on that node can be used. O2CB will not attempt to match the node name (hostname) with the specified IP address. For best performance, use of a private interconnect (lower latency) is recommended. The cluster.conf file is in a stanza format with two types of stanzas, namely, cluster and node. A typical cluster.conf will have one clus- ter stanza and multiple node stanzas. The cluster stanza has two parameters: node_count Total number of nodes in the cluster name Name of the cluster The node stanza has five parameters: ip_port IP port ip_address IP address number Unique node number from 0-254 name Hostname cluster Name of the cluster Users populating cluster.conf manually should follow the format strictly. As in, stanza header should start at the first column and end with a colon, stanza parameters should start after a tab, a blank line should demarcate each stanza and care taken to avoid stray white- spaces. The O2CB cluster timeouts are specified in /etc/sysconfig/o2cb and can be configured using the o2cb init script. These timeouts are used by the O2CB clusterstack to determine whether a node is dead or alive. While the use of default values is recom- mended, users can experiment with other values if the defaults are causing spurious fencing. The cluster timeouts are: Heartbeat Dead Threshold The Disk Heartbeat timeout is the number of two second iterations before a node is considered dead. The exact formula used to con- vert the timeout in seconds to the number of iterations is as follows: O2CB_HEARTBEAT_THRESHOLD = (((timeout in seconds) / 2) + 1) For e.g., to specify a 60 sec timeout, set it to 31. For 120 secs, set it to 61. The default for this timeout is 60 secs (O2CB_HEARTBEAT_THRESHOLD = 31). Network Idle Timeout The Network Idle timeout specifies the time in milliseconds before a network connection is considered dead. It defaults to 30000 ms. Network Keepalive Delay The Network Keepalive specifies the maximum delay in milliseconds before a keepalive packet is sent to another node to check whether it is alive or not. If the node is alive, it will respond. Its defaults to 2000 ms. Network Reconnect Delay The Network Reconnect specifies the minimum delay in milliseconds between connection attempts. It defaults to 2000 ms. EXAMPLES
A sample /etc/ocfs2/cluster.conf. cluster: node_count = 3 name = webcluster node: ip_port = 7777 ip_address = 192.168.0.107 number = 7 name = node7 cluster = webcluster node: ip_port = 7777 ip_address = 192.168.0.106 number = 6 name = node6 cluster = webcluster node: ip_port = 7777 ip_address = 192.168.0.110 number = 10 name = node10 cluster = webcluster SEE ALSO
mkfs.ocfs2(8) fsck.ocfs2(8) tunefs.ocfs2(8) debugfs.ocfs2(8) ocfs2console(8) AUTHORS
Oracle Corporation COPYRIGHT
Copyright (C) 2004, 2010 Oracle. All rights reserved. Version 1.6.4 September 2010 o2cb(7)
All times are GMT -4. The time now is 04:18 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy