DLM_TOOL(8) dlm DLM_TOOL(8)NAME
dlm_tool - a utility for the dlm and dlm_controld daemon
SYNOPSIS
dlm_tool [COMMAND] [OPTIONS] [ name ]
COMMANDS
ls
Display dlm_controld internal lockspace state.
status
Dump dlm_controld daemon state.
dump
Dump dlm_controld debug buffer.
dump_config
Dump dlm_controld config settings.
fence_ack nodeid
Quit trying to fence a node.
log_plock
Dump dlm_controld plock debug buffer.
plocks name
Dump posix locks from dlm_controld for the lockspace.
join name
Join a lockspace.
leave name
Leave a lockspace.
lockdebug name
Complete display of locks from the lockspace.
lockdump name
Minimal display of locks from the lockspace (deprecated).
OPTIONS -n Show all node information in ls.
-d 0|1 Resource directory off/on in join, default 0
-e 0|1 Exclusive create off/on in join, default 0
-f 0|1 FS (filesystem) flag off/on in join, default 0
-m mode Permission mode for lockspace device (octal), default 0600
-s Summary following lockdebug output (experiemental)
-v Verbose lockdebug output
-w Wide lockdebug output
-M Include MSTCPY locks in lockdump output
-h Print help, then exit
-V Print program version information, then exit
SEE ALSO dlm_controld(8), dlm.conf(5)dlm 2012-04-05 DLM_TOOL(8)
Check Out this Related Man Page
DLM_CONTROLD(8) cluster DLM_CONTROLD(8)NAME
dlm_controld - daemon that configures dlm according to cluster events
SYNOPSIS
dlm_controld [OPTIONS]
DESCRIPTION
The dlm lives in the kernel, and the cluster infrastructure (corosync membership and group management) lives in user space. The dlm in the
kernel needs to adjust/recover for certain cluster events. It's the job of dlm_controld to receive these events and reconfigure the kernel
dlm as needed. dlm_controld controls and configures the dlm through sysfs and configfs files that are considered dlm-internal interfaces.
The cman init script usually starts the dlm_controld daemon.
OPTIONS
Command line options override a corresponding setting in cluster.conf.
-D Enable debugging to stderr and don't fork.
See also dlm_tool dump in dlm_tool(8).
-L Enable debugging to log file.
See also logging in cluster.conf(5).
-K Enable kernel dlm debugging messages.
See also log_debug below.
-r num dlm kernel lowcomms protocol, 0 tcp, 1 sctp, 2 detect. 2 selects tcp if corosync rrp_mode is "none", otherwise sctp.
Default 2.
-g num groupd compatibility mode, 0 off, 1 on.
Default 0.
-f num Enable (1) or disable (0) fencing recovery dependency.
Default 1.
-q num Enable (1) or disable (0) quorum recovery dependency.
Default 0.
-d num Enable (1) or disable (0) deadlock detection code.
Default 0.
-p num Enable (1) or disable (0) plock code for cluster fs.
Default 1.
-l num Limit the rate of plock operations, 0 for no limit.
Default 0.
-o num Enable (1) or disable (0) plock ownership.
Default 1.
-t ms Plock ownership drop resources time (milliseconds).
Default 10000.
-c num Plock ownership drop resources count.
Default 10.
-a ms Plock ownership drop resources age (milliseconds).
Default 10000.
-P Enable plock debugging messages (can produce excessive output).
-h Print a help message describing available options, then exit.
-V Print program version information, then exit.
FILES cluster.conf(5) is usually located at /etc/cluster/cluster.conf. It is not read directly. Other cluster components load the contents into
memory, and the values are accessed through the libccs library.
Configuration options for dlm (kernel) and dlm_controld are added to the <dlm /> section of cluster.conf, within the top level <cluster>
section.
Kernel options
protocol
The network protocol can be set to tcp, sctp or detect which selects tcp or sctp based on the corosync rrp_mode configuration
(redundant ring protocol). The rrp_mode "none" results in tcp. Default detect.
<dlm protocol="detect"/>
timewarn
After waiting timewarn centiseconds, the dlm will emit a warning via netlink. This only applies to lockspaces created with the
DLM_LSFL_TIMEWARN flag, and is used for deadlock detection. Default 500 (5 seconds).
<dlm timewarn="500"/>
log_debug
DLM kernel debug messages can be enabled by setting log_debug to 1. Default 0.
<dlm log_debug="0"/>
clusternode/weight
The lock directory weight can be specified one the clusternode lines. Weights would usually be used in the lock server configura-
tions shown below instead.
<clusternode name="node01" nodeid="1" weight="1"/>
Daemon options
enable_fencing
See command line description.
<dlm enable_fencing="1"/>
enable_quorum
See command line description.
<dlm enable_quorum="0"/>
enable_deadlk
See command line description.
<dlm enable_deadlk="0"/>
enable_plock
See command line description.
<dlm enable_plock="1"/>
plock_rate_limit
See command line description.
<dlm plock_rate_limit="0"/>
plock_ownership
See command line description.
<dlm plock_ownership="1"/>
drop_resources_time
See command line description.
<dlm drop_resources_time="10000"/>
drop_resources_count
See command line description.
<dlm drop_resources_count="10"/>
drop_resources_age
See command line description.
<dlm drop_resources_age="10000"/>
plock_debug
Enable (1) or disable (0) plock debugging messages (can produce excessive output). Default 0.
<dlm plock_debug="0"/>
Disabling resource directory
Lockspaces usually use a resource directory to keep track of which node is the master of each resource. The dlm can operate without the
resource directory, though, by statically assigning the master of a resource using a hash of the resource name. To enable, set the per-
lockspace nodir option to 1.
<dlm>
<lockspace name="foo" nodir="1">
</dlm>
Lock-server configuration
The nodir setting can be combined with node weights to create a configuration where select node(s) are the master of all resources/locks.
These master nodes can be viewed as "lock servers" for the other nodes.
<dlm>
<lockspace name="foo" nodir="1">
<master name="node01"/>
</lockspace>
</dlm>
or,
<dlm>
<lockspace name="foo" nodir="1">
<master name="node01"/>
<master name="node02"/>
</lockspace>
</dlm>
Lock management will be partitioned among the available masters. There can be any number of masters defined. The designated master nodes
will master all resources/locks (according to the resource name hash). When no masters are members of the lockspace, then the nodes revert
to the common fully-distributed configuration. Recovery is faster, with little disruption, when a non-master node joins/leaves.
There is no special mode in the dlm for this lock server configuration, it's just a natural consequence of combining the "nodir" option
with node weights. When a lockspace has master nodes defined, the master has a default weight of 1 and all non-master nodes have weight of
0. An explicit non-zero weight can also be assigned to master nodes, e.g.
<dlm>
<lockspace name="foo" nodir="1">
<master name="node01" weight="2"/>
<master name="node02" weight="1"/>
</lockspace>
</dlm>
In which case node01 will master 2/3 of the total resources and node2 will master the other 1/3.
SEE ALSO dlm_tool(8), fenced(8), cman(5), cluster.conf(5)cluster 2009-01-18 DLM_CONTROLD(8)