Sponsored Content
Homework and Emergencies Homework & Coursework Questions NFS service not started in Redhat Cluster Post 302954597 by nitinredekar on Wednesday 9th of September 2015 08:40:39 AM
Old 09-09-2015
NFS service not started in Redhat Cluster

Hi All,

Need your help to resolve below error in cluter.

Code:
Sep  9 05:37:30 node2 rgmanager[2804]: Starting disabled service service:Prod_Service
Sep  9 05:37:30 node2 rgmanager[7545]: [lvm] HA LVM:  Improper setup detected
Sep  9 05:37:30 node2 rgmanager[7567]: [lvm] * initrd image needs to be newer than lvm.conf
Sep  9 05:37:30 node2 rgmanager[7607]: [lvm] Activating vg_prod-nfs/lv_Prod-nfs
Sep  9 05:37:30 node2 rgmanager[7637]: [lvm] Making resilient : lvchange -ay vg_prod-nfs/lv_Prod-nfs
Sep  9 05:37:30 node2 rgmanager[7662]: [lvm] Resilient command: lvchange -ay vg_prod-nfs/lv_Prod-nfs --config devices{filter=["a|/dev/sdb|","r|.*|"]}
Sep  9 05:37:31 node2 rgmanager[7738]: [fs] mounting /dev/dm-0 on /PROD-HOME
Sep  9 05:37:31 node2 kernel: EXT4-fs (dm-0): mounted filesystem with ordered data mode. Opts: 
Sep  9 05:37:31 node2 rgmanager[7822]: [nfsclient] No export path specified.
Sep  9 05:37:31 node2 rgmanager[2804]: start on nfsclient "Prod-client" returned 2 (invalid argument(s))
Sep  9 05:37:31 node2 rgmanager[2804]: #68: Failed to start service:Prod_Service; return value: 1
Sep  9 05:37:31 node2 rgmanager[2804]: Stopping service service:Prod_Service
Sep  9 05:37:31 node2 rgmanager[7892]: [nfsclient] No export path specified.
Sep  9 05:37:31 node2 rgmanager[2804]: stop on nfsclient "Prod-client" returned 2 (invalid argument(s))
Sep  9 05:37:31 node2 rgmanager[7962]: [fs] unmounting /PROD-HOME
Sep  9 05:37:31 node2 rgmanager[8017]: [lvm] HA LVM:  Improper setup detected
Sep  9 05:37:31 node2 rgmanager[8039]: [lvm] * initrd image needs to be newer than lvm.conf
Sep  9 05:37:31 node2 rgmanager[8072]: [lvm] Deactivating vg_prod-nfs/lv_Prod-nfs
Sep  9 05:37:31 node2 rgmanager[8094]: [lvm] Making resilient : lvchange -an vg_prod-nfs/lv_Prod-nfs
Sep  9 05:37:31 node2 rgmanager[8119]: [lvm] Resilient command: lvchange -an vg_prod-nfs/lv_Prod-nfs --config devices{filter=["a|/dev/sdb|","r|.*|"]}
Sep  9 05:37:31 node2 rgmanager[8149]: [lvm] Removing ownership tag (node2.xyz.com) from vg_prod-nfs/lv_Prod-nfs
Sep  9 05:37:31 node2 rgmanager[2804]: #12: RG service:Prod_Service failed to stop; intervention required
Sep  9 05:37:31 node2 rgmanager[2804]: Service service:Prod_Service is failed
Sep  9 05:37:31 node2 rgmanager[2804]: #13: Service service:Prod_Service failed to stop cleanly

I am not sure where to define nfs path.

This setup i am trying on Centos 6.5.

Last edited by nitinredekar; 09-10-2015 at 10:19 AM..
 

10 More Discussions You Might Find Interesting

1. Red Hat

Unable to start cluster service in redhat 5

Hi, I am trying to setup redhat native cluster 5, i installed all the required rpms and issued the command service cman start (cman is the cluster manager) this gives me output... #service cman start Starting cluster: Loading modules... done Mounting configfs... done ... (0 Replies)
Discussion started by: santhoshb
0 Replies

2. Red Hat

Redhat Cluster Fencing failed

Hello; I have 2 node Redhat Cluster (RHEL4 U4 and Cluster Suite) I'm using mc_data fiber channel switch for fencing when I want to fence manually using fence_mcdata -a x.x.x. -l xxx -p xxxx -n 5 -o disable following messages appears fencing node "test1" agent "fence_mcdata" reports:... (0 Replies)
Discussion started by: sakir19
0 Replies

3. High Performance Computing

cluster setup and nfs

hello, During the cluster installation, we are required to enter a cluster name. When this name is used? I have a shared disk between those 2 nodes. To test NFS failover, what should I do? PS: I am using VCS 5 thanks (5 Replies)
Discussion started by: melanie_pfefer
5 Replies

4. Solaris

All dependicies of nfs service is online but nfs is offline

Hi all in my server all nfs dependices are online but nfs client is offline root@BIWAPP1 # svcs -a|grep nfs disabled Sep_05 svc:/network/nfs/server:default online Sep_05 svc:/network/nfs/rquota:default online Sep_05 svc:/network/nfs/mapid:default online ... (9 Replies)
Discussion started by: spandhan
9 Replies

5. Red Hat

Samba smb not started on Redhat Enterprise

Hi i installed samba-client-3.0.23c-2 on Redhat Server but cat not start #service smb start. it gives error unrecognized smb I checked samba-common-3.0.23c-2 # rpm -qa |grep smb pam_smb-1.1.7-7.2.1 # rpm -qa |grep samba samba3-client-3.2.15-40.el5 Could someone help please? ... (1 Reply)
Discussion started by: lamoul
1 Replies

6. Red Hat

Redhat cluster

Hi Friends, I am trying to configure redhat high availability cluster on rhel5 32bit. As i am not having the RHN subscription, I have to intsall all the cluster suite manually. I am not aware what all are the rpms needs to be installed. Please let me know about the rpms needs to be installed.... (3 Replies)
Discussion started by: arumon
3 Replies

7. Red Hat

Redhat cluster error

hi all , I'm trying for the Red-hat cluster setup with the ' Enable shared storage support'. But hang up with a error that "Packages of set "Clustered Storage" are not present in any available repository". I made this setup by installing the rpm manually..I ve attached the rpm list and error... (0 Replies)
Discussion started by: sriniv666
0 Replies

8. Red Hat

Redhat 6.2 cluster

Hello Experts, Please help me with the exact steps on configuring two node cluster on RHEL 6.2, I failed to configure the simplest cluster by below steps, 1- install RHEL 6.2 64-bit on both nodes 2- add to hosts file of each server ( 1 IP in local NW and another in IP private NW). x.x.x.x ... (1 Reply)
Discussion started by: karmellove
1 Replies

9. Red Hat

Redhat cluster

HI, Can you guys help me to install redhat cluster on VMware Workstation. How to download cluster suite and how to configure on Workstation. (1 Reply)
Discussion started by: karthik9358
1 Replies

10. Red Hat

Query Regarding Redhat Cluster ?? Please help me

:b::b:Hello Expert, I have a query regarding redhat cluster configuration file (/etc/cluster/cluster.conf). Is there any process to change the location of the configuration file of the cluster. Is there any step in ricci process.. If yes then please explain me how the cluster node will... (0 Replies)
Discussion started by: Pradipta Kumar
0 Replies
FENCE_SANLOCK(8)					      System Manager's Manual						  FENCE_SANLOCK(8)

NAME
fence_sanlock - fence agent using watchdog and shared storage leases SYNOPSIS
fence_sanlock [OPTIONS] DESCRIPTION
fence_sanlock uses the watchdog device to reset nodes, in conjunction with three daemons: fence_sanlockd, sanlock, and wdmd. The watchdog device, controlled through /dev/watchdog, is available when a watchdog kernel module is loaded. A module should be loaded for the available hardware. If no hardware watchdog is available, or no module is loaded, the "softdog" module will be loaded, which emulates a hardware watchdog device. Shared storage must be configured for sanlock to use from all hosts. This is generally an lvm lv (non-clustered), but could be another block device, or NFS file. The storage should be 1GB of fully allocated space. After being created, the storage must be initialized with the command: # fence_sanlock -o sanlock_init -p /path/to/storage The fence_sanlock agent uses sanlock leases on shared storage to verify that hosts have been reset, and to notify fenced nodes that are still running, that they should be reset. The fence_sanlockd init script starts the wdmd, sanlock and fence_sanlockd daemons before the cluster or fencing systems are started (e.g. cman, corosync and fenced). The fence_sanlockd daemon is started with the -w option so it waits for the path and host_id options to be provided when they are available. Unfencing must be configured for fence_sanlock in cluster.conf. The cman init script does unfencing by running fence_node -U, which in turn runs fence_sanlock with the "on" action and local path and host_id values taken from cluster.conf. fence_sanlock in turn passes the path and host_id values to the waiting fence_sanlockd daemon. With these values, fence_sanlockd joins the sanlock lockspace and acquires a resource lease for the local host. It can take several minutes to complete these unfencing steps. Once unfencing is complete, the node is a member of the sanlock lockspace named "fence" and the node's fence_sanlockd process holds a resource lease named "hN", where N is the node's host_id. (To verify this, run the commands "sanlock client status" and "sanlock client host_status", which show state from the sanlock daemon, or "sanlock direct dump <path>" which shows state from shared storage.) When fence_sanlock fences a node, it tries to acquire that node's resource lease. sanlock will not grant the lease until the owner (the node being fenced) has been reset by its watchdog device. The time it takes to acquire the lease is 140 seconds from the victim's last lockspace renewal timestamp on the shared storage. Once acquired, the victim's lease is released, and fencing completes successfully. Live nodes being fenced When a live node is being fenced, fence_sanlock will continually fail to acquire the victim's lease, because the victim continues to renew its lockspace membership on storage, and the fencing node sees it is alive. This is by design. As long as the victim is alive, it must continue to renew its lockspace membership on storage. The victim must not allow the remote fence_sanlock to acquire its lease and con- sider it fenced while it is still alive. At the same time, a victim knows that when it is being fenced, it should be reset to avoid blocking recovery of the rest of the cluster. To communicate this, fence_sanlock makes a "request" on storage for the victim's resource lease. On the victim, fence_sanlockd, which holds the resource lease, is configured to receive SIGUSR1 from sanlock if anyone requests its lease. Upon receiving the signal, fence_sanlockd knows that it is a fencing victim. In response to this, fence_sanlockd allows its wdmd connection to expire, which in turn causes the watchdog device to fire, resetting the node. The watchdog reset will obviously have the effect of stopping the victim's lockspace membership renewals. Once the renewals stop, fence_sanlock will finally be able to acquire the victim's lease after waiting a fixed time from the final lockspace renewal. Loss of shared storage If access to shared storage with sanlock leases is lost for 80 seconds, sanlock is not able to renew the lockspace membership, and enters recovery. This causes sanlock clients holding leases, such as fence_sanlockd, to be notified that their leases are being lost. In response, fence_sanlockd must reset the node, much as if it was being fenced. Daemons killed/crashed/hung If sanlock, fence_sanlockd daemons are killed abnormally, or crash or hang, their wdmd connections will expire, causing the watchdog device to fire, resetting the node. fence_sanlock from another node will then run and acquire the victim's resource lease. If the wdmd daemon is killed abnormally or crashes or hangs, it will not pet the watchdog device, causing it to fire and reset the node. Time Values The specific times periods referenced above, e.g. 140, 80, are based on the default sanlock i/o timeout of 10 seconds. If sanlock is con- figured to use a different i/o timeout, these numbers will be different. OPTIONS
-o action The agent action: on Enable the local node to be fenced. Used by unfencing. off Disable another node. status Test if a node is on or off. A node is on if it's lease is held, and off is it's lease is free. metadata Print xml description of required parameters. sanlock_init Initialize sanlock leases on shared storage. -p path The path to shared storage with sanlock leases. -i host_id The host_id, from 1-128. STDIN PARAMETERS
Options can be passed on stdin, with the format key=val. Each key=val pair is separated by a new line. action=on|off|status See -o path=/path/to/shared/storage See -p host_id=num See -i FILES
Example cluster.conf configuration for fence_sanlock. (For cman based clusters in which fenced runs agents.) Also see cluster.conf(5), fenced(8), fence_node(8). <clusternode name="node01" nodeid="1"> <fence> <method name="1"> <device name="wd" host_id="1"/> </method> </fence> <unfence> <device name="wd" host_id="1" action="on"/> </unfence> </clusternode> <clusternode name="node02" nodeid="2"> <fence> <method name="1"> <device name="wd" host_id="2"/> </method> </fence> <unfence> <device name="wd" host_id="2" action="on"/> </unfence> </clusternode> <fencedevice name="wd" agent="fence_sanlock" path="/dev/fence/leases"/> Example dlm.conf configuration for fence_sanlock. (For non-cman based clusters in which dlm_controld runs agents.) Also see dlm.conf(5), dlm_controld(8). device wd /usr/sbin/fence_sanlock path=/dev/fence/leases connect wd node=1 host_id=1 connect wd node=2 host_id=2 unfence wd TEST
To test fence_sanlock directly, without clustering: 1. Initialize storage node1: create 1G lv on shared storage /dev/fence/leases node1: fence_sanlock -o sanlock_init -p /dev/fence/leases 2. Start services node1: service fence_sanlockd start node2: service fence_sanlockd start 3. Enable fencing node1: fence_sanlock -o on -p /dev/fence/leases -i 1 node2: fence_sanlock -o on -p /dev/fence/leases -i 2 This "unfence" step may take a couple minutes. 4. Verify hosts and leases node1: sanlock status s fence:1:/dev/fence/leases:0 r fence:h1:/dev/fence/leases:1048576:1 p 2465 node2: sanlock status s fence:2:/dev/fence/leases:0 r fence:h2:/dev/fence/leases:2097152:1 p 2366 node1: sanlock host_status lockspace fence 1 timestamp 717 2 timestamp 678 node2: sanlock host_status lockspace fence 1 timestamp 738 2 timestamp 678 5. Fence node2 node1: fence_sanlock -o off -p /dev/fence/leases -i 2 This may take a few minutes to return. When node2 is not dead before fencing, sanlock on node1 will log errors about failing to acquire the lease while node2 is still alive. This is expected. 6. Success node1 fence_sanlock should exit 0 after node2 is reset by its watchdog. SEE ALSO
fence_sanlockd(8), sanlock(8), wdmd(8) 2013-05-02 FENCE_SANLOCK(8)
All times are GMT -4. The time now is 02:18 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy