02-24-2012
I never worked with Veritas but I have some experience with other clustersoftware.
To me automatic fallback sounds dangerous. When node 1 goes down you most likely have some manual intervention to fix the cause of the failover. If node 1 would go up automatically without anyone checking it there is a significant chance that it will fail again for the same reason it initially failed. This would result in your database filesystem moving back and forth between the nodes...
9 More Discussions You Might Find Interesting
1. UNIX for Advanced & Expert Users
We recently transferred our Samba server to a cluster of two Sun systems running Solaris 9 with Veritas Cluster Server. We used Samba 3.0.7. Since then Samba seems to crash every 2 to 3 days, always around noon. The symptoms are that at first people who are connected lose the ability to access... (1 Reply)
Discussion started by: rhfrommn
1 Replies
2. UNIX for Dummies Questions & Answers
--fail seems to not be working in cURL. When using cURL to download multiple files, --fail is supposed to keep cURL from making files that don't exist, on 404 errors. But when I use --fail or -f, it still makes those files. I've tried it on both Cygwin cURL and Win32 cURL.
Anyone know anything... (2 Replies)
Discussion started by: Pulseczar
2 Replies
3. High Performance Computing
Dear All,
Can anyone explain about Pros and Cons of SUN and Veritas Cluster ?
Any comparison chart is highly appreciated.
Regards,
RAA (4 Replies)
Discussion started by: RAA
4 Replies
4. Solaris
Does Veritas Cluster work with IPMP on Solaris 10?
If anyone has set it up do you have a doc or tips?
I have heard several different statements ranging from , not working at all to Yes it works! Great How?
* Test and Base IPs????
* configure the MultiNICB agent ?
I can give details... (1 Reply)
Discussion started by: dfezz1
1 Replies
5. High Performance Computing
Hi,
Anybody know where I can download Veritas Cluster 5.0 for Solaris x86?
Thanks (2 Replies)
Discussion started by: gwhelan
2 Replies
6. Filesystems, Disks and Memory
After much hitting my head to the wall with this meesages
root@crasher /workzone/dvd4_sfb-sol_x64]#./installer -verbose -installonly
5.0MP3 Installation Program
Copyright (c) 2008 Symantec Corporation. All rights reserved. Symantec, the Symantec... (1 Reply)
Discussion started by: tristezo2k
1 Replies
7. Solaris
Yesterday my customer told me to expect a vcs upgrade to happen in the future. He also plans to stop using HDS and move to EMC.
Am thinking how to migrate to sun cluster setup instead.
My plan as follows leave the existing vcs intact as a fallback plan.
Then install and build suncluster on... (5 Replies)
Discussion started by: sparcguy
5 Replies
8. Red Hat
why does my cluster form but fail after a few minutes, or why do my multicast communications stop working after a short amount of time? (1 Reply)
Discussion started by: gema.utama
1 Replies
9. UNIX for Beginners Questions & Answers
Hi Experts,
I wanted to extend a veritas file system which is running on veritas cluster and mounted on node2 system.
#hastatus -sum
-- System State Frozen
A node1 running 0
A node2 running 0
-- Group State
-- Group System Probed ... (1 Reply)
Discussion started by: Skmanojkum
1 Replies
LEARN ABOUT DEBIAN
lamshrink
LAMSHRINK(1) LAM COMMANDS LAMSHRINK(1)
NAME
lamshrink - Shrink a LAM universe.
SYNOPSIS
lamshrink [-dhv] [-w delay] nodeid
OPTIONS
-d Print detailed debugging information.
-h Print useful information on this command.
-v Be verbose.
-w delay Notify processes on the doomed node and pause for delay seconds before proceeding.
nodeid Remove the LAM node with this ID.
DESCRIPTION
An existing LAM session, initiated by lamboot(1), can be shrunk to include less nodes with lamshrink. One node is removed for each invoca-
tion. At a minimum, the node ID is given on the command line. Once lamshrink completes, the node ID is invalid across the remaining nodes
(as can be seen by running lamnodes(1)).
Existing application processes on the target node can be warned of impending shutdown with the -w option. A LAM signal (SIGFUSE) will be
sent to these processes and lamshrink will then pause for the given number of seconds before proceeding with removing the node. By de-
fault, SIGFUSE is ignored. A different handler can be installed with ksignal(2).
All application processes on all remaining nodes are always informed of the death of a node. This is also done with a signal (SIGSHRINK),
which by default causes a process's runtime route cache to be flushed (to remove any cached information on the dead node). If this signal
is re-vectored for the purpose of fault tolerance, the old handler should be called at the beginning of the new handler. The signal does
not, by itself, give the process information on which node has been removed. One technique for getting this information is to query the
router for information on all relevant nodes using getroute(2). The dead node will cause this routine to return an error.
FAULT TOLERANCE
If enabled with lamboot(1), LAM will watch for nodes that fail. The procedure for removing a node that has failed is the same as lamshrink
after the warning step. In particular, the SIGSHRINK signal is delivered.
EXAMPLES
lamshrink -v n1 Remove LAM on n1. Report about important steps as
they are done.
lamshrink n30 -w 10
Inform all processes on LAM node 30, that the node will be dead in 10 seconds. Wait 10 seconds and remove the node. Operate silently.
SEE ALSO
lamboot(1), lamnodes(1), ksignal(2), getroute(2)
LAM 7.1.4 July, 2007 LAMSHRINK(1)