Sponsored Content
Operating Systems Solaris Sun cluster v3.2 - Is this right? Post 303020925 by psychocandy on Tuesday 31st of July 2018 09:17:54 AM
Old 07-31-2018
Sun cluster v3.2 - Is this right?

Running Sun Cluster v3.2 it appears. Two clustered physcial servers both running Solaris 10.

Both servers run a number of Oracle DBs etc. BUT I'm a bit concerned that its been set up but will never switch in the even of failure of one of the hosts?

Some of the cluster groups we've disabled (migrated to new servers). BUT noticed that only CHD-RG and WCKNO-RG have both servers listed?
Are both nodes supposed to be configured for these cluster groups? If, for example, server DAT-PL2 failed, then COGNOS10-RG would not switch because DAT-SL2 is not listed?

Config is as follows:-

Code:
# clrg status
 
=== Cluster Resource Groups ===
 
Group Name           Node Name     Suspended    Status
----------           ---------     ---------    ------
listener-p-rg        dat-pl2    No           Online
 
listener-s-rg        dat-sl2    No           Online
 
chprd-rg             dat-pl2    No           Online
 
vses-rg              dat-pl2    No           Online
 
vprd-rg              dat-sl2    No           Online
 
wckprd-rg            dat-sl2    No           Online
 
listener2-p-rg       dat-pl2    No           Online
 
listener2-s-rg       dat-sl2    No           Online
 
wckno-rg             dat-pl2    No           Offline
                          dat-sl2    No           Online
 
listenerWCK-p-rg     dat-pl2    No           Online
 
listenerEWF-p-rg     dat-pl2    No           Online
 
listenerEWF-s-rg     dat-sl2    No           Online
 
listenerWCK-s-rg     dat-sl2    No           Online
 
cognos10-rg          dat-pl2    No           Online
 
listenerXML-s-rg     dat-sl2    No           Online
 
listenerXML-p-rg     dat-pl2    No           Online
 
listenerFES-p-rg     dat-pl2    No           Online
 
listenerFES-s-rg     dat-sl2    No           Online
 
e5fin-rg             dat-sl2    No           Unmanaged
 
e5archive-rg         dat-sl2    No           Unmanaged
 
fes-rg               dat-sl2    No           Unmanaged
 
chd-rg               dat-pl2    No           Unmanaged
                     dat-sl2    No           Unmanaged


Last edited by Corona688; 07-31-2018 at 12:22 PM.. Reason: code tags
 

9 More Discussions You Might Find Interesting

1. High Performance Computing

SUN Cluster Vs Veritas Cluster

Dear All, Can anyone explain about Pros and Cons of SUN and Veritas Cluster ? Any comparison chart is highly appreciated. Regards, RAA (4 Replies)
Discussion started by: RAA
4 Replies

2. Solaris

sun cluster

hi can u please send me how to configure cluster in x86 machines. i am using VM ware. Is there any difference in X86 and SPARC cluster installations please help me thaks to all (1 Reply)
Discussion started by: sijocg
1 Replies

3. Solaris

Sun Cluster 3.2

Hi, I am getting these messages constantly at random times there is not visible impact but would like to know what these messages mean? Cluster.RGM.rgmd: resource group on node change to RG_ON_PENDING_R_RERT Cluster.RGM.rgmd: resource group on node change to RG_ONLINE (1 Reply)
Discussion started by: xor
1 Replies

4. Solaris

Sun Cluster

Hi All I hav working knowledge on solarois 10. Could anyone suggest me any doc or material for learning SUN cluster from scratch. Anything from basic level . Thanks in advance!! (1 Reply)
Discussion started by: ningy
1 Replies

5. Solaris

Sun cluster and Veritas cluster question.

Yesterday my customer told me to expect a vcs upgrade to happen in the future. He also plans to stop using HDS and move to EMC. Am thinking how to migrate to sun cluster setup instead. My plan as follows leave the existing vcs intact as a fallback plan. Then install and build suncluster on... (5 Replies)
Discussion started by: sparcguy
5 Replies

6. Solaris

Sun cluster 3.2

Hi I have a new mount point which is to added to to an existing resource group. I read many docs but I am not able to find the exact methods. Could anyone help me pls. (1 Reply)
Discussion started by: ningy
1 Replies

7. Solaris

Sun cluster 3.1

Hi, I have upgraded a Sparc T2000 server which is node 2, in a Sun Cluster 3.1 two node cluster from Solaris 10 Update 2, to 10 Update 7. This is a requirement for a NetApps solution as we currently have a Sun 3510 SAN. I am at a stage where I believe the two nodes will not communicate over the... (0 Replies)
Discussion started by: zetex
0 Replies

8. Solaris

Sun cluster

Any body can tell me where i can download sun cluster installation and configuration video. is there any option to install cluster in 32 bit virtual machine. please help (4 Replies)
Discussion started by: sunnybee
4 Replies

9. Solaris

Sun cluster 4.0 - zone cluster failover doubt

Hello experts - I am planning to install a Sun cluster 4.0 zone cluster fail-over. few basic doubts. (1) Where should i install the cluster s/w binaries ?. ( global zone or the container zone where i am planning to install the zone fail-over) (2) Or should i perform the installation on... (0 Replies)
Discussion started by: NVA
0 Replies
TRACE-CMD-RESTORE(1)													      TRACE-CMD-RESTORE(1)

NAME
trace-cmd-restore - restore a failed trace record SYNOPSIS
trace-cmd restore [OPTIONS] [command] cpu-file [cpu-file ...] DESCRIPTION
The trace-cmd(1) restore command will restore a crashed trace-cmd-record(1) file. If for some reason a trace-cmd record fails, it will leave a the per-cpu data files and not create the final trace.dat file. The trace-cmd restore will append the files to create a working trace.dat file that can be read with trace-cmd-report(1). When trace-cmd record runs, it spawns off a process per CPU and writes to a per cpu file usually called trace.dat.cpuX, where X represents the CPU number that it is tracing. If the -o option was used in the trace-cmd record, then the CPU data files will have that name instead of the trace.dat name. If a unexpected crash occurs before the tracing is finished, then the per CPU files will still exist but there will not be any trace.dat file to read from. trace-cmd restore will allow you to create a trace.dat file with the existing data files. OPTIONS
-c Create a partial trace.dat file from the machine, to be used with a full trace-cmd restore at another time. This option is useful for embedded devices. If a server contains the cpu files of a crashed trace-cmd record (or trace-cmd listen), trace-cmd restore can be executed on the embedded device with the -c option to get all the stored information of that embedded device. Then the file created could be copied to the server to run the trace-cmd restore there with the cpu files. If *-o* is not specified, then the file created will be called 'trace-partial.dat'. This is because the file is not a full version of something that trace-cmd-report(1) could use. -t tracing_dir Used with -c, it overrides the location to read the events from. By default, tracing information is read from the debugfs/tracing directory. -t will use that location instead. This can be useful if the trace.dat file to create is from another machine. Just tar -cvf events.tar debugfs/tracing and copy and untar that file locally, and use that directory instead. -k kallsyms Used with -c, it overrides where to read the kallsyms file from. By default, /proc/kallsyms is used. -k will override the file to read the kallsyms from. This can be useful if the trace.dat file to create is from another machine. Just copy the /proc/kallsyms file locally, and use -k to point to that file. -o output' By default, trace-cmd restore will create a trace.dat file (or trace-partial.dat if -c is specified). You can specify a different file to write to with the -o option. -i input By default, trace-cmd restore will read the information of the current system to create the initial data stored in the trace.dat file. If the crash was on another machine, then that machine should have the trace-cmd restore run with the -c option to create the trace.dat partial file. Then that file can be copied to the current machine where trace-cmd restore will use -i to load that file instead of reading from the current system. EXAMPLES
If a crash happened on another box, you could run: $ trace-cmd restore -c -o box-partial.dat Then on the server that has the cpu files: $ trace-cmd restore -i box-partial.dat trace.dat.cpu0 trace.dat.cpu1 This would create a trace.dat file for the embedded box. SEE ALSO
trace-cmd(1), trace-cmd-record(1), trace-cmd-report(1), trace-cmd-start(1), trace-cmd-stop(1), trace-cmd-extract(1), trace-cmd-reset(1), trace-cmd-split(1), trace-cmd-list(1), trace-cmd-listen(1) AUTHOR
Written by Steven Rostedt, <rostedt@goodmis.org[1]> RESOURCES
git://git.kernel.org/pub/scm/linux/kernel/git/rostedt/trace-cmd.git COPYING
Copyright (C) 2010 Red Hat, Inc. Free use of this software is granted under the terms of the GNU Public License (GPL). NOTES
1. rostedt@goodmis.org mailto:rostedt@goodmis.org 06/11/2014 TRACE-CMD-RESTORE(1)
All times are GMT -4. The time now is 03:23 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy