Sponsored Content
Operating Systems Solaris Solaris cluster critical issue Post 302936862 by achenle on Saturday 28th of February 2015 05:59:05 PM
Old 02-28-2015
How did you fix it on your test cluster, where you test these kind of things before trying them on critical production systems?
 

8 More Discussions You Might Find Interesting

1. High Performance Computing

Building a Solaris Cluster Express cluster in a VirtualBox on OpenSolaris

Provides a description of how to set up a Solaris Cluster Express cluster in a VirtualBox on OpenSolaris. More... (0 Replies)
Discussion started by: Linux Bot
0 Replies

2. Solaris

Issue while installing: Solaris 10 SPARC Recommended Patch Cluster (2009.10.23)

Hello, As explained, I've encountered an issue while installing Solaris 10 SPARC Recommended Patch Cluster (2009.10.23). Actually, patch no 120011-14 stops with the following error: ERROR: attribute verification of </var/run/.patchSafeMode/root/usr/bin/passwd> failed file type <f>... (6 Replies)
Discussion started by: a.mauger
6 Replies

3. Solaris

Sun Cluster configuration issue

I am using VMware Workstation-7 on Windows-XP host . I am trying to configure Solaris 10-X86 guest os based 2 nodes Sun Cluster . I have added one extra Virtual Lan adapter on my VMware with another subnet (that I would like to put for SUN Cluster private communication). I have... (0 Replies)
Discussion started by: sanjee
0 Replies

4. Solaris

Sun Cluster switching issue

I have installed sun cluster 3.2 on two sprac servers. Configured a failover resource group. Added a LogicalHostname resource to it. LogicalHostname is also added to /etc/hosts with ip address. I am able to access cluster by share ip used for logical hostname but when i try to switch the resource... (0 Replies)
Discussion started by: ahmadnauman
0 Replies

5. Solaris

Sun Cluster 3.2 Issue

Hello everyone, I have two Solaris 10 servers that are on cluster. The cluster is a Sun Cluster 3.2 I have a script cronned that stop/start a ressource in a resource group everyday. Today I have checked the status of the ressources and I found that my ressource group have a "Error--stop... (1 Reply)
Discussion started by: adilyos
1 Replies

6. Red Hat

Linux Cluster failover issue

Hi Guys, I am not much aware of clusters but i have few questions can someone provide the overview as it would be very helpful for me. How can i perform cluster failover test to see all the services are failing back to other node ? If it is using veritas cluster then what kind of... (2 Replies)
Discussion started by: munna529
2 Replies

7. AIX

Cluster communication issue

Hi, I am using Power HA7.1.1 SP5 AIx 7.1 My both cluster nodes are independently working. RG informations are not updating each other. Node A shows that node B is down and vice versa. RG1 is running node A, RG2 running on node B. === clRGinfo From Node B === RG01 OFFLINE ... (2 Replies)
Discussion started by: sunnybee
2 Replies

8. Solaris

Patching Procedure in Solaris 10 with sun cluster having Solaris zone

Hi Gurus I am not able to find the patching procedure for solaris 10 ( sol10 u11) to latest patchset with sun cluster having failover zones so that same I should follow. Take an instance, there are sol1 and sol2 nodes and having two failover zones like sozone1-rg and sozone2-rg and currently... (1 Reply)
Discussion started by: nick101
1 Replies
Apache::TestSmoke(3pm)					User Contributed Perl Documentation				    Apache::TestSmoke(3pm)

NAME
Apache::TestSmoke - Special Tests Sequence Failure Finder SYNOPSIS
# get the usage and the default values % t/SMOKE -help # repeat all tests 5 times and save the report into # the file 'myreport' % t/SMOKE -times=5 -report=myreport # run all tests default number of iterations, and repeat tests # default number of times % t/SMOKE # same as above but work only the specified tests % t/SMOKE foo/bar foo/tar # run once a sequence of tests in a non-random mode # e.g. when trying to reduce a known long sequence that fails % t/SMOKE -order=rotate -times=1 foo/bar foo/tar # show me each currently running test # it's not the same as running the tests in the verbose mode % t/SMOKE -verbose # run t/TEST, but show any problems after *each* tests is run # useful for bug reports (it actually runs t/TEST -start, then # t/TEST -run for each test separately and finally t/TEST -stop % t/SMOKE -bug_mode # now read the created report file DESCRIPTION
The Problem When we try to test a stateless machine (i.e. all tests are independent), running all tests once ensures that all tested things properly work. However when a state machine is tested (i.e. where a run of one test may influence another test) it's not enough to run all the tests once to know that the tested features actually work. It's quite possible that if the same tests are run in a different order and/or repeated a few times, some tests may fail. This usually happens when some tests don't restore the system under test to its pristine state at the end of the run, which may influence other tests which rely on the fact that they start on pristine state, when in fact it's not true anymore. In fact it's possible that a single test may fail when run twice or three times in a sequence. The Solution To reduce the possibility of such dependency errors, it's helpful to run random testing repeated many times with many different srand seeds. Of course if no failures get spotted that doesn't mean that there are no tests inter-dependencies, which may cause a failure in production. But random testing definitely helps to spot many problems and can give better test coverage. Resolving Sequence Problems When this kind of testing is used and a failure is detected there are two problems: 1. First is to be able to reproduce the problem so if we think we fixed it, we could verify the fix. This one is easy, just remember the sequence of tests run till the failed test and rerun the same sequence once again after the problem has been fixed. 2. Second is to be able to understand the cause of the problem. If during the random test the failure has happened after running 400 tests, how can we possibly know which previously running tests has caused to the failure of the test 401. Chances are that most of the tests were clean and don't have inter-dependency problem. Therefore it'd be very helpful if we could reduce the long sequence to a minimum. Preferably 1 or 2 tests. That's when we can try to understand the cause of the detected problem. This utility attempts to solve both problems, and at the end of each iteration print a minimal sequence of tests causing to a failure. This doesn't always succeed, but works in many cases. This utility: 1. Runs the tests randomly until the first failure is detected. Or non-randomly if the option -order is set to repeat or rotate. 2. Then it tries to reduce that sequence of tests to a minimum, and this sequence still causes to the same failure. 3. (XXX: todo): then it reruns the minimal sequence in the verbose mode and saves the output. 4. It reports all the successful reductions as it goes to STDOUT and report file of the format: smoke-report-<date>.txt. In addition the systems build parameters are logged into the report file, so the detected problems could be reproduced. 5. Goto 1 and run again using a new random seed, which potentially should detect different failures. Reduction Algorithm Currently for each reduction path, the following reduction algorithms get applied: 1. Binary search: first try the upper half then the lower. 2. Random window: randomize the left item, then the right item and return the items between these two points. t/SMOKE.PL t/SMOKE.PL is driving this module, if you don't have it, create it: #!perl use strict; use warnings FATAL => 'all'; use FindBin; use lib "$FindBin::Bin/../Apache-Test/lib"; use lib "$FindBin::Bin/../lib"; use Apache::TestSmoke (); Apache::TestSmoke->new(@ARGV)->run; usually Makefile.PL converts it into t/SMOKE while adjusting the perl path, but you create t/SMOKE in first place as well. AUTHOR
Stas Bekman perl v5.14.2 2011-02-08 Apache::TestSmoke(3pm)
All times are GMT -4. The time now is 01:29 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy