Sponsored Content
Top Forums UNIX for Advanced & Expert Users VCS triggerring panic on 1 node, root disk under SVM Post 302811113 by amlanroy on Thursday 23rd of May 2013 04:29:13 AM
Old 05-23-2013
VCS triggerring panic on 1 node, root disk under SVM

We have two node cluster with OS disk mirrored under SVM. There is slight disk problem on one of the mirror disk causing cluster to panic.
Failure of one mirror disk causing VCS to panic the node. Why VCS is not able to write /var filesystem, as one of the disk is healthy.

--------------------------------------------------------------------------------------------
From VCS engine log,

2013/04/20 02:34:18 VCS INFO V-16-1-50135 User root fired command: hagrp -unfreeze ORACLE_PRASAPDB_Group from localhost
2013/04/21 10:38:14 VCS INFO V-16-1-10196 Cluster logger started
2013/04/21 10:38:14 VCS NOTICE V-16-1-11022 VCS engine (had) started


From OS messages,

Apr 21 04:00:26 prdagwn1 genunix: [ID 111219 kern.notice] dumping to /dev/dsk/c1t0d0s1, offset 65536, content: kernel Apr 21 04:00:31 prdagwn1 scsi: [ID 365881 kern.info] /pci@400/pci@0/pci@8/scsi@0 (mpt0):
Apr 21 04:00:31 prdagwn1 Log info 31140000 received for target 1.
Apr 21 04:00:31 prdagwn1 scsi_status=0, ioc_status=8048, scsi_state=c
Apr 21 04:01:32 prdagwn1 md_stripe: [ID 641072 kern.warning] WARNING: md: d20: read error on /dev/dsk/c1t1d0s0 Apr 21 04:02:32 prdagwn1 last message repeated 1 time Apr 21 04:03:33 prdagwn1 md_stripe: [ID 641072 kern.warning] WARNING: md: d20: write error on /dev/dsk/c1t1d0s0 Apr 21 04:07:42 prdagwn1 genunix: [ID 409368 kern.notice] ^M100% done: 405019 pages dumped, compression ratio 3.71, Apr 21 04:07:42 prdagwn1 genunix: [ID 851671 kern.notice] dump succeeded

I can see the md has problem in write access on disk slice c1t1d0s0 which has /var included. If the directory /var is in question with access problem, VCS GAB would trigger system panic due to client process failure as VCS daemon 'had' can't access the /var directory.

Oracle suggested to replace disk1 after initial investigation. This is the same disk that failed 3 weeks back

Is there any setting we can do in VCS to accomodate delayed write in disk under SVM ? Or do we move from SVM to VxVM ?
 

10 More Discussions You Might Find Interesting

1. Solaris

Disk replacement with svm

I dont even know what raid level this is, but its raid 5 mirrored from the looks of it. I have a failed disk (t12) within this mirror. What is the best way to replace this disk? 2 things concern me, isn't there a command to prepare the disk for a hot swap? and what should i do with the... (3 Replies)
Discussion started by: BG_JrAdmin
3 Replies

2. Solaris

Why didn't she panic? (Sol 10 + SVM + HDS)

Hi folks, the following incident occured today: by mistake one of our renowned administrators deleted the complete zoning for a 25K domain running solaris 10. Thus the system lost all of it's external disks. We've got oracle datafiles and oracle software residing on those lost... (6 Replies)
Discussion started by: MikaBaghinen
6 Replies

3. Solaris

mirroring root disk using svm - but no free slices for metadb's

Hi all, we have an existing system that was configured using just one of the (two) internal disks. I want to mirror the disk using SVM, but have realised there is no free slice for creating the metadb's. Is there a workaround I can use for this? In the past we have always kept slice 7 free -... (8 Replies)
Discussion started by: badoshi
8 Replies

4. Solaris

Root Disk mirroring in SVM

Dear All, Please help me to configure root mirroring using SVM in Solaris 9. Thanks and Regards, Lakkireddy BR (3 Replies)
Discussion started by: lbreddy
3 Replies

5. Solaris

SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node

Hi, Is it possible to have a Solaris cluster of 2 nodes at SITE-A using SVM and creating metaset using say 2 LUNs (on SAN). Then replicating these 2 LUNs to remote site SITE-B via storage based replication and then using these LUNs by importing them as a metaset on a server at SITE-B which is... (0 Replies)
Discussion started by: dn2011
0 Replies

6. Hardware

Cannot mount root - Panic

Hi Guys, I just tried to reboot my server, but it seems to be panicing and unable to mount the root disk. The filesystem is supposed to be ufs, so im not sure why its talking about vfs. At the moment it's just stuck in a loop of Reboot-panic-reboot. Any ideas? A. SunOS Release 5.10... (1 Reply)
Discussion started by: A-Train
1 Replies

7. Solaris

Root disk mirroring in SVM

I tried doing rootdisk mirroring in my local host , i added a new Ide disk in my system and copied the prtvtoc from root disk to the newly added disk, and then when i tried to add database replicas on both the disks, it was added for boot disk but for the newly added disk i gave the error, which... (6 Replies)
Discussion started by: Laxxi
6 Replies

8. Solaris

VCS - 3 node - IP change

Hey Admins, I need steps to do an IP changes on 3-node VCS cluster. All IP's (public, cluster IP's, backup IP, etc.)want to change. We have a downtime scheduled. Please let me know the steps to do so. Thanks in advance. (5 Replies)
Discussion started by: snchaudhari2
5 Replies

9. UNIX for Beginners Questions & Answers

SVM recover root

Hi,, i have a solaris 10 machine , that let the root password expire.. I am trying to recover from failsafe mode. In the old days it was rater simple , boot to cdrom , mount the root disk, /a cd /a and wipe out root pw string.. Now i have to try and do it remotely , from failsafe and this... (2 Replies)
Discussion started by: goya
2 Replies

10. AIX

Crash dump and Panic message : RSCT Dead Man Switch Timeout for HACMP; halting non-responsive node

Dear all i have two aix system -Model : P770 -OS version: AIX 6.1 -patch level : 6100-07-04-1216 -ha version : HACMP v 6.1.0.8 -host : A, B last Wednesday, my B system suddenly went down with crash dump. after 1 minute, A system went down with crash dump. I checked the dump of A-system... (6 Replies)
Discussion started by: tomato00
6 Replies
volreattach(8)						      System Manager's Manual						    volreattach(8)

NAME
volreattach - Reattaches disk drives that have once again become accessible SYNOPSIS
/usr/sbin/volreattach [-rb] [accessname...] /usr/sbin/volreattach -c accessname OPTIONS
The volreattach utility has the following options: Calls the volrecover utility to attempt to recover stale plexes of any volumes on the failed disk. Performs the reattach operation in the background. Checks whether a reattach is possible. No operation is performed, but the name of the disk group and disk media name at which the disk can be reattached is displayed. DESCRIPTION
The volreattach utility attempts to reattach disks using the same disk group and media names they had before detachment. This operation may be necessary if a disk has a transient failure, or if some disk drivers are unloaded or unloadable when the Logical Storage Manager is started, causing disks to enter the failed state. If the problem is fixed, the volreattach command can be used to reattach the disks without plexes being flagged as stale, as long as the reattach happens before any volumes on the disk are started. The volreattach command is called from the voldiskadm menus as part of disk recovery. The volreattach utility tries to find a disk with a disk group and disk ID matching that of the disk(s) being reattached. If the matching disk is found, the reattach operation may still fail if the original cause (or some other cause) for the disk failure still exists. EXIT CODES
A zero exit status is returned if it is possible to perform a reattach. Otherwise, non-zero is returned. SEE ALSO
volintro(8), voldiskadm(8), volrecover(8) volreattach(8)
All times are GMT -4. The time now is 11:27 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy