I lost my system volume in a power outage, but fortunately I had a dual boot and I could boot into an older opensolaris version and my raidz2 7 drive pool was still fine. I even scrubbed it, no errors. However, the older os has some smb problems so I wanted to upgrade to opensolaris11. I accidentally did the auto install option and it overwrote one of my data drives. This is really important data, but fortunately I have a raidz2, so how bad can it be?... BAD! Booting back into the earlier solaris version I cannot access my pool, This is what zpool status shows:
zpool status
pool: brick3
state: UNAVAIL
scrub: none requested
config:
I tried to export and re-import it, and I get the following message:
I wanted to update this thread in case anyone is stupid enough to have this same problem. In desperation I installed the new opensolaris11 on a the right drive this time, and then I was able to import my pool in a degraded but functional state. Since this is a raidz2 pool I should be able to restore just fine. I don't know what opensolaris11 did to my pool, but then again I don't really care either if I can have my data back. Whew that was scary. I vow to back up everything really important. zfs is great but you can still loose everything in the blink of an eye.
I thought I was OK, but zpool scrub hangs forever at 20% across multiple cold boots, and importing from oracle solaris 11 live hangs forever... I what do I do now, older versions of open solaris will not import the pool anymore...
---------- Post updated at 07:36 AM ---------- Previous update was at 12:48 AM ----------
Well it's 4AM and now I am getting mad. I think this whole bloody mess is caused by Oracle's 'great new solaris 11' package. My advice, don't touch this piece.
To recap, I wanted to try the best newest Solaris release for my super critical file server, so I downloaded from Oracle what I thought was the right iso. I booted, and just as the boot menu came up I got a phone call. When I came back the thing had automatically installed a new version of solaris over one of the drives in the zpool. (BAD #1 ).
Since I have a raidz2 I wasn't too worried ( at first ) and I booted into my original 2009 opensolaris. However, I got the errors shown in the above posts. I exported and could not re-import my pool. Oracale had done something to that drive breaking my entire pool even though it is a radz2 and should be tolerant of 2 drive failures at the same time. ( BAD #2 ).
Since I could do nothing with my pool with my old OS, I tried on the new Oracle solaris, and could indeed import my pool in degraded state because of the one overwritten drive. Fine, I wanted to scrub everything first ( I don't know if this was wise or not ) so I did zpool scrub, which eventually hung forever at 11%. All access to the drive was similarly hung. Rebooting the machine did not change this situation which seemed increasingly dire ( BAD #3 )
I finally got out of this problem by unplugging drives to fault the pool and rebooting in single user mode. Eventually I was able to stop the scrub with the "zpool scrub -s" command, in single user mode. And when I rebooted, I could access my pool again. My first priority at this point was to back up all data immediately. I began to copy off my most important stuff, but unfortunately before I could copy even a fraction off the file system hung again ( BAD #4 ).
Googling around I found most causes for hanging zpool commands are related to hardware failure, so going on a hunch, I figured Oracle phased out or screwed up the drivers for my disks. I still could not import my pool in the old opensolaris OS, because of whatever the Oracle install wrote on that drive. So I booted in Oracle solaris, in single user mode, and did "zpool offline <pool> <drive>" and it worked! Then I rebooted into good old opensolaris, and imported my pool. It worked!!!!
So at this point it is now more like 5AM and I have backed up most of my critical data, way more than I could before anyway. It appears at this point that I was correct and the drivers for either my motherboard or my hard drive controller card were broken by the Oracle release in a way that let it silently trash my zpool. I have a SIIG 2 drive sata and a MSI n1996 motherboard, not sure which the problem is with, but whichever, it works fine in opensolaris 2009 and previous versions.
I just want to warn people that are not real Solaris experts from even trying this Oracle package. Personally I am migrating to fbsd as soon as I can...
Hi guys, I appreciate any help in this regard, we have lost sensitive data in the company.
One box with 2 disk mirrored and a 3ware controller handling 13 disks in a raidz2 pool. Suddenly the box restart and keeps "Reading ZFS config" for hours.
Unplugging disk by disk we isolate the disk... (3 Replies)
Hey all!
I was hoping someone knew anything about this one...
I know with Solaris Volume Manager the default Database Replica size is 8192 blocks (4MB approximately)
Now I know you can increase this amount but is there any point?
The reason I am asking this is that I've setup mirroring on... (2 Replies)
Hi all,
I recently started exploring Solaris 10.
I am testing metadevices now.
I have been reading about the state databases here: 6.State Database (Overview) (Solaris Volume Manager Administration Guide) - Sun Microsystems
So I created 3 metadbs on 2 slices (6 in total; c1t1d0s3... (3 Replies)
Hi,
When I execute one of my shellscript I am getting the below mentioned error message .This application takes 2input files which have the records counts 26463 and 1178046
exec(2): insufficient swap or memory available.
exec(2): insufficient swap or memory available.
exec(2): insufficient swap... (3 Replies)
Hi,
when navigating on application webpages (peoplesoft applications) the users receive :
INSUFFICIENT AVAILABLE MEMORY.
I issued vmstat on UNIX server ( where is hosted Web server and application server and DB). Here is the result :
:vmstat 2 10
System configuration: lcpu=4 mem=30720MB... (8 Replies)
Good morning,
I have Solstice disk suite installed on my server.
One disk broke so I sostitute it.
A replica was present on this disk.
I had delete and then recreate it with commands metadb -d and metadb -a.
Now when I inquire for the status of replicas I see this:
stp11# metadb -i
... (2 Replies)
First I would like to thank this forum for assisting me in setting up my 1st sunbox.
Could not have done it if it had not been for you guys and google :D
I have mirrored my box and have SUCCESSFULLY tested booting from the rootdisk and rootmirror successfully.
I am now looking at configuring... (2 Replies)
We have an application running on Win2K and this application ftp files to HP-UX using ftpdc as user id. The files are created in HP-UX with following permissions:
-rw-r----- 1 ftpdc users 968321 Apr 12 22:57 aaaa.txt
There is a job that runs on HP-UX trying to modify this file using the... (7 Replies)
Hello,
We are using Solstice Disk Suite on Solaris 2.7.
We want to add two striped volume with six disks.
On each disk, we take a slice and we create the stripe.
That I want to know :
Is it necessary to add two replicas on the same slice on the new disks, as made before on the others... (1 Reply)