Solaris 10 Live Upgrade Issue


 
Thread Tools Search this Thread
Operating Systems Solaris Solaris 10 Live Upgrade Issue
# 1  
Old 05-07-2012
Solaris 10 Live Upgrade Issue

Hi Guys, having an issue with running Live Upgrade on a t5240 runiing Solaris 10 5/08.

The system has the required patches 121430, and Live upgrade was updated from the install media sol-10-u10-ga2-sparc-dvd.iso

The following boot environments were created
solenv1 and solenv2 with the lucreate command and this completed sucessfully.

Code:
# lustatus
Boot Environment           Is       Active Active    Can    Copy
Name                       Complete Now    On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1                    yes      yes    yes       no     -
solenv2                    yes      no     no        yes    -

I had an issue when running the lupgrade command, it got to 98% complete

Code:
# luupgrade -u -k /var/tmp/LDMupgrade/autoreg  -n solenv2 -s /sol_dvd

64459 blocks
miniroot filesystem is <lofs>
Mounting miniroot at </sol_dvd/Solaris_10/Tools/Boot>
#######################################################################
 NOTE: To improve products and services, Oracle Solaris communicates
 configuration data to Oracle after rebooting.

 You can register your version of Oracle Solaris to capture this data
 for your use, or the data is sent anonymously.

 For information about what configuration data is communicated and how
 to control this facility, see the Release Notes or
 Redirecting....

 INFORMATION: After activated and booted into new BE <solenv2>,
 Auto Registration happens automatically with the following Information

autoreg=disable
#######################################################################
Validating the contents of the media </sol_dvd>.
The media is a standard Solaris media.
The media contains an operating system upgrade image.
The media contains <Solaris> version <10>.
Constructing upgrade profile to use.
Locating the operating system upgrade program.
Checking for existence of previously scheduled Live Upgrade requests.
Creating upgrade profile for BE <solenv2>.
Determining packages to install or upgrade for BE <solenv2>.
Performing the operating system upgrade of the BE <solenv2>.
CAUTION: Interrupting this process may leave the boot environment unstable
or unbootable.
Read from remote host server: Connection reset by peer
Connection to server closed.

Since this happened, lustatus shows the following output

Code:
# lustatus
Boot Environment           Is       Active Active    Can    Copy
Name                       Complete Now    On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1                    yes      yes    yes       no     -
solenv2                    yes      no     no        no     UPDATING

/etc/lutab shows the following

Code:
# cat /etc/lutab
# DO NOT EDIT THIS FILE BY HAND. This file is not a public interface.
# The format and contents of this file are subject to change.
# Any user modification to this file may result in the incorrect
# operation of Live Upgrade.
1:solenv1:C:0
1:/:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:1
1:boot-device:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:2
2:solenv2:C:0
2:/:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:1
2:boot-device:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:2

What can I do remove the BE solenv2 and start again?

Any suggestions would be most appreciated

Thanks
# 2  
Old 05-07-2012
This User Gave Thanks to DukeNuke2 For This Post:
# 3  
Old 05-07-2012
Hi Duke, thanks for that, I had a look at that document, early this morning ( was at the end of an 18 hour session so may not have been focussing properly)

I have just tried it again

Code:
# lustatus
Boot Environment           Is       Active Active    Can    Copy
Name                       Complete Now    On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1                    yes      yes    yes       no     -
solenv2                    yes      no     no        no     UPDATING

I was looking at the can delete column, which indicate it couldn't be deleted, but when actually trying to run it, I get the following

Code:
# ludelete solenv2
INFORMATION: Removing invalid lock file.
ERROR: mount: The state of /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0 is not okay
        and it was attempted to be mounted read/write
mount: Please run fsck and try again
ERROR: cannot mount mount point </.alt.tmp.b-8og.mnt> device </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0>
ERROR: failed to mount file system </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0> on </.alt.tmp.b-8og.mnt>
ERROR: cannot mount boot environment by icf file </etc/lu/ICF.2>
WARNING: Unable to mount ABE <solenv2>: cannot complete lumk_iconf
WARNING: Unable to determine disk partition configuration information for BE <solenv2>.
ERROR: mount: The state of /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0 is not okay
        and it was attempted to be mounted read/write
mount: Please run fsck and try again
ERROR: cannot mount mount point </.alt.tmp.b-hpg.mnt> device </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0>
ERROR: failed to mount file system </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0> on </.alt.tmp.b-hpg.mnt>
ERROR: cannot mount boot environment by icf file </etc/lu/ICF.2>
ERROR: Cannot mount BE <solenv2>.
mount: Please run fsck and try again
luupdall: WARNING: Could not mount the Root Slice of BE:"solenv2".
mount: The state of /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0 is not okay
        and it was attempted to be mounted read/write

However, output from lustatus has now changed

Code:
# lustatus
Boot Environment           Is       Active Active    Can    Copy
Name                       Complete Now    On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1                    yes      yes    yes       no     -
solenv2                    yes      no     no        yes    -

Would I be right in thinking that I can perform an lucreate, using the same names again?
# 4  
Old 05-07-2012
you have to delete solenv2 before you can create the environment with the same name again... but there seems to be a problem with the mountpoints for the new environment. they seem to be on a external device (SAN?) which can not be mounted properly (from what i can see in the error messages...).
This User Gave Thanks to DukeNuke2 For This Post:
# 5  
Old 05-07-2012
Hi Duke, yes that is correct the disks are on a SAN.

I have tried running fsck on the specified disk

Code:
# cat /etc/lutab
# DO NOT EDIT THIS FILE BY HAND. This file is not a public interface.
# The format and contents of this file are subject to change.
# Any user modification to this file may result in the incorrect
# operation of Live Upgrade.
1:solenv1:C:0
1:/:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:1
1:boot-device:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:2
2:solenv2:C:0
2:/:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:1
2:boot-device:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:2

# echo | format
Searching for disks...done


AVAILABLE DISK SELECTIONS:
       0. c6t60A9800057396D64685A4D7A51725458d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 3072>
          /scsi_vhci/ssd@g60a9800057396d64685a4d7a51725458
       1. c6t60A9800057396D64685A4D7A52675733d0 <NETAPP-LUN-7320-500.08GB>
          /scsi_vhci/ssd@g60a9800057396d64685a4d7a52675733
       2. c6t60A9800057396D64685A51774F6E712Dd0 <NETAPP-LUN-0.2 cyl 5118 alt 2 hd 16 sec 512>
          /scsi_vhci/ssd@g60a9800057396d64685a51774f6e712d
       3. c6t60A9800057396D64685A62414D324E54d0 <NETAPP-LUN-7320-50.00GB>
          /scsi_vhci/ssd@g60a9800057396d64685a62414d324e54
       4. c6t60A9800057396D64685A62414D345041d0 <NETAPP-LUN-7320-10.00GB>
          /scsi_vhci/ssd@g60a9800057396d64685a62414d345041
       5. c6t60A9800057396D6468344D7A4E795678d0 <NETAPP-LUN-0.2 cyl 5118 alt 2 hd 16 sec 512>
          /scsi_vhci/ssd@g60a9800057396d6468344d7a4e795678
       6. c6t60A9800057396D6468344D7A4F4A3561d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 2048>
          /scsi_vhci/ssd@g60a9800057396d6468344d7a4f4a3561
       7. c6t60A9800057396D6468344D7A4F356151d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 4096>
          /scsi_vhci/ssd@g60a9800057396d6468344d7a4f356151
       8. c6t60A9800057396D6468344D7A4F424571d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 1024>
          /scsi_vhci/ssd@g60a9800057396d6468344d7a4f424571
       9. c6t60A9800057396D6468344D7A4F456848d0 <NETAPP-LUN-0.2 cyl 5118 alt 2 hd 16 sec 256>
          /scsi_vhci/ssd@g60a9800057396d6468344d7a4f456848
      10. c6t60A9800057396D6468344D7A4F483044d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 2048>
          /scsi_vhci/ssd@g60a9800057396d6468344d7a4f483044
Specify disk (enter its number): Specify disk (enter its number):

below is the output
Quote:
# fsck -F ufs /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0
** /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0
** Last Mounted on /a
** Phase 1 - Check Blocks and Sizes
INCORRECT DISK BLOCK COUNT I=400 (416 should be 224)
CORRECT? y

FRAGMENT 49976 DUP I=5828 LFN 0
<snip>
EXCESSIVE DUPLICATE FRAGMENTS I=5828
CONTINUE? y

<snip>

** Phase 2 - Check Pathnames
DIRECTORY CORRUPTED I=1497 OWNER=root MODE=40755
SIZE=512 MTIME=May 7 00:07 2012
DIR=?

SALVAGE? yes

MISSING '.' I=1497 OWNER=root MODE=40755
SIZE=512 MTIME=May 7 00:07 2012
DIR=?

FIX? yes

<snip>

FILESYSTEM MAY STILL BE INCONSISTENT.
224419 files, 10919276 used, 110132692 free (6596 frags, 13765762 blocks, 0.0% fragmentation)

***** FILE SYSTEM WAS MODIFIED *****
ORPHANED DIRECTORIES REATTACHED; DIR LINK COUNTS MAY NOT BE CORRECT.
***** FILE SYSTEM IS BAD *****

***** PLEASE RERUN FSCK *****
#
A rerun of fsck brings

Quote:
224419 files, 10920222 used, 110131746 free (5818 frags, 13765741 blocks, 0.0% fragmentation)

***** FILE SYSTEM WAS MODIFIED *****
ORPHANED DIRECTORIES REATTACHED; DIR LINK COUNTS MAY NOT BE CORRECT.

***** PLEASE RERUN FSCK *****
I am then able to delete the BE solenv2

Quote:
# ludelete solenv2
Updating boot environment configuration database.
Updating boot environment description database on all BEs.
Updating all boot environment configuration databases.
#
# lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1 yes yes yes no -
#
So it looks like i'm getting back on track, thanks very much for your input Smilie

As an aside, looking at the results form the secondary run of fsck, should I rerun it?

Last edited by Revo; 05-07-2012 at 10:41 AM.. Reason: spelling mistake
# 6  
Old 05-07-2012
as you've deleted the BE i don't think it is neccessary to re-run fsck... just try to create the BE again after creating a new fs (newfs) on the target disk(s).
This User Gave Thanks to DukeNuke2 For This Post:
# 7  
Old 05-07-2012
Hi Duke, reran the lucreate command, and this seemed to work fine

Code:
# newfs /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0
newfs: /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0 last mounted as /.alt.tmp.b-s3b.mnt
newfs: construct a new file system /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0: (y/n)? y
Warning: 2048 sector(s) in last cylinder unallocated
/dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0:      245825536 sectors in 40011 cylinders of 48 tracks, 128 sectors
        120032.0MB in 2501 cyl groups (16 c/g, 48.00MB/g, 5824 i/g)
super-block backups (for fsck -F ufs -o b=#) at:
 32, 98464, 196896, 295328, 393760, 492192, 590624, 689056, 787488, 885920,
Initializing cylinder groups:
.................................................
super-block backups for last 10 cylinder groups at:
 244882848, 244981280, 245079712, 245178144, 245276576, 245366816, 245465248,
 245563680, 245662112, 245760544


Code:
# lucreate -c solenv1 -m /:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:ufs -n solenv2 
Determining types of file systems supported
Validating file system requests
Preparing logical storage devices
Preparing physical storage devices
Configuring physical storage devices
Configuring logical storage devices
Analyzing system configuration.
Updating boot environment description database on all BEs.
Updating system configuration files.
The device </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0> is not a root device for any boot environment; cannot get BE ID.
Creating configuration for boot environment <solenv2>.
Source boot environment is <solenv1>.
Creating file systems on boot environment <solenv2>.
Creating <ufs> file system for </> in zone <global> on </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0>.
Mounting file systems for boot environment <solenv2>.
Calculating required sizes of file systems for boot environment <solenv2>.
Populating file systems on boot environment <solenv2>.
Analyzing zones.
Mounting ABE <solenv2>.
Generating file list.
Copying data from PBE <solenv1> to ABE <solenv2>.
100% of filenames transferred
Finalizing ABE.
Fixing zonepaths in ABE.
Unmounting ABE <solenv2>.
Fixing properties on ZFS datasets in ABE.
Reverting state of zones in PBE <solenv1>.
Making boot environment <solenv2> bootable.
Population of boot environment <solenv2> successful.
Creation of boot environment <solenv2> successful.


Code:
# luupgrade -u -k /var/tmp/LDMupgrade/autoreg -n solenv2 -s /sol_dvd
miniroot filesystem is <lofs>
Mounting miniroot at </sol_dvd/Solaris_10/Tools/Boot>
#######################################################################
 NOTE: To improve products and services, Oracle Solaris communicates
 configuration data to Oracle after rebooting. 

 You can register your version of Oracle Solaris to capture this data
 for your use, or the data is sent anonymously. 

 For information about what configuration data is communicated and how
 to control this facility, see the Release Notes or
 www.oracle.com/goto/solarisautoreg. 
 
 INFORMATION: After activated and booted into new BE <solenv2>,
 Auto Registration happens automatically with the following Information 

autoreg=disable
#######################################################################
Validating the contents of the media </sol_dvd>.
The media is a standard Solaris media.
The media contains an operating system upgrade image.
The media contains <Solaris> version <10>.
Constructing upgrade profile to use.
Locating the operating system upgrade program.
Checking for existence of previously scheduled Live Upgrade requests.
Creating upgrade profile for BE <solenv2>.
Determining packages to install or upgrade for BE <solenv2>.
Performing the operating system upgrade of the BE <solenv2>.
CAUTION: Interrupting this process may leave the boot environment unstable 
or unbootable.
May  7 17:33:01 <server> ufs: NOTICE: /a: bad dir ino 1729 at offset 0: mangled entry
Upgrading Solaris: 14% completed

Is the bad dir ino message something to worry about?
I've had a look on google but cant seem to find anything that matches
Login or Register to Ask a Question

Previous Thread | Next Thread

5 More Discussions You Might Find Interesting

1. Solaris

Solaris patching issue with Live Upgrade

I have Solaris-10 sparc box with ZFS file-system, which is running two non global zones. I am in process of applying Solaris Recommended patch cluster via Live Upgrade. Though I have enough space in root file-system of both zones, everytime I run installcluster, it fails with complaining less... (7 Replies)
Discussion started by: solaris_1977
7 Replies

2. Solaris

Live Upgrade Issue

I tried a live upgrade for one my solaris 10u8 server which didnt go sucessfull and after that i now have following mounts in memory. df: cannot statvfs /.alt.sol10u8_2/var: No such file or directory df: cannot statvfs /.alt.sol10u8_2/var/run: No such file or directory df: cannot statvfs... (0 Replies)
Discussion started by: fugitive
0 Replies

3. Solaris

Live upgrade issue with ZFS Root

I have created a solaris10 update9 zfs root flash archive for sun4v environment which i 'm tryin to use for upgrading solaris10 update8 zfs root based server using live upgrade. following is my current system status lustatus Boot Environment Is Active Active Can Copy Name Complete Now... (0 Replies)
Discussion started by: fugitive
0 Replies

4. Solaris

Solaris Live Upgrade issue with Zones

I 'm running solaris10 u6 with 141414-02. My system is T5220 running 2 Ldoms and 7 zones on primary domain. I 'm tryin to create booth environment on my alternate root disk after breaking the SVM mirroring but it does not go well and stuck at some point , and i have to reboot the system to get rid... (1 Reply)
Discussion started by: fugitive
1 Replies

5. Solaris

Live upgrade Issue

Hi, I upgraded solaris 10 x86 from update 3 to update 7 with zones installed n UFS file system . The global zone was updated but the non global zone still shows update 3 what could be the reason for this and how can i update the local zones to update 7 (0 Replies)
Discussion started by: fugitive
0 Replies
Login or Register to Ask a Question