Hi Guys, having an issue with running Live Upgrade on a t5240 runiing Solaris 10 5/08.
The system has the required patches 121430, and Live upgrade was updated from the install media sol-10-u10-ga2-sparc-dvd.iso
The following boot environments were created
solenv1 and solenv2 with the lucreate command and this completed sucessfully.
Code:
# lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1 yes yes yes no -
solenv2 yes no no yes -
I had an issue when running the lupgrade command, it got to 98% complete
Code:
# luupgrade -u -k /var/tmp/LDMupgrade/autoreg -n solenv2 -s /sol_dvd
64459 blocks
miniroot filesystem is <lofs>
Mounting miniroot at </sol_dvd/Solaris_10/Tools/Boot>
#######################################################################
NOTE: To improve products and services, Oracle Solaris communicates
configuration data to Oracle after rebooting.
You can register your version of Oracle Solaris to capture this data
for your use, or the data is sent anonymously.
For information about what configuration data is communicated and how
to control this facility, see the Release Notes or
Redirecting....
INFORMATION: After activated and booted into new BE <solenv2>,
Auto Registration happens automatically with the following Information
autoreg=disable
#######################################################################
Validating the contents of the media </sol_dvd>.
The media is a standard Solaris media.
The media contains an operating system upgrade image.
The media contains <Solaris> version <10>.
Constructing upgrade profile to use.
Locating the operating system upgrade program.
Checking for existence of previously scheduled Live Upgrade requests.
Creating upgrade profile for BE <solenv2>.
Determining packages to install or upgrade for BE <solenv2>.
Performing the operating system upgrade of the BE <solenv2>.
CAUTION: Interrupting this process may leave the boot environment unstable
or unbootable.
Read from remote host server: Connection reset by peer
Connection to server closed.
Since this happened, lustatus shows the following output
Code:
# lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1 yes yes yes no -
solenv2 yes no no no UPDATING
/etc/lutab shows the following
Code:
# cat /etc/lutab
# DO NOT EDIT THIS FILE BY HAND. This file is not a public interface.
# The format and contents of this file are subject to change.
# Any user modification to this file may result in the incorrect
# operation of Live Upgrade.
1:solenv1:C:0
1:/:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:1
1:boot-device:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:2
2:solenv2:C:0
2:/:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:1
2:boot-device:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:2
What can I do remove the BE solenv2 and start again?
Hi Duke, thanks for that, I had a look at that document, early this morning ( was at the end of an 18 hour session so may not have been focussing properly)
I have just tried it again
Code:
# lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1 yes yes yes no -
solenv2 yes no no no UPDATING
I was looking at the can delete column, which indicate it couldn't be deleted, but when actually trying to run it, I get the following
Code:
# ludelete solenv2
INFORMATION: Removing invalid lock file.
ERROR: mount: The state of /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0 is not okay
and it was attempted to be mounted read/write
mount: Please run fsck and try again
ERROR: cannot mount mount point </.alt.tmp.b-8og.mnt> device </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0>
ERROR: failed to mount file system </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0> on </.alt.tmp.b-8og.mnt>
ERROR: cannot mount boot environment by icf file </etc/lu/ICF.2>
WARNING: Unable to mount ABE <solenv2>: cannot complete lumk_iconf
WARNING: Unable to determine disk partition configuration information for BE <solenv2>.
ERROR: mount: The state of /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0 is not okay
and it was attempted to be mounted read/write
mount: Please run fsck and try again
ERROR: cannot mount mount point </.alt.tmp.b-hpg.mnt> device </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0>
ERROR: failed to mount file system </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0> on </.alt.tmp.b-hpg.mnt>
ERROR: cannot mount boot environment by icf file </etc/lu/ICF.2>
ERROR: Cannot mount BE <solenv2>.
mount: Please run fsck and try again
luupdall: WARNING: Could not mount the Root Slice of BE:"solenv2".
mount: The state of /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0 is not okay
and it was attempted to be mounted read/write
However, output from lustatus has now changed
Code:
# lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1 yes yes yes no -
solenv2 yes no no yes -
Would I be right in thinking that I can perform an lucreate, using the same names again?
you have to delete solenv2 before you can create the environment with the same name again... but there seems to be a problem with the mountpoints for the new environment. they seem to be on a external device (SAN?) which can not be mounted properly (from what i can see in the error messages...).
Hi Duke, yes that is correct the disks are on a SAN.
I have tried running fsck on the specified disk
Code:
# cat /etc/lutab
# DO NOT EDIT THIS FILE BY HAND. This file is not a public interface.
# The format and contents of this file are subject to change.
# Any user modification to this file may result in the incorrect
# operation of Live Upgrade.
1:solenv1:C:0
1:/:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:1
1:boot-device:/dev/dsk/c6t60A9800057396D64685A4D7A51725458d0s0:2
2:solenv2:C:0
2:/:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:1
2:boot-device:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:2
# echo | format
Searching for disks...done
AVAILABLE DISK SELECTIONS:
0. c6t60A9800057396D64685A4D7A51725458d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 3072>
/scsi_vhci/ssd@g60a9800057396d64685a4d7a51725458
1. c6t60A9800057396D64685A4D7A52675733d0 <NETAPP-LUN-7320-500.08GB>
/scsi_vhci/ssd@g60a9800057396d64685a4d7a52675733
2. c6t60A9800057396D64685A51774F6E712Dd0 <NETAPP-LUN-0.2 cyl 5118 alt 2 hd 16 sec 512>
/scsi_vhci/ssd@g60a9800057396d64685a51774f6e712d
3. c6t60A9800057396D64685A62414D324E54d0 <NETAPP-LUN-7320-50.00GB>
/scsi_vhci/ssd@g60a9800057396d64685a62414d324e54
4. c6t60A9800057396D64685A62414D345041d0 <NETAPP-LUN-7320-10.00GB>
/scsi_vhci/ssd@g60a9800057396d64685a62414d345041
5. c6t60A9800057396D6468344D7A4E795678d0 <NETAPP-LUN-0.2 cyl 5118 alt 2 hd 16 sec 512>
/scsi_vhci/ssd@g60a9800057396d6468344d7a4e795678
6. c6t60A9800057396D6468344D7A4F4A3561d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 2048>
/scsi_vhci/ssd@g60a9800057396d6468344d7a4f4a3561
7. c6t60A9800057396D6468344D7A4F356151d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 4096>
/scsi_vhci/ssd@g60a9800057396d6468344d7a4f356151
8. c6t60A9800057396D6468344D7A4F424571d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 1024>
/scsi_vhci/ssd@g60a9800057396d6468344d7a4f424571
9. c6t60A9800057396D6468344D7A4F456848d0 <NETAPP-LUN-0.2 cyl 5118 alt 2 hd 16 sec 256>
/scsi_vhci/ssd@g60a9800057396d6468344d7a4f456848
10. c6t60A9800057396D6468344D7A4F483044d0 <NETAPP-LUN-0.2 cyl 6398 alt 2 hd 16 sec 2048>
/scsi_vhci/ssd@g60a9800057396d6468344d7a4f483044
Specify disk (enter its number): Specify disk (enter its number):
below is the output
Quote:
# fsck -F ufs /dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0
** /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0
** Last Mounted on /a
** Phase 1 - Check Blocks and Sizes
INCORRECT DISK BLOCK COUNT I=400 (416 should be 224)
CORRECT? y
FRAGMENT 49976 DUP I=5828 LFN 0
<snip>
EXCESSIVE DUPLICATE FRAGMENTS I=5828
CONTINUE? y
***** FILE SYSTEM WAS MODIFIED *****
ORPHANED DIRECTORIES REATTACHED; DIR LINK COUNTS MAY NOT BE CORRECT.
***** PLEASE RERUN FSCK *****
I am then able to delete the BE solenv2
Quote:
# ludelete solenv2
Updating boot environment configuration database.
Updating boot environment description database on all BEs.
Updating all boot environment configuration databases.
#
# lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
solenv1 yes yes yes no -
#
So it looks like i'm getting back on track, thanks very much for your input
As an aside, looking at the results form the secondary run of fsck, should I rerun it?
Last edited by Revo; 05-07-2012 at 10:41 AM..
Reason: spelling mistake
as you've deleted the BE i don't think it is neccessary to re-run fsck... just try to create the BE again after creating a new fs (newfs) on the target disk(s).
Hi Duke, reran the lucreate command, and this seemed to work fine
Code:
# newfs /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0
newfs: /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0 last mounted as /.alt.tmp.b-s3b.mnt
newfs: construct a new file system /dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0: (y/n)? y
Warning: 2048 sector(s) in last cylinder unallocated
/dev/rdsk/c6t60A9800057396D6468344D7A4F356151d0s0: 245825536 sectors in 40011 cylinders of 48 tracks, 128 sectors
120032.0MB in 2501 cyl groups (16 c/g, 48.00MB/g, 5824 i/g)
super-block backups (for fsck -F ufs -o b=#) at:
32, 98464, 196896, 295328, 393760, 492192, 590624, 689056, 787488, 885920,
Initializing cylinder groups:
.................................................
super-block backups for last 10 cylinder groups at:
244882848, 244981280, 245079712, 245178144, 245276576, 245366816, 245465248,
245563680, 245662112, 245760544
Code:
# lucreate -c solenv1 -m /:/dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0:ufs -n solenv2
Determining types of file systems supported
Validating file system requests
Preparing logical storage devices
Preparing physical storage devices
Configuring physical storage devices
Configuring logical storage devices
Analyzing system configuration.
Updating boot environment description database on all BEs.
Updating system configuration files.
The device </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0> is not a root device for any boot environment; cannot get BE ID.
Creating configuration for boot environment <solenv2>.
Source boot environment is <solenv1>.
Creating file systems on boot environment <solenv2>.
Creating <ufs> file system for </> in zone <global> on </dev/dsk/c6t60A9800057396D6468344D7A4F356151d0s0>.
Mounting file systems for boot environment <solenv2>.
Calculating required sizes of file systems for boot environment <solenv2>.
Populating file systems on boot environment <solenv2>.
Analyzing zones.
Mounting ABE <solenv2>.
Generating file list.
Copying data from PBE <solenv1> to ABE <solenv2>.
100% of filenames transferred
Finalizing ABE.
Fixing zonepaths in ABE.
Unmounting ABE <solenv2>.
Fixing properties on ZFS datasets in ABE.
Reverting state of zones in PBE <solenv1>.
Making boot environment <solenv2> bootable.
Population of boot environment <solenv2> successful.
Creation of boot environment <solenv2> successful.
Code:
# luupgrade -u -k /var/tmp/LDMupgrade/autoreg -n solenv2 -s /sol_dvd
miniroot filesystem is <lofs>
Mounting miniroot at </sol_dvd/Solaris_10/Tools/Boot>
#######################################################################
NOTE: To improve products and services, Oracle Solaris communicates
configuration data to Oracle after rebooting.
You can register your version of Oracle Solaris to capture this data
for your use, or the data is sent anonymously.
For information about what configuration data is communicated and how
to control this facility, see the Release Notes or
www.oracle.com/goto/solarisautoreg.
INFORMATION: After activated and booted into new BE <solenv2>,
Auto Registration happens automatically with the following Information
autoreg=disable
#######################################################################
Validating the contents of the media </sol_dvd>.
The media is a standard Solaris media.
The media contains an operating system upgrade image.
The media contains <Solaris> version <10>.
Constructing upgrade profile to use.
Locating the operating system upgrade program.
Checking for existence of previously scheduled Live Upgrade requests.
Creating upgrade profile for BE <solenv2>.
Determining packages to install or upgrade for BE <solenv2>.
Performing the operating system upgrade of the BE <solenv2>.
CAUTION: Interrupting this process may leave the boot environment unstable
or unbootable.
May 7 17:33:01 <server> ufs: NOTICE: /a: bad dir ino 1729 at offset 0: mangled entry
Upgrading Solaris: 14% completed
Is the bad dir ino message something to worry about?
I've had a look on google but cant seem to find anything that matches
I have Solaris-10 sparc box with ZFS file-system, which is running two non global zones. I am in process of applying Solaris Recommended patch cluster via Live Upgrade.
Though I have enough space in root file-system of both zones, everytime I run installcluster, it fails with complaining less... (7 Replies)
I tried a live upgrade for one my solaris 10u8 server which didnt go sucessfull and after that i now have following mounts in memory.
df: cannot statvfs /.alt.sol10u8_2/var: No such file or directory
df: cannot statvfs /.alt.sol10u8_2/var/run: No such file or directory
df: cannot statvfs... (0 Replies)
I have created a solaris10 update9 zfs root flash archive for sun4v environment which i 'm tryin to use for upgrading solaris10 update8 zfs root based server using live upgrade.
following is my current system status
lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now... (0 Replies)
I 'm running solaris10 u6 with 141414-02. My system is T5220 running 2 Ldoms and 7 zones on primary domain. I 'm tryin to create booth environment on my alternate root disk after breaking the SVM mirroring but it does not go well and stuck at some point , and i have to reboot the system to get rid... (1 Reply)
Hi,
I upgraded solaris 10 x86 from update 3 to update 7 with zones installed n UFS file system . The global zone was updated but the non global zone still shows update 3 what could be the reason for this and how can i update the local zones to update 7 (0 Replies)