Sponsored Content
Operating Systems Solaris Metastat shows state needs maintenance Post 303037776 by gull04 on Tuesday 13th of August 2019 09:15:29 AM
Old 08-13-2019
HI,

You've stated that the storage team did the migration of IBM Storage to Nimble storage, I'm guessing that that is where the problem lies. Or there could be issues with the vpath software and Nimble storage, does the Solaris version support Nimble?

I would have tackled this a different way;
  1. Added the new disk to the running system as normal.
  2. Extended the metadevice into a four way mirror.
  3. Remove the original disk.
  4. Recreate the metadb.

I'm not sure why you would attempt to do this using SAN replication, the only exception I would make is where you can replicate the device at a block level ensuring tha the boot block and everything else would come over.

You may get away with installing the boot block and re-labeling the disks, but I'd rather adopt the add the disks and extend the mirror approach.

Given the current situation, you may want to try the metarecover options to recover the individual devices.

Regards

Gull04

--- Post updated at 03:15 PM ---

Hi,

Just out of curiosity, have you made any required changes to /etc/vfstab - you don't mention it.

Regards

Gull04

Last edited by gull04; 08-13-2019 at 10:20 AM.. Reason: Additional Information.
This User Gave Thanks to gull04 For This Post:
 

10 More Discussions You Might Find Interesting

1. Solaris

SVM metastat -- needs maintenance

Running Solaris 9 with SVM. I'm not that familiar with it, but metastat output gives "needs maintenance" message on 2 of the mirrors. There are no errors in /var/adm/messages. What do I need to do to fix this error? Thanks. (14 Replies)
Discussion started by: dangral
14 Replies

2. Solaris

Help on metastat

hi all, can someone pls pass on your suggestion? Firs thing I am testing a script which checks for the pattern 'Needs Maintenance' from metastat output and prints some messages in the screen. So i need to simulate an error in mirrored disk for metastat to give this message 'Needs Maintenance'.... (3 Replies)
Discussion started by: srirammad007
3 Replies

3. Solaris

Softpartition State: Errored in Command MetaStat

Hi people, I have on problem when execute the command METASTAT... d60: Soft Partition Device: d10 State: Errored Size: 12582912 blocks (6.0 GB) Someone help me? Thank you very much (4 Replies)
Discussion started by: denisgomes
4 Replies

4. Solaris

both mirrors in needs maintenance state.

Hi, Ii am facing the belwo problem: d50: Mirror Submirror 0: d30 State: Needs maintenance Submirror 1: d40 State: Needs maintenance Pass: 1 Read option: roundrobin (default) Write option: parallel (default) Size: 212176648 blocks (101 GB) d30:... (3 Replies)
Discussion started by: sag71155
3 Replies

5. Solaris

Metastat shows "maintenance" and "last-erred"

Hi All, Sorry to post a problem for my first post but I'm in a bit of a pickle at the minute! I have an Ultra45 connected to a Storedge 3100 series, 2 internal, 2 external disks with a db application running on the external disks. Now everything is working fine and we've had no downtime or... (4 Replies)
Discussion started by: TheSteed
4 Replies

6. Solaris

How to clear maintenance state for apache2 server?

Can any one of you suggest me the method to get apache server in online from maintenance mode. I tried in the following way, but couldn't get that service to online. bash-3.00# svcs -a | grep apache legacy_run 9:51:55 lrc:/etc/rc3_d/S50apache offline 9:51:22... (3 Replies)
Discussion started by: Sesha
3 Replies

7. Solaris

metastat |grep Needs

Dear, Required an script such that : If metastat |grep Needs , results in some output then this command to be executed for the same : opcmsg object=metastat a=OS msg_grp=OpC severity=critical msg_text="Need maintenance for the system $line" With regards, Mjoshi (3 Replies)
Discussion started by: mjoshi87
3 Replies

8. Solaris

svcs command shows the state as disabled

Hi I need to export a directory named /sybase from my solaris machine via NFS. The svcs command shows the state as disabled. Please let me know how to export the directory. Once the directory is exported from the solaris machine it has to be mounted locally in an aix machine. Can some one... (2 Replies)
Discussion started by: newtoaixos
2 Replies

9. AIX

Open firmware state to running state

Hi Admins, I am having a whole system lpar in open firmware state on HMC. How can I bring it to running state ? Let me know. Thanks. (2 Replies)
Discussion started by: snchaudhari2
2 Replies

10. Solaris

SSH : Restarting too quickly, changing state to maintenance

Hi, I'm new to Solaris. I have an issue with ssh service. When I restart the service it exits with an exit status of 0 $svcadm restart svc:/network/ssh:default $echo $? 0 $ However, the service goes into maintenance mode after restart. I'm able to connect even though the service is in... (3 Replies)
Discussion started by: maverick_here
3 Replies
stmsboot(1M)															      stmsboot(1M)

NAME
stmsboot - administration program for the Solaris I/O multipathing feature SYNOPSIS
/usr/sbin/stmsboot [-d | -e | -u | -L | -l controller_number] The Solaris I/O multipathing feature is a multipathing solution for storage devices that is part of the Solaris operating environment. This feature was formerly known as Sun StorEdge Traffic Manager (STMS) or MPxIO. The stmsboot program is an administrative command to manage enumeration of fibre channel devices under Solaris I/O multipathing. Solaris I/O multipathing-enabled devices are enumerated under scsi_vhci(7D), providing multipathing capabilities. Solaris I/O multipathing-disabled devices are enumerated under the physical controller. In the /dev and /devices trees, Solaris I/O multipathing-enabled devices receive new names that indicate that they are under Solaris I/O multipathing control. This means a device will have a different name from its original name (following installation) when it is under Solaris I/O multipathing control. The stmsboot command automatically updates /etc/vfstab and dump configuration to reflect the device names changes when enabling or disabling Solaris I/O multipathing. A reboot is required for changes to take effect. The following options are supported: -e Enables Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports. Following this enabling, you are prompted to reboot. During the reboot, vfstab and the dump configuration will be updated to reflect the device name changes. -d Disables Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports. Following this disabling, you are prompted to reboot. During the reboot, vfstab and the dump configuration will be updated to reflect the device name changes. -u Updates vfstab and the dump configuration after you have manually modified the configuration to have Solaris I/O multipathing enabled or disabled on specific fp(7D) controller ports. This option prompts you to reboot. During the reboot, vfstab and the dump configura- tion will be updated to reflect the device name changes. -L Display the device name changes from non-Solaris I/O multipathing device names to Solaris I/O multipathing device names. -l controller_number Display the device name changes from non-Solaris I/O multipathing device names to Solaris I/O multipathing device names for the speci- fied controller. Along with its primary function of enabling or disabling Solaris I/O multipathing, the stmsboot command is used to update vfstab and the dump configuration to reflect device name changes. For a system to function properly, you must configure the applications that consume the devices by old names to use the new names. The -L and -l options display the mapping between the old and new device names. These options work after the changes made to the Solaris I/O multipathing configuration have taken effect. For example, you can use these options following the reboot after invoking stmsboot -e. The old device names must exist in order to display the mappings. Example 1: Enabling Solaris I/O Multipathing Following OS Upgrade To enable Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports run: # stmsboot -e Example 2: Disabling Solaris I/O Multipathing To disable Solaris I/O multipathing on all fibre channel (fp(7D)) controller ports, run: # stmsboot -d Example 3: Enabling Solaris I/O Multipathing on Selected Ports You want to enable Solaris I/O multipathing on some fibre channel controller ports and disable the feature on the rest. You edit the fp.conf file (see fp(7D)) to enable or disable Solaris I/O multipathing on specific controller ports. You then run the following command to have vfstab and the dump configuration updated to reflect the new device names: # stmsboot -u See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Architecture |SPARC | +-----------------------------+-----------------------------+ |Availability |SUNWcsu, SUNWcslr | +-----------------------------+-----------------------------+ |Interface Stability |Obsolete | +-----------------------------+-----------------------------+ SEE ALSO
dumpadm(1M), ufsdump(1M), dumpdates(4), vfstab(4), fcp(7D), fctl(7D), fp(7D), qlc(7D), scsi_vhci(7D) Consult the Sun StorEdge Disk Tray [or Subsystem] Administrator's Guide for the T3, 3910, 3960, 6120, and 6320 storage subsystems. Sun StorEdge Traffic Manager Installation and Configuration Guide Solaris I/O multipathing is not supported on all devices. After enabling Solaris I/O multipathing, only those devices that Solaris I/O mul- tipathing supports are placed under Solaris I/O multipathing control. Non-supported devices remain as before. For Solaris releases prior to the current release, the -e and -d options remove the mpxio-disable property entries from fp.conf file (see fp(7D)) and add a global mpxio-disable entry to fp.conf. The current release of the Solaris operating system does not support the mpxio-disable property. Solaris I/O multipathing is always enabled. If you want to disable multipathing, you must use the mechanisms provided by the HBA drivers. See fp(7D). Enabling Solaris I/O Multipathing on a Sun StorEdge Disk Array The following applies to Sun StoreEdge T3, 3910, 3960, 6120, and 6320 storage subsystems. To place your Sun StorEdge disk subsystem under Solaris I/O multipathing control, in addition to enabling Solaris I/O multipathing, the mp_support of the subsystem must be set to mpxio mode. The preferred sequence is to change the subsystem's mp_support to mpxio mode, then run stmsboot -e. If Solaris I/O multipathing is already enabled but the subsystem's mp_support is not in mpxio mode, then change the mp_support to mpxio mode and run stmsboot -u. Refer to the Sun StorEdge Administrator's Guide for your subsystem for more details. ufsdump Users The ufsdump command keeps records of the filesystem dumps in /etc/dumpdates (see dumpdates(4)). Among other items, the records contain device names. An effect of the "active" stmsboot options (-e, -d, and -u) is to change the device name of a storage device. The stmsboot command does not modify the dumpdates file. Because of this, the dumpdates records will refer to the old device names, that is, the device names that were in effect before you ran stmsboot. The effect of this device name-dumpdates disagreement is that, following use of stms- boot, ufsdump will be processed as if no previous dump had ever been made, thus dumping the entire filesystem (effectively, a level 0 dump). Procedure to Use stmsboot in Sun Cluster Environment If possible, use stmsboot -e before you start installing Sun Cluster software. After you run stmsboot, you install Sun Cluster software as you normally would. If you install Sun Cluster software before running stmsboot, you must use the following procedure. On each machine in the cluster on which you want to enable the Solaris multipathing feature, enter: # stmsboot -e ...and allow the system to reboot. When the system comes up, enter the following two commands: 1. # /usr/cluster/bin/scdidadm -C 2. # /usr/cluster/bin/scdidadm -r The preceding commands update did mappings with new device names while preserving did instance numbers for disks that are connected to multiple cluster nodes. did instance numbers of the local disks might not be preserved. For this reason, the did disk names for local disks might change. 3. Update /etc/vfstab to reflect any new did disk names for your local disks. 4. Reboot the system. To disable the Solaris multipathing feature, use stmsboot -d (instead of stmsboot -e), then follow the procedure above. To view mappings between the old and new device names, run stmsboot -L. To view did device name mappings, run /usr/cluster/bin/scdidadm -L. 3 Mar 2005 stmsboot(1M)
All times are GMT -4. The time now is 08:52 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy