Sponsored Content
Operating Systems Solaris Solaris 10 branded zone with IPMP Post 302872513 by aeroforce on Friday 8th of November 2013 01:34:22 PM
Old 11-08-2013
Solaris 10 branded zone with IPMP

All.

I am trying to create a 10 branded zone on a Sol 11.1 T5. The Global is using IPMP...so aggregating is out of the question. Has anyone successfully created a branded zone with IPMP? If they have can you please show me the steps you took to get this to run.

Thanks
 

10 More Discussions You Might Find Interesting

1. Solaris

Postprocess issue installing Solaris 9 Branded Zone

Hello, I am attempting to install a newly configured Solaris 9 zone using the flash archive image provided on the sun website. Each time I run the install I get a failure during the Postprocess: Creating mount points. Following is the log: Log File: /var/tmp/s9-zone.install.25457.log... (0 Replies)
Discussion started by: brialt1
0 Replies

2. Solaris

Three questions - migrating to branded zone

1. i read that VxFS (3.x) installed on a solaris 8 will not work on a branded zone. fair enough, but what is the impact of having the packages installed when the flarcreate is done? will the container fail to boot with those packages installed? 2. NTP client and server doesnt work on branded... (11 Replies)
Discussion started by: frustin
11 Replies

3. Solaris

Prerequisites for Solaris10 Branded Zone within Solaris 11 Express

Dear all, we are trying to install a Solaris10 Branded Zone on a Solaris 11 Express system getting this message: ***The image patch level is downrev for running in a solaris10 branded zone*** The Sparc Solaris 10 system is running Release 11/06 and Patch 119254-70 is installed. Does the... (0 Replies)
Discussion started by: xunil321
0 Replies

4. Solaris

Link Based IPMP on Shared IP Solaris Zone

Hi, This may have already been raised previously so sorry for the duplication. What I want to achieve is have a physical server using link based IPMP setup in the global zone (not problem doing that) and then create a zone set as Shared-IP so when the servers NIC has an issue the IP will... (0 Replies)
Discussion started by: giles.cardew
0 Replies

5. Solaris

Solaris 9 branded zone on a Solaris 10 server

I have a T5220 running Solaris 10. I have a application that can only run on Solaris 9. I want to build a branded Solaris 9 zone for this application on the Solaris 10 Server. What steps would I need to accomplish this? (2 Replies)
Discussion started by: soupbone38
2 Replies

6. Solaris

Branded Solaris 10 zone no network

I am trying to create a branded 10 zone on a Solaris 11.1 control domain. I am using a flar image (cpio) from an existing LDOM. Here are the steps taken after flarcreate. #zonecfg -z <device> >create -b >set brand=solaris10 >set zonepath=/zonez/<device> >set ip-type=exclusive >add net... (1 Reply)
Discussion started by: aeroforce
1 Replies

7. Solaris

NFS server on Solaris 10 branded zone

I am trying to get a NFS share on a Solaris 10 branded zone. I have read that this is not possible...but I was hoping someone here could tell me different. share -F nfs -o ro,anon=0 /dir This would really surprise me that you cannot do this. Thanks (2 Replies)
Discussion started by: aeroforce
2 Replies

8. Solaris

Branded zone Solaris 9

Hi, I want to install branded zone sol 9 in sol 10. I keep getting same error. Please advice me. Thanks. (12 Replies)
Discussion started by: mzainal
12 Replies

9. Solaris

Solaris 11 Global zone patching having Solaris 10 branded zone

I am planning to do solaris 11 global zone patching having solaris 10 branded zone. I have a doubts on step 8 specially Can someone clear my step 8 doubts or if anything wrong between step 1 to step 9 please correct that also as I have pretty good idea about Step 10 mean patching in solaris 10... (2 Replies)
Discussion started by: amity
2 Replies

10. Solaris

Help setting up a bhyve branded zone

So I started with following the example on the omnios page for bhyve (sorry, I can't provide the link because I'm a post shy of the 5 to give a URL): And found that in the end I was able to create a running bhyve zone, but accessing it via zlogin results in a blank console. I've tried Debian 9... (14 Replies)
Discussion started by: stratacast1
14 Replies
scstat(1M)						  System Administration Commands						scstat(1M)

NAME
scstat - monitor the status of a Sun Cluster configuration SYNOPSIS
scstat [-DWginpqv [v]] [-h node] DESCRIPTION
Note - Beginning with the Sun Cluster 3.2 release, Sun Cluster software includes an object-oriented command set. Although Sun Cluster software still supports the original command set, Sun Cluster procedural documentation uses only the object-oriented command set. For more infor- mation about the object-oriented command set, see the Intro(1CL) man page. The scstat command displays the current state of Sun Cluster components. Only one instance of the scstat command needs to run on any machine in the Sun Cluster configuration. When run without any options, scstat displays the status for all components of the cluster. This display includes the following informa- tion: o A list of cluster members o The status of each cluster member o The status of resource groups and resources o The status of every path on the cluster interconnect o The status of every disk device group o The status of every quorum device o The status of every IP network multipathing (IPMP) group and public network adapter From a non-global zone, referred to simply as a zone, you can run all forms of this command except the -i option. When you run the scstat command from a non-global zone, the output is the same as when run from the global zone except that no status information is displayed for IP network multipathing groups or public network adapters. You need solaris.cluster.device.read, solaris.cluster.transport.read, solaris.cluster.resource.read, solaris.cluster.node.read, solaris.cluster.quorum.read, and solaris.cluster.system.read RBAC authorization to use this command without options. See rbac(5). Resources and Resource Groups The resource state, resource group state, and resource status are all maintained on a per-node basis. For example, a given resource has a distinct state on each cluster node and a distinct status on each cluster node. The resource state is set by the Resource Group Manager (RGM) on each node, based only on which methods have been invoked on the resource. For example, after the STOP method has run successfully on a resource on a given node, the resource's state will be OFFLINE on that node. If the STOP method exits nonzero or times out, then the state of the resource is Stop_failed. Possible resource states include: Online, Offline, Start_failed, Stop_failed, Monitor_failed, Online_not_monitored, Starting, and Stopping. Possible resource group states are: Unmanaged, Online, Offline, Pending_online, Pending_offline, Error_stop_failed, Online_faulted, and Pending_online_blocked. In addition to resource state, the RGM also maintains a resource status that can be set by the resource itself by using the API. The field Status Message actually consists of two components: status keyword and status message. Status message is optionally set by the resource and is an arbitrary text string that is printed after the status keyword. Descriptions of possible values for a resource's status are as follows: DEGRADED The resource is online, but its performance or availability might be compromised in some way. FAULTED The resource has encountered an error that prevents it from functioning. OFFLINE The resource is offline. ONLINE The resource is online and providing service. UNKNOWN The current status is unknown or is in transition. Device Groups Device group status reflects the availability of the devices in that group. The following are possible values for device group status and their descriptions: DEGRADED The device group is online, but not all of its potential primaries (secondaries) are up. For two-node connectivity, this status basically indicates that a stand-by primary does not exist, which means a failure of the primary node will result in a loss of access to the devices in the group. OFFLINE The device group is offline. There is no primary node. The device group must be brought online before any of its devices can be used. ONLINE The device group is online. There is a primary node, and devices within the group are ready for I/O. WAIT The device group is between one status and another. This status might occur, for example, when a device group is going from offline to online. IP Network Multipathing Groups IP network multipathing (IPMP) group status reflects the availability of the backup group and the adapters in the group. The following are possible values for IPMP group status and their descriptions: OFFLINE The backup group failed. All adapters in the group are offline. ONLINE The backup group is functional. At least one adapter in the group is online. UNKNOWN Any other state than those listed before. This could result when an adapter is detached or marked as down by Solaris commands such as if_mpadm or ifconfig. The following are possible values for IPMP adapter status and their descriptions: OFFLINE The adapter failed or the backup group is offline. ONLINE The adapter is functional. STANDBY The adapter is on standby. UNKNOWN Any other state than those listed before. This could result when an adapter is detached or marked as down by Solaris commands such as if_mpadm or ifconfig. OPTIONS
You can specify command options to request the status for specific components. If more than one option is specified, the scstat command prints the status in the specified order. The following options are supported: -D Shows status for all disk device groups. You can use this option in the global zone or in a non-global zone. For ease of administration, use this form of the command in the global zone. Output is the same when run from a zone as when run from the global zone. You need solaris.cluster.device.read RBAC authorization to use this command option. See rbac(5). -g Shows status for all resource groups. You can use this option in the global zone or in a non-global zone. For ease of administration, use this form of the command in the global zone. Output is the same when run from a zone as when run from the global zone. You need solaris.cluster.resource.read RBAC authorization to use this command option. See rbac(5). -h node Shows status for the specified node (node) and status of the disk device groups of which this node is the primary node. Also shows the status of the quorum devices to which this node holds reservations of the resource groups to which the node is a potential master, and holds reservations of the transport paths to which the node is attached. You need solaris.cluster.device.read, solaris.cluster.transport.read, solaris.cluster.resource.read, solaris.clus- ter.node.read, solaris.cluster.quorum.read, and solaris.cluster.system.read RBAC authorization to use this command option. See rbac(5). -i Shows status for all IPMP groups and public network adapters. You can use this option only in the global zone. -n Shows status for all nodes. You can use this option in the global zone or in a non-global zone. For ease of administration, use this form of the command in the global zone. Output is the same when run from a zone as when run from the global zone. You need solaris.cluster.node.read RBAC authorization to use this command option. See rbac(5). -p Shows status for all components in the cluster. Use with -v to display more verbose output. You can use this option in the global zone or in a non-global zone. For ease of administration, use this form of the command in the global zone. Output is the same when run from a zone as when run from the global zone, except that no status for IPMP groups or public net- work adapters is displayed. You need solaris.cluster.device.read, solaris.cluster.transport.read, solaris.cluster.resource.read, solaris.clus- ter.node.read, solaris.cluster.quorum.read, and solaris.cluster.system.read RBAC authorization to use -p with -v. See rbac(5). -q Shows status for all device quorums and node quorums. You can use this option in the global zone or in a non-global zone. For ease of administration, use this form of the command in the global zone. Output is the same when run from a zone as when run from the global zone. You need solaris.cluster.quorum.read RBAC authorization to use this command option. See rbac(5). -v[v] Shows verbose output. -W Shows status for cluster transport path. You can use this option in the global zone or in a non-global zone. For ease of administration, use this form of the command in the global zone. Output is the same when run from a zone as when run from the global zone. You need solaris.cluster.transport.read RBAC authorization to use this command option. See rbac(5). EXAMPLES
Example 1 Using the scstat Command The following command displays the status of all resource groups followed by the status of all components related to the specified host: % scstat -g -h host The output that is displayed appears in the order in which the options are specified. These results are the same results you would see by typing the two commands: % scstat -g and % scstat -h host EXIT STATUS
The following exit values are returned: 0 The command completed successfully. nonzero An error has occurred. ATTRIBUTES
See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Availability |SUNWsczu | +-----------------------------+-----------------------------+ |Interface Stability |Evolving | +-----------------------------+-----------------------------+ SEE ALSO
Intro(1CL), cluster(1CL), if_mpadm(1M), ifconfig(1M), scha_resource_setstatus(1HA), scha_resource_setstatus(3HA), attributes(5) NOTES
An online quorum device means that the device was available for contributing to the formation of quorum when quorum was last established. From the context of the quorum algorithm, the device is online because it actively contributed to the formation of quorum. However, an online quorum device might not necessarily continue to be in a healthy enough state to contribute to the formation of quorum when quorum is re-established. The current version of Sun Cluster does not include a disk monitoring facility or regular probes to the quorum devices. Sun Cluster 3.2 10 Jul 2006 scstat(1M)
All times are GMT -4. The time now is 05:08 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy