Sponsored Content
Top Forums UNIX for Advanced & Expert Users Supported but not advertised 1000baseT/Full only in one node of RH cluster Post 302346249 by fpmurphy on Friday 21st of August 2009 10:28:36 AM
Old 08-21-2009
Looking at the source code, if a NIC is set in low power mode, it will not autonegotiate 1000baseT.

Suggest you dump the NVRAM of this NIC and a "good" NIC and see if they are the same.
 

10 More Discussions You Might Find Interesting

1. HP-UX

i-node full on /var

Can anyone tell me how would I troubleshoot when /var becomes full with inodes? This is on HP11.11 system. Where used is 92%, ifree is 1891 iuse is 88%. Thanks. (3 Replies)
Discussion started by: catwomen
3 Replies

2. Shell Programming and Scripting

The other node name of a SUN cluster

Hello, Under ksh I have to run a script on one of the nodes of a Solaris 8 cluster which at some time must execute a command on the alternate node: # rsh <name> "command" I have to implement this script on all the clusters of my company (a lot of...). Fortunately, the names of the two nodes... (11 Replies)
Discussion started by: heartwork
11 Replies

3. HP-UX

Node can't join cluster

Need help guys! when running cmrunnode batch i'm getting this error cmrunnode : Waiting for cluster to... (1 Reply)
Discussion started by: Tris
1 Replies

4. HP-UX

MC/SG Fail to join cluster node

Hi, Please advise me whereas I have two node cluster server configured with MC/SG. Application and DB are running on Node 1, while Node 2 is standby. All the volume group devices are part of cluster environment. There is only one package running at node 1. Node 2 is having the problem to... (1 Reply)
Discussion started by: rauphelhunter
1 Replies

5. High Performance Computing

Setting up 2 node cluster using solaris 10

hi, i am trying to setup a 2 node cluster environment. following is what i have; 1. 2 x sun ultra60 - 450MHz procs, 1GB RAM, 9GB HDD, solaris 10 2. 2 x HBA cards 3. 2 x Connection leads to connect ultra60 with D1000 4. 1 x D1000 storage box. 5. 3 x 9GB HDD + 2 x 36GB HDD first of all,... (1 Reply)
Discussion started by: solman17
1 Replies

6. Solaris

Active Sun cluster node?

I now the logical name and Virtual IP of the cluster. How can I find the active sun cluster node without having root access? (3 Replies)
Discussion started by: sreeniatbp
3 Replies

7. HP-UX

Identify cluster active node

Hello, Is there any way to identify the active node in a HP-UX cluster without root privileges? (3 Replies)
Discussion started by: psimoes79
3 Replies

8. Solaris

How to remove single node cluster

Hi Gurus, I am very new to clustering and for test i have created a single node cluster, now i want to remove the system from cluster. Did some googling however as a newbee in cluster unable to co related the info. Please help Thanks (1 Reply)
Discussion started by: kumarmani
1 Replies

9. Solaris

SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node

Hi, Is it possible to have a Solaris cluster of 2 nodes at SITE-A using SVM and creating metaset using say 2 LUNs (on SAN). Then replicating these 2 LUNs to remote site SITE-B via storage based replication and then using these LUNs by importing them as a metaset on a server at SITE-B which is... (0 Replies)
Discussion started by: dn2011
0 Replies

10. AIX

Cluster node not starting

Setting up HACMP 6.1 on a two node cluster. The other node works fine and can start properly on STABLE state (VGs varied, FS mounted, Service IP aliased). However, the other node is always stuck on ST_JOINING state. Its taking forever and you can't stop the cluster as well or recover from script... (2 Replies)
Discussion started by: depam
2 Replies
ccs_tool(8)                                                                                                                            ccs_tool(8)

NAME
ccs_tool - The tool used to make online updates of CCS config files. SYNOPSIS
ccs_tool [OPTION].. <command> DESCRIPTION
ccs_tool is part of the Cluster Configuration System (CCS). It is used to make online updates to cluster.conf. It can also be used to upgrade old style (GFS <= 6.0) CCS archives to the new xml cluster.conf format. OPTIONS
-h Help. Print out the usage. -V Print the version information. sub-commands have their own options, see below for more detail COMMANDS
addnode [options] <node> [<fenceoption=value>]... Adds a new node to the cluster configuration file. Fencing device options are specified as key=value pairs (as many as required) and are entered into the configuration file as is. See the documentation for your fencing agent for more details (eg a powerswitch fence device may need to know which port the node is connected to). Options: -v <votes> Number of votes for this node (mandatory) -n <nodeid> Node id for this node (optional) -i <interface> Network interface to use for this node. Mandatory if the cluster is using multicast as transport. Forbidden if not. -m <multicast> Multicast address for cluster. Only allowed on the first node to be added to the file. Subsequent nodes will use either multicast or broadcast depending on the properties of the first node. -f <fencedevice> Name of fence device to use for this node. The fence device section must already have been added to the file, probably using the addfence command. -c <file> Config file to use. Defaults to /etc/cluster/cluster.conf -o <file> Output file. Defaults to the same as -c -C Don't run "ccs_tool update" after changing file. This will happen by default if the input file is the same as the output file. -F Force a "ccs_tool update" even if the input and output files are different. delnode [options] <node> Delete a node from the cluster configuration file. Note: there is no "edit" command so to change the properties of a node you must delete it and add it back in with the new properties. Options: -c <file> Config file to use. Defaults to /etc/cluster/cluster.conf -o <file> Output file. Defaults to the same as -c -C Don't run "ccs_tool update" after changing file. This will happen by default if the input file is the same as the output file. -F Force a "ccs_tool update" even if the input and output files are different. addfence [options] <name> <agent> [<option>=<value>]... Adds a new fence device section to the cluster configuration file. <agent> is the name of the fence agent that controls the device. the options following are entered as key-value pairs. See the fence agent documentation for details about these. eg: you may need to enter the IP address and username/password for a powerswitch fencing device. Options: -c <file> Config file to use. Defaults to /etc/cluster/cluster.conf -o <file> Output file. Defaults to the same as -c -C Don't run "ccs_tool update" after changing file. This will happen by default if the input file is the same as the output file. -F Force a "ccs_tool update" even if the input and output files are different. delfence [options] <node> Deletes a fencing device from the cluster configuration file. delfence will allow you to remove a fence device that is in use by nodes. This is to allow changes to be made, but be aware that it may produce an invalid configuration file if you don't add it back in again. Options: -c <file> Config file to use. Defaults to /etc/cluster/cluster.conf -o <file> Output file. Defaults to the same as -c -C Don't run "ccs_tool update" after changing file. This will happen by default if the input file is the same as the output file. -F Force a "ccs_tool update" even if the input and output files are different. lsnode [options] List the nodes in the configuration file. This is (hopefully obviously) not necessarily the same as the nodes currently in the clus- ter, but it should be a superset. Options: -v Verbose. Lists all the properties of the node, and the node-specific properties of the fence device too. -c <file> Config file to use. Defaults to /etc/cluster/cluster.conf lsfence [options] List all the fence devices in the cluster configuration file. Options: -v Verbose. Lists all the properties of the fence device rather than just the names and agents. -c <file> Config file to use. Defaults to /etc/cluster/cluster.conf create [options] <clustername> Create a new, skeleton, configuration file. Note that "create" on its own will not create a valid configuration file. Fence agents and nodes will need to be added to it before handing it over to ccsd. The new configuration file will have a version number of 1. Subsequent addnode/delnode/addfence/delfence operations will increment the version number by 1 each time. Options: -c <file> Config file to create. Defaults to /etc/cluster/cluster.conf addnodeids Adds node ID numbers to all the nodes in cluster.conf. In RHEL4, node IDs were optional and assigned by cman when a node joined the cluster. In RHEL5 they must be pre-assigned in cluster.conf. This command will not change any node IDs that are already set in clus- ter.conf, it will simply add unique node ID numbers to nodes that do not already have them. SEE ALSO
cluster.conf(5) ccs_tool(8)
All times are GMT -4. The time now is 10:32 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy