09-08-2009
Your server is part of an HACMP cluster that uses heartbeats to detect all nodes are alive. The error messages tells you that one of those heartbeats is not working properly because of high (CPU) load. You said that you use 70% of the server's paging space. If this paging space is on rootvg and if rootvg disks are 100% busy during paging space usage that is the likely cause for heartbeat pakets getting lost. So the second relevant part of the error message is the hint about excessive disk I/O which is causing high memory contention
You don't not to worry as long as at least one heartbeat keeps on working. However make sure that the paging space is not being used frequently! You DB has grown over the time and you may need to get more RAM for the (all) cluster nodes
10 More Discussions You Might Find Interesting
1. AIX
Hi, I'a a new member here.
My company just bought p570 with 8 LPAR (previously we have p650 with 4 LPAR).
Did anyone have procedure how to setup NIM server (NIM LPAR) and how to install other new LPAR to use the NIM server (as client).
Appreciate your help and thank you very much.
Rgds,
David (0 Replies)
Discussion started by: dshg
0 Replies
2. AIX
Hi-
I'm using NIM functionality (AIX5.3) to backup all AIX Servers but some Servers are in the DmZ and many tcpip ports (nfs, ping,etc...) should be open and... it's really a security risks!
As anyone experience about NIM Backup through Firewall?
Which tcpip ports should be open?
Could we... (1 Reply)
Discussion started by: nymus7
1 Replies
3. AIX
I try to migrate a NIM server from one server to another.
I try to do a mksysb on NIM server
restore the NIM server's mksysb to a client through NIM installation
shutdown NIM server
start newly installed client as NIM server
Does anyone do this before? who can give me some suggestion? (1 Reply)
Discussion started by: yanzhang
1 Replies
4. AIX
Can any one help.....
How will do migration through NIM server? (4 Replies)
Discussion started by: AIXlearner
4 Replies
5. AIX
:b:Hi...
I need help to configure nim client on nim server..
can i define aix 5.3.4.0 on aix 5.3.7.0 nim server.. while i m configuring nim client on nim server its getting msg that images not same.. i need to confirm that both version should be same or not..
Thanks.. (5 Replies)
Discussion started by: sumathi.k
5 Replies
6. AIX
Hi All,
Please excuse the possibly naive question but I'm trying to clone/install a new AIX 5.3 LPAR on a p570 from a mksysb image file using nim. Has anyone done this before and if so, what would the exact command look like?
Does it even remotely resemble something like
nim -o... (1 Reply)
Discussion started by: combustables
1 Replies
7. AIX
Guys,
We are planning to upgrade one of our NIM server to AIX 6.1 from 5.3...
Since the server itself is a NIM Server we can't perform it via NIM & I'm choosing to do CD install.. The Install method would be Upgrade installation.
Is there anything special that I need to consider before... (5 Replies)
Discussion started by: kkeng808
5 Replies
8. AIX
Could you please let me know, if it is possible to have NIM server running on one volume group and other applications like oracle running on other volume group. Do we need to have a dedicated server for only AIX NIM server?. I am new to AIX and planning to install NIM server on a test server. which... (3 Replies)
Discussion started by: saikiran_1984
3 Replies
9. AIX
HELLO ALL
i have installed aix 6.2 , and install sysback 6.1 over Nim , and cinfigure it by Nim AND sysback smitty menu with create spot and lppsource and make TSM configration for that, i take image backup(installation image) successfully but when i want to restore this image , the boot cycle... (5 Replies)
Discussion started by: nancy_ghawanmeh
5 Replies
10. AIX
Using nimadm:
nimadm -j nimadmvg -c sap024 -s spot_6100 -l lpp_6100 -d "hdisk1" -Y
Initializing the NIM master.
Initializing NIM client sap024.
0505-205 nimadm: The level of bos.alt_disk_install.rte installed in SPOT
spot_6100 (6.1.3.4) does not match the NIM master's level (7.1.1.2).... (2 Replies)
Discussion started by: sciacca75
2 Replies
cmruncl(1m) cmruncl(1m)
NAME
cmruncl - run a high availability cluster
SYNOPSIS
cmruncl [-f] [-v] [-n node_name...] [-t | -w none]
DESCRIPTION
cmruncl causes all nodes in a configured cluster or all nodes specified to start their cluster daemons and form a new cluster.
To start a cluster, a user must either be superuser(UID=0), or have an access policy of FULL_ADMIN allowed in the cluster configuration
file. See access policy in cmquerycl(1m).
This command should only be run when the cluster is not active on any of the configured nodes. This command verifies the network configu-
ration before causing the nodes to start their cluster daemons. If a cluster is already running on a subset of the nodes, the cmrunnode
command should be used to start the remaining nodes and force them to join the existing cluster.
If node_name is not specified, the cluster daemons will be started on all the nodes in the cluster. All nodes in the cluster must be
available for the cluster to start unless a subset of nodes is specified.
Options
cmruncl supports the following options:
-f Force cluster startup without warning message and continuation prompt that are printed with the -n option.
-v Verbose output will be displayed.
-t Test only. Provide an assessment of the package placement without affecting the current state of the nodes or packages.
The -w option is not required with the -t option as -t does not validate network connectivity, but assumes that all the
nodes can meet any external dependencies such as EMS resources, package subnets, and storage.
-n node_name...
Start the cluster daemon on the specified subset of node(s).
-w none By default network probing is performed to check that the network connectivity is the same as when the cluster was config-
ured. Any anomalies are reported before the cluster daemons are started. The -w none option disables this probing. The
option should only be used if this network configuration is known to be correct from a recent check.
RETURN VALUE
cmruncl returns the following value:
0 Successful completion.
1 Command failed.
EXAMPLES
Run the cluster daemon:
cmruncl
Run the cluster daemons on node1 and node2:
cmruncl -n node1 -n node2
AUTHOR
cmruncl was developed by HP.
SEE ALSO
cmquerycl(1m), cmhaltcl(1m), cmhaltnode(1m), cmrunnode(1m), cmviewcl(1m), cmeval(1m).
Requires Optional Serviceguard Software cmruncl(1m)