Sponsored Content
Operating Systems HP-UX ServiceGuard cluster & volume group failover Post 302253412 by Wotan31 on Friday 31st of October 2008 02:08:20 PM
Old 10-31-2008
Quote:
Originally Posted by vbe
And not the other packages?

??

Is this VG a new VG? and you forgot to add in your <packagename>.cfg file ?
The other packages do not have VG's assigned to them, only this one.

It is a new VG, and it has been added to the packagename.cfg file.

The VG comes from SAN disks that are presented to both nodes. Everything works fine, except for the fact that a manual vg activation has to occur any time we move the package from one node to the other.
 

10 More Discussions You Might Find Interesting

1. HP-UX

HP Serviceguard failover - what doesn't get moved across?

Excuse the basic nature of the question, I've zero experience with regards to this and I'm just looking for a little clarity... When using serviceguard for a failover from one machine to another, what doesn't get taken from one machine to the other? For example, I was told that OS users and... (1 Reply)
Discussion started by: stefcha
1 Replies

2. High Performance Computing

sun Cluster resource group cant failover

I have rcently setup a 4 node cluster running sun cluster 3.2 and I have installed 4 zones on each node. when installing the zones I had to install the zone on all nodes the on the last node do a zlogin -C <zonename> this worked ok. theni I tried to siwitch the zone to node a thei work... (14 Replies)
Discussion started by: lesliek
14 Replies

3. HP-UX

Serviceguard Cluster Questions..

Hi, I'm fairly new to HP-UX. I am trying to find out if it is possible to create a cluster with 3 active nodes and 1 passive node. Everything I have found online so far just describes active/active and active/passive. Is it possible to have 3 active nodes and 1 passive node in a cluster? If... (2 Replies)
Discussion started by: DtbCollumb
2 Replies

4. AIX

backup & restore a volume group

Dear All, I would like to ask about saving & restoring a user defined volume group. i have a user defined volume group, named as datavg. i want to save it & to restore it into different size of physical volume. currently, datavg consist of 4 pv, three fisrt pv size are 100GB & one pv size is... (1 Reply)
Discussion started by: donybangetgitu
1 Replies

5. Solaris

Sun Cluster 3.1 failover

Hi, We have two sun SPARC server in Clustered (Sun Cluster 3.1). For some reason, System 1 failed over to System 2. Where can I find the logs which could tell me the reason for this failover? Thanks (5 Replies)
Discussion started by: Mack1982
5 Replies

6. Gentoo

How to failover the cluster ?

How to failover the cluster ? GNU/Linux By which command, My Linux version 2008 x86_64 x86_64 x86_64 GNU/Linux What are the prerequisites we need to take while failover ? if any Regards (3 Replies)
Discussion started by: sidharthmellam
3 Replies

7. Solaris

Sun cluster 4.0 - zone cluster failover doubt

Hello experts - I am planning to install a Sun cluster 4.0 zone cluster fail-over. few basic doubts. (1) Where should i install the cluster s/w binaries ?. ( global zone or the container zone where i am planning to install the zone fail-over) (2) Or should i perform the installation on... (0 Replies)
Discussion started by: NVA
0 Replies

8. UNIX for Dummies Questions & Answers

How to create a volume group, logical volume group and file system?

hi, I want to create a volume group of 200 GB and then create different file systems on that. please help me out. Its becomes confusing when the PP calculating PP. I don't understand this concept. (2 Replies)
Discussion started by: kamaldev
2 Replies

9. Linux

Volume is mounted on two ServiceGuard nodes

Hey! I'm running a HP ServiceGuard cluster with three nodes. One of the packages was moved (not by me) from one node to another a few weeks ago. I just noticed that one of the volume groups is still mounted on the old node. Oops! When I run df, less space is used on the old node than the new. ... (0 Replies)
Discussion started by: tobiasvl
0 Replies

10. Red Hat

No space in volume group. How to create a file system using existing logical volume

Hello Guys, I want to create a file system dedicated for an application installation. But there is no space in volume group to create a new logical volume. There is enough space in other logical volume which is being mounted on /var. I know we can use that logical volume and create a virtual... (2 Replies)
Discussion started by: vamshigvk475
2 Replies
CRM_MON(8)							  [FIXME: manual]							CRM_MON(8)

NAME
crm_mon - monitor the cluster's status SYNOPSIS
crm_mon [-V] -d -pfilename -h filename crm_mon [-V] [-1|-n|-r] -h filename crm_mon [-V] [-n|-r] -X filename crm_mon [-V] [-n|-r] -c|-1 crm_mon [-V] -i interval crm_mon -? DESCRIPTION
The crm_mon command allows you to monitor your cluster's status and configuration. Its output includes the number of nodes, uname, uuid, status, the resources configured in your cluster, and the current status of each. The output of crm_mon can be displayed at the console or printed into an HTML file. When provided with a cluster configuration file without the status section, crm_mon creates an overview of nodes and resources as specified in the file. OPTIONS
--help, -? Provide help. --verbose, -V Increase the debug output. --interval seconds, -i seconds Determine the update frequency. If -i is not specified, the default of 15 seconds is assumed. --group-by-node, -n Group resources by node. --inactive, -r Display inactive resources. --as-console, -c Display the cluster status on the console. --one-shot, -1 Display the cluster status once on the console then exit (does not use ncurses). --as-html filename, -h filename Write the cluster's status to the specified file. --daemonize, -d Run in the background as a daemon. --pid-file filename, -p filename Specify the daemon's pid file. --xml-file filename, -X filename Specify an XML file containing a cluster configuration and create an overview of the cluster's configuration. EXAMPLES
Display your cluster's status and get an updated listing every 15 seconds: crm_mon Display your cluster's status and get an updated listing after an interval specified by -i. If -i is not given, the default refresh interval of 15 seconds is assumed: crm_mon -i interval[s] Display your cluster's status on the console: crm_mon -c Display your cluster's status on the console just once then exit: crm_mon -1 Display your cluster's status and group resources by node: crm_mon -n Display your cluster's status, group resources by node, and include inactive resources in the list: crm_mon -n -r Write your cluster's status to an HTML file: crm_mon -h filename Run crm_mon as a daemon in the background, specify the daemon's pid file for easier control of the daemon process, and create HTML output. This option allows you to constantly create HTML output that can be easily processed by other monitoring applications: crm_mon -d -p filename -h filename Display the cluster configuration laid out in an existing cluster configuration file (filename), group the resources by node, and include inactive resources. This command can be used for dry-runs of a cluster configuration before rolling it out to a live cluster. crm_mon -r -n -X filename FILES
/var/lib/heartbeat/crm/cib.xml--the CIB (minus status section) on disk. Editing this file directly is strongly discouraged. AUTHOR
crm_mon was written by Andrew Beekhof. [FIXME: source] 07/05/2010 CRM_MON(8)
All times are GMT -4. The time now is 08:11 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy