Volume is mounted on two ServiceGuard nodes


 
Thread Tools Search this Thread
Operating Systems Linux Volume is mounted on two ServiceGuard nodes
# 1  
Old 02-21-2014
Volume is mounted on two ServiceGuard nodes

Hey! I'm running a HP ServiceGuard cluster with three nodes. One of the packages was moved (not by me) from one node to another a few weeks ago.

I just noticed that one of the volume groups is still mounted on the old node. Oops! When I run df, less space is used on the old node than the new.

What's the safest way to get rid of it on the old node with no corruption? Off the top of my head I'm thinking something along the lines of:

Code:
mount -oremount,ro VOLUME
umount VOLUME
vgchange -an VOLUME_GROUP

Any tips? Thanks!
Login or Register to Ask a Question

Previous Thread | Next Thread

9 More Discussions You Might Find Interesting

1. Red Hat

No space in volume group. How to create a file system using existing logical volume

Hello Guys, I want to create a file system dedicated for an application installation. But there is no space in volume group to create a new logical volume. There is enough space in other logical volume which is being mounted on /var. I know we can use that logical volume and create a virtual... (2 Replies)
Discussion started by: vamshigvk475
2 Replies

2. Solaris

In Solaries 10 how to mount multiple volume on same mounted point

Hi , I am completely stuck and not getting any clue to come out this . So looking for help Q : I have salaries 10 in server with that Dell Equallogic storage connected. in dell Equlalogic in i have 70 TB storage . I created 7 volumes 10 TB each . In Solaries 10 i have syslog server i... (1 Reply)
Discussion started by: Roahn Tiwari
1 Replies

3. UNIX for Dummies Questions & Answers

Confusion Regarding Physical Volume,Volume Group,Logical Volume,Physical partition

Hi, I am new to unix. I am working on Red Hat Linux and side by side on AIX also. After reading the concepts of Storage, I am now really confused regarding the terminologies 1)Physical Volume 2)Volume Group 3)Logical Volume 4)Physical Partition Please help me to understand these concepts. (6 Replies)
Discussion started by: kashifsd17
6 Replies

4. HP-UX

MC Serviceguard and WebSphere ND 6.1

I am a Lawson Software consultant assisting a client with a Lawson upgrade. One of our components is WebSphere Application Server ND 6.1. Our client is using MC Serviceguard for failover. When testing failover, my Websphere instance is trying to start, but generating this error, referencing the old... (4 Replies)
Discussion started by: mrvitas
4 Replies

5. AIX

Basic Filesystem / Physical Volume / Logical Volume Check

Hi! Can anyone help me on how I can do a basic check on the Unix filesystems / physical volumes and logical volumes? What items should I check, like where do I look at in smit? Or are there commands that I should execute? I need to do this as I was informed by IBM that there seems to be... (1 Reply)
Discussion started by: chipahoys
1 Replies

6. HP-UX

ServiceGuard cluster & volume group failover

I have a 2-node ServiceGuard cluster. One of the cluster packages has a volume group assigned to it. When I fail the package over to the other node, the volume group does not come up automatically on the other node. I have to manually do a "vgchange -a y vgname" on the node before the package... (5 Replies)
Discussion started by: Wotan31
5 Replies

7. UNIX for Advanced & Expert Users

Ownership problem using a CIFS-mounted volume

Hello, I am trying to use a storage service for backing large amounts (terabytes) of data. The service uses Linux machines and allows mounting of their disks using the CIFS/SMB protocol. I do have the option of using rsync directly over the network without mounting. But in order to... (4 Replies)
Discussion started by: same1290
4 Replies

8. Solaris

How to resize mirror volume in veritas volume manager 3.5 on Solaris 9 OE

Hi all, I have a problem with vxvm volume which is mirror with two disks. when i am try to increase file system, it is throwing an ERROR: can not allocate 5083938 blocks, ERROR: can not able to run vxassist on this volume. Please find a sutable solutions. Thanks and Regards B. Nageswar... (0 Replies)
Discussion started by: nageswarb
0 Replies

9. UNIX for Advanced & Expert Users

LVM - Extending Logical Volume within Volume Group

Hello, I have logical volume group of 50GB, in which I have 2 logical volumes, LogVol01 and LogVol02, both are of 10GB. If I extend LogVol01 further by 10GB, then it keeps the extended copy after logical volume 2. I want to know where it keeps this information Regards Himanshu (3 Replies)
Discussion started by: ghimanshu
3 Replies
Login or Register to Ask a Question
cmruncl(1m)															       cmruncl(1m)

NAME
cmruncl - run a high availability cluster SYNOPSIS
cmruncl [-f] [-v] [-n node_name...] [-t | -w none] DESCRIPTION
cmruncl causes all nodes in a configured cluster or all nodes specified to start their cluster daemons and form a new cluster. To start a cluster, a user must either be superuser(UID=0), or have an access policy of FULL_ADMIN allowed in the cluster configuration file. See access policy in cmquerycl(1m). This command should only be run when the cluster is not active on any of the configured nodes. This command verifies the network configu- ration before causing the nodes to start their cluster daemons. If a cluster is already running on a subset of the nodes, the cmrunnode command should be used to start the remaining nodes and force them to join the existing cluster. If node_name is not specified, the cluster daemons will be started on all the nodes in the cluster. All nodes in the cluster must be available for the cluster to start unless a subset of nodes is specified. Options cmruncl supports the following options: -f Force cluster startup without warning message and continuation prompt that are printed with the -n option. -v Verbose output will be displayed. -t Test only. Provide an assessment of the package placement without affecting the current state of the nodes or packages. The -w option is not required with the -t option as -t does not validate network connectivity, but assumes that all the nodes can meet any external dependencies such as EMS resources, package subnets, and storage. -n node_name... Start the cluster daemon on the specified subset of node(s). -w none By default network probing is performed to check that the network connectivity is the same as when the cluster was config- ured. Any anomalies are reported before the cluster daemons are started. The -w none option disables this probing. The option should only be used if this network configuration is known to be correct from a recent check. RETURN VALUE cmruncl returns the following value: 0 Successful completion. 1 Command failed. EXAMPLES
Run the cluster daemon: cmruncl Run the cluster daemons on node1 and node2: cmruncl -n node1 -n node2 AUTHOR
cmruncl was developed by HP. SEE ALSO
cmquerycl(1m), cmhaltcl(1m), cmhaltnode(1m), cmrunnode(1m), cmviewcl(1m), cmeval(1m). Requires Optional Serviceguard Software cmruncl(1m)