Sponsored Content
Full Discussion: VCS - 3 node - IP change
Operating Systems Solaris VCS - 3 node - IP change Post 302921015 by snchaudhari2 on Tuesday 14th of October 2014 08:49:40 AM
Old 10-14-2014
VCS - 3 node - IP change

Hey Admins,

I need steps to do an IP changes on 3-node VCS cluster. All IP's (public, cluster IP's, backup IP, etc.)want to change. We have a downtime scheduled. Please let me know the steps to do so.

Thanks in advance.
 

9 More Discussions You Might Find Interesting

1. High Performance Computing

What does the 'probe' do in VCS?

Weird question I know, but I'm intrigued. Say for instance you have an Application in an SG with the usual start/stop/monitor. To successfully probe does it just check for existence of the script/file ? Thanks... (1 Reply)
Discussion started by: itsupplies
1 Replies

2. Solaris

SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node

Hi, Is it possible to have a Solaris cluster of 2 nodes at SITE-A using SVM and creating metaset using say 2 LUNs (on SAN). Then replicating these 2 LUNs to remote site SITE-B via storage based replication and then using these LUNs by importing them as a metaset on a server at SITE-B which is... (0 Replies)
Discussion started by: dn2011
0 Replies

3. Solaris

VCS Clusters

:)Hi, can someone please explain VCS clustering and where do we need VCS clusters ..? :o:)Thanks in advance :o:) (1 Reply)
Discussion started by: amitbisht9
1 Replies

4. Solaris

VCS on Solaris: VCS ERROR V-16-2-13077 (host2) Agent is unable to offline resource(DiskReservation)

hi, dear all I get a problem "VCS ERROR V-16-2-13077 " on VCS 4.1 for Solaris 10. I can not offline the host2 when the raid is bad. I don't know the reason and how to offline host2 and switch to host1. please help me, thank you! the message of engine_A.log is : ... (2 Replies)
Discussion started by: ForgetChen
2 Replies

5. UNIX for Advanced & Expert Users

VCS triggerring panic on 1 node, root disk under SVM

We have two node cluster with OS disk mirrored under SVM. There is slight disk problem on one of the mirror disk causing cluster to panic. Failure of one mirror disk causing VCS to panic the node. Why VCS is not able to write /var filesystem, as one of the disk is healthy. ... (1 Reply)
Discussion started by: amlanroy
1 Replies

6. Homework & Coursework Questions

Accessing one UNIX node from another node of the same server

Hi Experts, I am in need of running a script from one node say node 1 via node 2. My scheduling tool dont have access to node2 , so i need to invoke the list file from node1 but the script needs to run from node2. because the server to which i am hitting, is having access only for the node... (5 Replies)
Discussion started by: arun1377
5 Replies

7. HP-UX

Mount FIle systems from node-1 onto node-2

Hi, We have HP UX service guard cluster on OS 11.23. Recently 40+ LUNs presented to both nodes by SAN team but I was asked to mount them on only one node. I created required VGs/LVs, created VxFS and mounted all of them and they are working fine. Now client requested those FS on 2nd node as... (4 Replies)
Discussion started by: prvnrk
4 Replies

8. AIX

Change the ID of a SVC Node

Hi Folks, I'm currently in a position where I am building the seed equipment for a Data Centre Migration, I'm familiar with some of the equipment - all of which I've listed below. SAN Storage - EMC VNX5800 SAN Switches - Brocade DCX 8510-4 SAN Management - IBM SVC 2145-DH8 four node cluster... (2 Replies)
Discussion started by: gull04
2 Replies

9. Solaris

VCS not active

Dear All, My current system run Solaris 9 SPARC 64 bit , and VXVM , VCS 4.1. I check all the node information . ls -l /etc/vx/*.exclude /etc/vx/*.exclude: No such file or directory root@devuardbs01 # vxdctl license All features are available: Mirroring Root Mirroring Concatenation... (0 Replies)
Discussion started by: linux_user
0 Replies
VOTEQUORUM_LEAVING(3)				    Corosync Cluster Engine Programmer's Manual 			     VOTEQUORUM_LEAVING(3)

NAME
votequorum_leaving - Tell other nodes that we are leaving the cluster SYNOPSIS
#include <corosync/votequorum.h> int votequorum_leaving(votequorum_handle_t handle); DESCRIPTION
The votequorum_leaving function is used to tell the other nodes in the cluster that this node is leaving. They will (when the node actually leaves) reduce quorum to keep the cluster running without this node. This function should only be called if it is known that the node is being shut down for a known reason and could be out of the cluster for an extended period of time. Normal behaviour is for the cluster to reduce the total number of votes, but NOT expected_votes when a node leave the cluster, so the clus- ter could become inquorate. This is correct behaviour and is ther eto prevent split-brain. Do NOT call this function unless you know what you are doing. RETURN VALUE
This call returns the CS_OK value if successful, otherwise an error is returned. ERRORS
The errors are undocumented. SEE ALSO
votequorum_overview(8), votequorum_initialize(3), votequorum_finalize(3), votequorum_dispatch(3), votequorum_fd_get(3), corosync Man Page 2009-01-26 VOTEQUORUM_LEAVING(3)
All times are GMT -4. The time now is 07:14 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy