Bonding, IP alias, Virtual IP, 2 nodes cluster


 
Thread Tools Search this Thread
Special Forums UNIX and Linux Applications High Performance Computing Bonding, IP alias, Virtual IP, 2 nodes cluster
# 1  
Old 07-28-2008
Bonding, IP alias, Virtual IP, 2 nodes cluster

Hi ! I have a simple setup of 2 PC (with linux Red-Hat) where the first PC is the primary machine and the second the backup. I use DRBD for data replication and Red-Hat cluster suite for HA (High Availability). I have tested both.

Now I NEED a COMMON IP ADDRESS (or Master/unique IP address) for my two system. When the primary machine fails (or shutdown), the backup system will do the job. From the external world, client who connect to my cluster will "connect to one address" or virtual IP and won't know if this is the primary system or the backup.

For now, my 2 machines have the same /etc/hosts file:

which contains: 10.60.1.23 PrimaryMachine
10.60.1.77 BackupMachine

Maybe I need to change the IP address for something local (192.168.1.0), I don't know. Theses address where created dynamically I think with DHCP when I installed Red-Hat on both machine.

I red many document about BONDING, IP ALIAS and VIRTUAL IP but now I am LOST !!!

What I do really need (keep it as simple as possible) and how to do it ???

Thank you very much.
Login or Register to Ask a Question

Previous Thread | Next Thread

10 More Discussions You Might Find Interesting

1. Red Hat

RedHat Cluster: Nodes won't see each other

Hi All; I try to build a Redhat Cluster (CentOS 6) on vmware. But each node sees the other down like: # clustat Cluster Status for mycluster @ Wed Apr 8 11:01:38 2015 Member Status: Quorate Member Name ID Status ------ ---- ... (1 Reply)
Discussion started by: Meacham12
1 Replies

2. Red Hat

RedHat Cluster: Nodes won't see each other

Hi All; I try to build a Redhat Cluster (CentOS 6) on vmware. But each node sees the other down like: # clustat Cluster Status for mycluster @ Wed Apr 8 11:01:38 2015 Member Status: Quorate Member Name ID Status ------ ---- ... (0 Replies)
Discussion started by: Meacham12
0 Replies

3. UNIX for Advanced & Expert Users

Arbitrator for 2 nodes ocfs cluster

Is there any way to create a arbitrary node for ocfs2 on a virtual machine (others are physical servers) so it won't go panic when one of physical server goes down? This is for load balanced application servers. Any setting example or tips? Thanks. (0 Replies)
Discussion started by: malayo
0 Replies

4. AIX

Re-cluster 2 HACMP 5.2 nodes

Hi, A customer I'm supporting once upon a time broke their 2 cluster node database servers so they could use the 2nd standby node for something else. Now sometime later they want to bring the 2nd node back into the cluster for resilance. Problem is there are now 3 VG's that have been set-up... (1 Reply)
Discussion started by: elcounto
1 Replies

5. Red Hat

How to troubleshoot a 1000 nodes Apache cluster?

Hi all. May I get some expert advice on troubleshooting performance issues of a 1000 nodes Apache LB cluster. Users report slow loading/response of webpages. Different websites are hosted on this cluster for different clients. But all are reporting the same issue. Could you please let me know... (1 Reply)
Discussion started by: admin_xor
1 Replies

6. Solaris

What is the procedure to reboot cluster nodes

Hi we have 2 solaris 10 servers in veritas cluster. also we have oracle cluster on the database end. now we have a requirement to reboot both the servers as it has been running for more than a year. can any one tell what is the procedure to bring down the cluster services in both the nodes... (7 Replies)
Discussion started by: newtoaixos
7 Replies

7. Solaris

Need advise on setting up solaris 10 2 nodes cluster

I am new to setting up sun solaris 10 cluster, I have 2 sun sparc t3-1 servers (identical), going to use them as web servers (sun one java web server 7), looking for data replication and real time fail over. My question is do I need external storage to configure the cluster? or I can just use... (3 Replies)
Discussion started by: spitfire2011
3 Replies

8. Red Hat

Centos/rhel 5 cluster 3 nodes with out Quorum

Hi all, i have 3 nodes cluster (Centos 5 cluster suit) with out quorum disk, node vote = 1, the value of a quorum = 2, when 2 nodes going offline, cluster services are destoys. How i can save the cluster and all services(move all services to one alive node) with out quorum disk when other... (3 Replies)
Discussion started by: Flomaster
3 Replies

9. Emergency UNIX and Linux Support

Rebooting 3 to 1 Cluster nodes.

hello Gurus, My current set up is 3 to 1 Cluster (SUN Cluster 3.2) running oracle database. Task is to reboot the servers. My query is about the procedure to do the same. My understanding is suspend the databases to avoid switchover. Then execute the command scshutdown to down the cluster... (4 Replies)
Discussion started by: EmbedUX
4 Replies

10. UNIX for Dummies Questions & Answers

IP Alias, Bonding or Virtual IP, 2 nodes Cluster, which one to use ?

Hi ! I have a simple setup of 2 PC (with linux Red-Hat) where the first PC is the primary machine and the second the backup. I use DRBD for data replication and Red-Hat cluster suite for HA (High Availability). I have tested both. Now I NEED a COMMON IP ADDRESS (or Master/unique IP address) for... (3 Replies)
Discussion started by: Danny Gilbert
3 Replies
Login or Register to Ask a Question
clusvcadm(8)						       Red Hat Cluster Suite						      clusvcadm(8)

NAME
clusvcadm - Cluster User Service Administration Utility SYNOPSIS
clusvcadm [-d <service> ] [-e <service> ] [-l] [-u] [-S] [-m <member> ] [-r <service> ] [-R <service> ] [-M <service> ] [-Z <service> ] [-U <service> ] [-s <service> ] [-v] DESCRIPTION
The clusvcadm command allows an administrator to enable, disable, relocate, and restart user services in a cluster. In order to perform cluster service operations, the cluster daemons must be running (and have quorum) on the member system on which the command is invoked. OPTIONS
-d <service> Stops and disables the user service named service -e <service> Enables and starts the user service named service -l Lock services in preparation for cluster shutdown. This should only be used if the administrator intends to perform a global, clus- ter wide shutdown. This prevents services from starting (but not stopping, like -Z does). Once the cluster quorum is dissolved, this state is reset. If a new instance of rgmanager boots while others are locked, the behavior is undefined. -u Unlock resource group managers. This allows services to start again. -S Display whether the resource group managers are locked or not. This can be used to verify the correct operation of the -l and -u options, but is only useful for debugging. -m <member> When used in conjunction with either the -e or -r options, this specifies the preferred target member on which to start the service. -r <service> Relocates the user service named service to another cluster member. -R <service> Restarts the user service named service on the cluster member on which it is currently running. -M <service> Use a special "migration" operation to move the user service named service to another cluster member. Currently, this is only use- ful for virtual machines. Use of migration usually requires special configuration of the local virtual machine manager in order to work correctly. -Z <service> Freezes the service named service on the cluster member on which it is currently running. This will prevent status checks of the service as well as failover in the event the node fails or rgmanager is stopped. -U <service> Unfreezes the user service named service on the cluster member on which it is currently running. This will re-enable status checks. -s <service> Stops the service named service until a member transition or until it is enabled again. -v Display version information and exit. NOTES
Executing -U (unfreeze) on a service which was frozen in the started state while the service owner is offline results in an undefined (and possibly dangerous) condition. Manually ensure all resources are clear before doing this. SEE ALSO
clustat(8) Jan 2005 clusvcadm(8)