I have Solaris-9 server, V240.
I got alert that one of the interface on IPMP configuration, is failed. Found that two IPs (192.168.120.32 and 192.168.120.35) are not pingable from this server. These two IPs were plumbed on another server and that is decommissioned now. That is the reason, they are not pingable. For immediate fix, I plumbed both these IPs on another server and after that I was able to ping. I have seen this behaviour in other server, so I knew this may be the cause. But even after all IPs are pingable from routing table, I can't remove FAILED flag from ce0 interface.
I ran "pkill -HUP in.mpathd" on one terminal twice and checked /var/adm/messages on another session
Please suggest, what I am missing here and should check ?
Thanks
Last edited by solaris_1977; 09-06-2019 at 02:51 AM..
I have read your post#1 countless times and I must confess that I am at a loss to understand your question. Sorry about that I cannot give you a specific answer as a result.
So what I will do is bash some keys a provide some general network interface information as it pertains to Solaris 9. I apologize if you already know all this but we have to start somewhere. This might be a long post before I'm finished, I don't know, it's just going to be as it comes (into my head).
Why are you seemingly just plumbing missing IP addresses that you can't ping onto another system? With IPMP the same IP address is aggregated across two or more NICs (on the same machine).
If you want to configure IPMP you would do that BEFORE you 'plumb'. For example if you have interfaces bge0 and bge1, you would create an aggregate interface 'aggr1' for example and after that you would plumb and configure only aggr1. You would not try to configure bge0 and bge1 individually any more.
Now Solaris 9 will look for files /etc/hostname.<interface> at boot time and try to plumb those interfaces. If this system was restored from a different hardware platform, then you might for example have a file /etc/hostname.ce0 existing causing Solaris to try to plumb ce0 at boot-time when ce0 doesn't actually exist on this hardware. To stop Solaris from trying to plumb ce0 simply delete the /etc/hostname.ceo file.
When Solaris finds a file /etc/hostname.<interface> at boot-time, it reads the hostname from this file and then (assuming the interface is not configured for DHCP of course) goes to /etc/hosts and looks up the IP address it should use on this interface.
If you aggregate bge0 and bge1 into aggr1, then a file /etc/hostname.aggr1 is created which Solaris will try to plumb at boot-time.
Now, you are trying to get a FAIL message for ce0 to disappear, yes? I can think of only two possibilities why a system would complain about ce0 FAIL:
1. File /etc/hostname.ce0 exists but actual interface ce0 does not exist on this hardware. Delete the file.
2. The interface ce0 does not exist on this platform but is included in an aggregate IPMP configuration that has been restored from a different hardware platform. Down the aggregate interface and delete the IPMP configuration, then recreate the aggregate with interfaces that do exist on this platform and exclude ce0 which doesn't.
Aggregating interfaces has nothing to do with other systems on the LAN. Provided the network cables from the aggregated interfaces go to network switch(es) that understand multi-pathing then all should be well.
I'm going to stop there. If I've completely misunderstood your question then please give us a clue what this is about please.
I am sorry to have confused you. I clubbed two issues in one. I will re-word this issue.
IPMP is already configured on this server. Suddenly I got alert that IPMP group is failed over due to some error. When I logged into the server, I found that ce2 was in FAILED status, instead of the usual INACTIVE state.
/etc/hostname.ce2 file is there and the physical interface is also present. There was never any change in its setup. Physically I can see light blinking on network port behind the server. But since this interface is in FAILED states, IPMP is broken. Running snoop on ce2, is not giving me any result. To test this, I tried to detach bge0 and it is not working
---------------------UPDATE-----------------
Found that cable had problem. After replacing that, I was able to fix this issue
Last edited by solaris_1977; 09-06-2019 at 11:44 PM..
This User Gave Thanks to solaris_1977 For This Post:
Thanks for the update.
If a NIC suddenly fails, and no admin did something to your system or to the LAN switch then the next idea is hardware.
The IPMP concept is quite different from the port aggregation concept.
Does the latter exist in Solaris 9 at all? In the early days you had to purchase SunTrunking software.
These 2 Users Gave Thanks to MadeInGermany For This Post:
@MadeInGermany........................That's an interesting point you make. AFAIR port aggregation was around long before multi-pathing (IPMP) as it's a simpler technology (isn't it??).
I assumed that since this is Solaris 9 we were talking aggregation and, from the posts, it sounded to me that one port going down (perhaps by unplugging the cable) stopped all communication thereby indicating that the other aggregated port was already down.
Perhaps I misunderstood the question in the first place. I had real difficulty getting a handle on it.
Yes, okay, I know that we techies are continuing a thread that's already tagged as solved.
Hi all,
Just a few questions ->
Is an "OFFLINE" interface going back to "ONLINE" consider as a failback by IPMP ?
I have "FAILBACK=no" in my /etc/default/mpathd; however when i do the following
(igb0 and igb7 are in the same ipmp link based group)
q1) why does "if_mpadm -r igb7" cause... (0 Replies)
Hi all,
I been reading examples of how to setup IPMP and how it differs from Etherchannel. However, i am still unsure of how it really works and i hope gurus here can shed some light on the questions I have below while i will lab it up for my own test ->
q1) for IPMP, there is no such thing... (23 Replies)
hi all,
i start with solaris 11 and i am disapointed by the change on ip managing.
i want to set a ipmp over tow aggregate but i dont find any doc and i am lost with the new commande
switch1
net0 aggregate1 |
net1 aggregate1 |-----|
|... (1 Reply)
All.
I am trying to create a 10 branded zone on a Sol 11.1 T5. The Global is using IPMP...so aggregating is out of the question. Has anyone successfully created a branded zone with IPMP? If they have can you please show me the steps you took to get this to run.
Thanks (4 Replies)
Can any one please explain me the concept behind IPMP in solaris clustering.Basic explanation would be really appreciated...
Thanks in Advance
vks (2 Replies)
A problem happened with me, I was configuring IP for two network interfaces, and when I rebooted the system, everything is working but after like 3 or 5 minutes it will tell me that the whole IPMP group has failed !
I tried to troubleshoot, so I found that the gateway is not reachable..so I... (4 Replies)
Does Veritas Cluster work with IPMP on Solaris 10?
If anyone has set it up do you have a doc or tips?
I have heard several different statements ranging from , not working at all to Yes it works! Great How?
* Test and Base IPs????
* configure the MultiNICB agent ?
I can give details... (1 Reply)
Hello All,
I work for a Health care company at a local trauma hospital. I maintain a Picture Archiving and Communication System (PAC's). Basically, any medical images (X-Ray, CT, MRI, Mammo, etc) are stored digitally on the servers for viewing and dictation from diagnostic stations. I took over... (10 Replies)