My company have encountered similar problems, and we have seen that some settings need to be set.
Here are the settings which have to be implemented.
Each child fibre device (fscsiX) has to have the following two modes set:
Additionally, every hdisk device needs to be changed (which I didn't see mentioned in the post).
Lastly, you may want to check that the hcheck_interval is NOT set to 0, as then it won't check at all. Usual recommendation is to set to 30 (but 10 should be sufficient).
Moderator's Comments:
Use code tags please.
Last edited by zaxxon; 01-17-2011 at 12:02 PM..
Reason: code tags
UPDATE: Sorry: The hcheck_interval idea was already mentioned by smurphy. I should have moved on to page 2.
One other thing to check is your "hcheck_interval" which is set at the disk level. The hcheck_interval tells your system how often to check, or re-check, FAILED paths and inactive ENABLED paths (in the case of "algorithm" being set to "fail_over") to ensure they are still connected and functioning. I suggest setting your hcheck_interval to 3600 (once an hour). You'll have to set this on all your disks individually. If the hcheck_interval is set to "0", then this disables it and the disk will never automatically change out of a FAILED or MISSING state.
Remember that MPIO is not like etherchannels, where it automatically re-enables all the paths as soon as the plug is back in. Something has to occur on the disk side to make it recheck them. Either the hcheck_interval comes around again, or you unplug your secondary fiber car which will cause AIX to suddenly start sending checks for all your disks down all the paths, FAILED or MISSING, and try to find a path that is working and it will set it back to ENABLED if it finds one.
Also, you can re-enable the paths manually by doing a chdev on it:
You can also see which path is being used by watching for numbers increasing in the output of "iostat -m":
Hello,
I have some concerns over the disk management of my AIX system.
For example server1
/ > lspv
hdisk0 00fa6d1288c820aa rootvg active
hdisk1 00fa6d1288c8213c vg_2 active
hdisk2 00c1cc14d6de272b ... (6 Replies)
This is getting very confusing for me, and appreciate if someone can help.
Platform: Power VM ( Virtual I/O Server)
ioslevel 2.1.3.10-FP23
# oslevel -s
6100-05-00-0000
Storage: IBM DS4300
Two HBAs - Dual Port Fibre Adapter Channels
Each card has two ports , so a total of 4 ports going... (3 Replies)
Dear Solaris Experts,
We are upgrading from sun4u to T4 systems and one proposal is to use LDOMs and also zones within LDOMs.
Someone advised using only zones and not LDOMs because the new machines have fewer chips and if a chip or a core fails then it doesn't impact the zones, but impacts... (3 Replies)
On a particular LPAR, I was running AIX 5.3 TL 3. On Monday I did an update of the LPAR to 5.3 TL 9 SP2. The install was smooth, but then I ran into a problem.
The MPIO driver does not work with LSI's StoreAge (SVM4). I did some looking, and looks like
5.3 TL3 = IBM.MPIO 5.3.0.30
5.3... (0 Replies)
Hi. I am IT manager/developer for a small organization. I have been doing as-needed linux server administration for several years and am by no means an expert. I've built several of my own servers, and our org is currently using hosting services for our servers and I am relatively happy.
We... (3 Replies)
Hi folks,
does anybody have a link to a documentation how to implement native MPIO on AIX? We are using EMC PowerPath and Datacore SanSymphony/Cambex for this so far and I wasn't able to find a good description on that topic. All I know so far is that mkpath, chpath and lspath are used to... (3 Replies)
We are looking at running MPIO for it's redundancy and load balancing benefits. Does anyone know what pieces of software or modules are needed on the VIO server to get load balancing to work. Remember we are using EMC's DMX3500 storage system. We no longer want to use Powerpath. :rolleyes: ... (2 Replies)
My product have around 10-15 programs/services running in the sun box, which together completes a task, sequentially. Several instances of the each program/service are running in the unix box, to manage the load and for risk-management reasons. As of now, we dont follow a strict strategy in... (2 Replies)