Recently I've been doing some performance changes for inter-lpar configs and completing some testing. So far test have been successful and promising but only completed in a dev environment.
Here is what I've done
created a separated vlan for inter-lpar communications separate from the VIO.
adjusted the MTU to 65280, and enabled Large Send.
transferred have increased by aprox 70% in some test. lowest was a 50% increase from the original configurations on other servers. I did see a improvement in terms of CPU as well compared to using default settings.
so my main question is this....
overall it was great. the problem which i was expecting was sftp/scp.
using nmon i did confirm they were transferring on the correct ports but transfer times were extremely slow.
I was using a 6GB file for test, on ftp best avg time was 14 seconds (all test were completed by a average of 10 test) when transferring 6 6GB files at the same time, avg time was 23 seconds.
Using sftp/scp transfer times were 2:30 min. 6 files approx 5 min.
so my next step was to try changing the algorithm on these processes which i personally was trying to avoid. I didn't run 6 concerent test with this one but the avg time was around 1:30 min.
Here was what i used for sftp:
here is what i used for scp:
from ifconfig on this interface:
when using MTU=65280 and largesend what would be the best thing to adjust in terms of network tuning.
As I said this is just dev testing and I know real cases will be different but I'm just trying to see what improvements I can make since some people have been complaining about slow transfers between systems and this was the best i can think of at the moment.
Hi,
I have 2 machines in production environment:
1. redhat machine for application
2. DB machine (oracle)
The application doing a lot of small read&writes from and to the DB machine.
The problem is that after some few hours the network from the application to the DB becomes very slow and... (4 Replies)
I have identical M5000 machines that are needing to transfer very large amounts of data between them. These are fully loaded machines, and I've already checked IO, memory usage, etc... I get poor network performance even when the machines are idle or copying via loopback. The 10 GB NICs are... (7 Replies)
hello i have a ubuntu ssh server that i can acess from any of my comnputers but only if they are on the same wireless network as the server. i tested trhis my tehtehring my samsung blackjack to my windows partition and installing openssh to windows it works when windows is on the wireless but no... (1 Reply)
I have a Teradata Machine, using MP-RAS Unix, with a 1000 Intel Ethernet card and a Cisco switch.
If I configure the ethernet card and the switch to auto, so they negotiate to 1000, or configure the ethernet card and switch manually to 1000Full or 100Full, the velocity is very very low.
Only... (2 Replies)
About 4 years ago I wrote this tool inspired by Rob Urban's collect tool for DEC's Tru64 Unix. What makes this tool as different as collect was in its day is its ability to run at a low overhead and collect tons of stuff. I've expanded the general concept and even include data not available in... (0 Replies)
I am running FC4:
Linux maincomp 2.6.13-1.1532_FC4smp
I recently changed the OS from windows XP, and have a feeling that for some reason my wireless network card is slower on Fedora Core 4. The Belkin PCI 802.11b card was automatically detected and configured by FC4 when I installed the OS,... (0 Replies)
I am posting this as an FYI. I burned a lot of time on digging for this solution. The problem was performance between newer AIX p series boxes and a sun 4500 with a fiber nic. I tried many misc. solutions but this turned out to fix the issue. My bandwidth benchmark went from around 20MB/sec to... (1 Reply)