05-23-2006
Quote:
Originally Posted by Corona688
What speed are you getting now?
If the link is slow, no amount of tuning will make it faster. If the link is fast, tuning won't make much difference either. So, yeah. Make the link faster.
currently,
766 Kbytes/s (approx)
10 More Discussions You Might Find Interesting
1. UNIX for Dummies Questions & Answers
Can this be done? Code samples welcome and encouraged. (2 Replies)
Discussion started by: idesaj
2 Replies
2. Shell Programming and Scripting
Hi everyone.
I wrote a ftp script for sending files.
while read FNAME
do
# Begin ftp
ftp -i -n $HOST << END
user $USER $PASSWD
put $FNAME $FNAME
quit
END
# End ftp
done < ftp_sending_list.dat
HOST, USER and PASSWD are my account data.
This script quits ftp for many times and this is... (1 Reply)
Discussion started by: Euler04
1 Replies
3. HP-UX
Hi gurus,
I have a problem with ftp access. The first 2 test e.g. Test A & Test B was successful with the file size 64kb (800++ numbers). The third test with file size 120kb was failed. The error is "Netout :Connection reset by peer". No password entered manually since the test run from the... (3 Replies)
Discussion started by: yeazas
3 Replies
4. AIX
Hi,
Does anyone know if it is possible to tar files larger than 2GB? The reason being is they want me to dump a single file (which is around 20GB) to a tape drive and they will restore it on a Solaris box. I know the tar have a limitation of 2GB so I am thinking of a way how to overcome this.... (11 Replies)
Discussion started by: depam
11 Replies
5. UNIX for Dummies Questions & Answers
Hello Unix Gurus,
I am new to Unix so need some help on this.
I am using the following commands:
1) mv -f Inputpath/*. outputpath
2) cp Inputpath/*. outputpath
3) rm -rf somepath/*
4) Find Inputpath/*.
Now I get the following error with... (18 Replies)
Discussion started by: pchegoor
18 Replies
6. UNIX for Dummies Questions & Answers
My unzip command doesn't work for files that are greater than 4GB. Consider my file name is unzip -p -a filename.zip, the command doesn't work since the size of the file is larger. I need to know the corresponding 7z command for the same. This is my Unix shell script program:
if
then
... (14 Replies)
Discussion started by: chandraprakash
14 Replies
7. Shell Programming and Scripting
I need to backup my database but the files are very large and the TAR command will not let me. I searched aids and found that I could do something with the mknod, COMPRESS and TAR command using them together. I appreciate your help. (10 Replies)
Discussion started by: frizcala
10 Replies
8. UNIX for Dummies Questions & Answers
Hi,
I have one large files of 100000 rows with header column.
Eg:
Emp Code, Emp Name
101,xxx
102,YYY
103,zzz
...
...
I want to split the files into smaller files with only 30000 rows each..File 1,2 and 3 must have 30000 rows and file 4 must contain 10000 rows.
But the column... (1 Reply)
Discussion started by: Nivas
1 Replies
9. UNIX for Beginners Questions & Answers
I need help modifying these two scripts to do the following:
- print files in (MB) instead of (KB)
- only select files larger than 500MB -> these will be mailed out daily
- Select all files regardless of size all in (MB) -> these will be mailed out once a week
this is what i have so far and... (5 Replies)
Discussion started by: donpasscal
5 Replies
10. UNIX for Beginners Questions & Answers
I am new at developing EXPECT scripts. I'm trying to create a script that will automatically connect to a several UNIX (sun solaris and HPUX) database server via FTP and pull the sizes of the listener/alert log files from specified server directory on the remote machines.
1. I want the script... (7 Replies)
Discussion started by: mikebantor
7 Replies
LEARN ABOUT OSX
ftp_geturl
ftp::geturl(n) ftp client ftp::geturl(n)
__________________________________________________________________________________________________________________________________________________
NAME
ftp::geturl - Uri handler for ftp urls
SYNOPSIS
package require Tcl 8.2
package require ftp::geturl ?0.2.1?
::ftp::geturl url
_________________________________________________________________
DESCRIPTION
This package provides a command which wraps around the client side of the ftp protocol provided by package ftp to allow the retrieval of
urls using the ftp schema.
API
::ftp::geturl url
This command can be used by the generic command ::uri::geturl (See package uri) to retrieve the contents of ftp urls. Internally it
uses the commands of the package ftp to fulfill the request.
The contents of a ftp url are defined as follows:
file The contents of the specified file itself.
directory
A listing of the contents of the directory in key value notation where the file name is the key and its attributes the asso-
ciated value.
link The attributes of the link, including the path it refers to.
BUGS, IDEAS, FEEDBACK
This document, and the package it describes, will undoubtedly contain bugs and other problems. Please report such in the category ftp of
the Tcllib SF Trackers [http://sourceforge.net/tracker/?group_id=12883]. Please also report any ideas for enhancements you may have for
either package and/or documentation.
SEE ALSO
ftpd, mime, pop3, smtp
KEYWORDS
ftp, internet, net, rfc 959
CATEGORY
Networking
ftp 0.2.1 ftp::geturl(n)