Due to server problem,the server accepts the connection and then do nothing,it makes the above script hang, is it possible to set time out for ftp client?
Besides, ftpPid value is process id of "tee" command but not ftp.
Would you tell me why?
I am trying to automate a process in one of two ways:
1) ftp .txt files from a unix server to a Windows 2000 server. I want to do this in a script. Can I supply the login and password to the Windows server inside the script?
2) I tried this and failed: sftp .txt files from the unix server to... (2 Replies)
I have an ftp process using vpn, my box is receiving data hit and miss from the host system. Any tool that can be used within unix (ksh) to monitor that ftp process? (2 Replies)
Hello script experts,
I am newbie to shell script. But I have to write a shell script (ASAP) where I need to ftp a file on daily basis to a remote server, and send an email with final status. I I should have a properties file with hostname, Userid, and pwd. And a shall script file should read... (1 Reply)
hi #!/bin/bash
SERVER=10.89.40.35
USER=xyz
PASSWD=xyz
ftp -in $SERVER<<EOF
user $USER $PASSWD
mkdir PPL
cd /path of remote dir
lcd /path of local dir
hash
bin
put <file name>
bye
<<EOF
The above ftp script i have to schedule in crontab at a particular instance of time run daily.... (2 Replies)
:confused:
Hi
Can anyone tell me how to check the FTP process status in linux .
eg,I'm ftpping some files from "A" server to "B" server through FTP command ,
after ftpping these files , I 'm suppose to delete the files which are successfully ftpped from "A" server ,but when I 'm checking the... (1 Reply)
my server gets text based files by ftp and one of my scripts opens and parses it.
but sometimes my script opens the file for parsing before the ftp ends. so while the file growing it parses it and moves to another location. because of this i usually get "end of file " error while opening files.
... (2 Replies)
Plzz Some One Help in this matter
I have scripts to load the data into tables and to copy files from ftp to our system.
we use to run the scripts every day....
we hav the files in the FTP server and we hav to bring the files to our system and we hav to load the data into the tables.
We... (0 Replies)
Hi,
I am using net::ftp for transferring files now i am trying in the same Linux server as a result ftp is very fast but if the server is other location (remote) then the file transferred will be time consuming.
So i want try putting FTP part as a background process. I am unaware how to do... (5 Replies)
Hi Guys,
Good day ULF! I have a general "auto-ftp" code which looks something like this:
#!/bin/sh
cd $1
ftp -v -n $2 << EOF
user $3 $4
prompt
cd $5
bin
mput $6
quit
EOF
This works pretty well, but I'm thinking of how can I make a code by maybe including on this code, for... (5 Replies)
Team,
I have multiple batchjobs running in VM, if I do ps -ef |grep java or tomcat I am getting multiple process list.
How do I get my exact tomcat process running and that is unique? via shell script? (4 Replies)
Discussion started by: Ghanshyam Ratho
4 Replies
LEARN ABOUT DEBIAN
bup-margin
bup-margin(1) General Commands Manual bup-margin(1)NAME
bup-margin - figure out your deduplication safety margin
SYNOPSIS
bup margin [options...]
DESCRIPTION
bup margin iterates through all objects in your bup repository, calculating the largest number of prefix bits shared between any two
entries. This number, n, identifies the longest subset of SHA-1 you could use and still encounter a collision between your object ids.
For example, one system that was tested had a collection of 11 million objects (70 GB), and bup margin returned 45. That means a 46-bit
hash would be sufficient to avoid all collisions among that set of objects; each object in that repository could be uniquely identified by
its first 46 bits.
The number of bits needed seems to increase by about 1 or 2 for every doubling of the number of objects. Since SHA-1 hashes have 160 bits,
that leaves 115 bits of margin. Of course, because SHA-1 hashes are essentially random, it's theoretically possible to use many more bits
with far fewer objects.
If you're paranoid about the possibility of SHA-1 collisions, you can monitor your repository by running bup margin occasionally to see if
you're getting dangerously close to 160 bits.
OPTIONS --predict
Guess the offset into each index file where a particular object will appear, and report the maximum deviation of the correct answer
from the guess. This is potentially useful for tuning an interpolation search algorithm.
--ignore-midx
don't use .midx files, use only .idx files. This is only really useful when used with --predict.
EXAMPLE
$ bup margin
Reading indexes: 100.00% (1612581/1612581), done.
40
40 matching prefix bits
1.94 bits per doubling
120 bits (61.86 doublings) remaining
4.19338e+18 times larger is possible
Everyone on earth could have 625878182 data sets
like yours, all in one repository, and we would
expect 1 object collision.
$ bup margin --predict
PackIdxList: using 1 index.
Reading indexes: 100.00% (1612581/1612581), done.
915 of 1612581 (0.057%)
SEE ALSO bup-midx(1), bup-save(1)BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown-bup-margin(1)