Did not use 'wait' yet.
How I understand by now the wait works only for child processes, started background.
Is there any other way to watch completion of any, not related process (at least, a process, owned by the same user?)
I need to start a background process, witch will be waiting... (2 Replies)
Hi Guys, I'm running a cron that mails me a file every night. The file is based on form input from the web and each evening after I mail it I delete and start a new file by the same name. It's a csv file so I preload a line of headers into the file. So far, so good. On days when no one inputs... (5 Replies)
Hi,
We currently have an Oracle database running and it is creating lots of processes in the /proc directory that are 1000M in size. The size of the /proc directory is now reading 26T. How can this be if the root file system is only 13GB?
I have seen this before we an Oracle temp file... (6 Replies)
find . -type d -print 2>/dev/null|awk '!/\.$/ {for (i=1;i<NF;i++){d=length($i);if ( d < 5 && i != 1 )d=5;printf("%"d"s","|")}print "---"$NF}' FS='/'
Can someone explain how this works..??
How can i add directory size to be listed in the above command's output..?? (1 Reply)
Hi ,
I am new to shell scripting. i have a requirement say i will receive a file in a directory say /xyz.if that file stays in that directory more than 30 min i need to get a mail to my outlook.this should run for every 20 min in crontab.
can anyone help me? (8 Replies)
I have a script that runs a console/terminal command on the server and what is want is for each of the multiple success reports fed back from the clients (echo-ed out onto the conosle) to be counted and after x number of reports reboot the server.
The Details:
The command (program) is... (0 Replies)
Hi,
I have(ksh):
...
while ]
do
fullPath=$(grep -s '/' $1 | wc -l)
echo $fullPath
if ]; then
fi
echo "Wrong, full path"
shift
done
...
I tried to do: (5 Replies)
To find the whole size of a particular directory i use "du -sk /dirname".. but after finding the direcory's size how do i make conditions like if the size of the dir is more than 1 GB i hav to delete some of the files inside the dir (0 Replies)
I have been searching both on Unix.com and Google and have not been able to find the answer to my question. I think it is partly because I can't come up with the right search terms.
Recently, my virtual server switched storage devices and I think the problem may be related to that change.... (2 Replies)
he below looks in $dir for any pattern of fileone. As is, it executes but only returns File found if the exact format in the script exsists. Why isn't a pattern of fileone being looked for and if it is in $dir, File found. I think that is what should happen. Thank you :).
dir=/path/to
if... (5 Replies)
Discussion started by: cmccabe
5 Replies
LEARN ABOUT DEBIAN
genbackupdata
GENBACKUPDATA(1) General Commands Manual GENBACKUPDATA(1)NAME
genbackupdata - generate backup test data
SYNOPSIS
genbackupdata [--chunk-size=SIZE] [--config=FILE] [-c=SIZE] [--create=SIZE] [--depth=DEPTH] [--dump-config] [--dump-setting-names]
[--file-size=SIZE] [--generate-manpage=TEMPLATE] [-h] [--help] [--list-config-files] [--log=FILE] [--log-keep=N] [--log-level=LEVEL]
[--log-max=SIZE] [--max-files=MAX-FILES] [--no-default-configs] [--output=FILE] [--quiet] [--seed=SEED] [--version]
DESCRIPTION
genbackupdata generates test data sets for performance testing of backup software. It creates a directory tree filled with files of dif-
ferent sizes. The total size and the distribution of sizes between small and big are configurable. The program can also modify an exist-
ing directory tree by creating new files, and deleting, renaming, or modifying existing files. This can be used to generate test data for
successive generations of backups.
The program is deterministic: with a given set of parameters (and a given pre-existing directory tree), it always creates the same output.
This way, it is possible to reproduce backup tests exactly, without having to distribute the potentially very large test sets.
The data set consists of plain files and directories. Files are either small text files or big binary files. Text files contain the
"lorem ipsum" stanza, binary files contain randomly generated byte streams. The percentage of file data that is small text or big binary
files can be set, as can the sizes of the respective file types.
Files and directories are named "fileXXXX" or "dirXXXX", where "XXXX" is a successive integer, separate successions for files and directo-
ries. There is an upper limit to how many files a directory may contain. After the file limit is reached, a new sub-directory is created.
The first set of files go into the root directory of the test set.
You have to give one of the options --create, --delete, --rename, or --modify for the program to do anything. You can, however, give more
than one of them, if DIR already exists. (Giving the same option more than once means that only the last instance is counted.) (DIR) is
created if it doesn't exist already.
OPTIONS --chunk-size=SIZE
generate data in chunks of this size (default: 16384)
--config=FILE
add FILE to config files
-c, --create=SIZE
how much data to create (default: 0)
--depth=DEPTH
depth of directory tree (default: 3)
--dump-config
write out the entire current configuration
--dump-setting-names
write out all names of settings and quit
--file-size=SIZE
size of one file (default: 16384)
--generate-manpage=TEMPLATE
fill in manual page TEMPLATE
-h, --help
show this help message and exit
--list-config-files
list all possible config files
--log=FILE
write log entries to FILE
--log-keep=N
keep last N logs (10)
--log-level=LEVEL
log at LEVEL, one of debug, info, warning, error, critical, fatal (default: debug)
--log-max=SIZE
rotate logs larger than SIZE, zero for never (default: 0)
--max-files=MAX-FILES
max files/dirs per dir (default: 128)
--no-default-configs
clear list of configuration files to read
--output=FILE
write output to FILE, instead of standard output
--quiet
do not report progress
--seed=SEED
seed for random number generator (default: 0)
--version
show program's version number and exit
EXAMPLES
Create data for the first generation of a backup:
genbackupdata --create=10G testdir
Modify an existing set of backup data to create a new generation:
genbackupdata -c 5% -d 2% -m 5% -r 0.5% testdir
The above command can be run for each new generation.
GENBACKUPDATA(1)