this is a simple script . can it help find your way ?
certainly above code with its structure can be write better and more efficient . i wrote it very quickly with my limited skill .
Hi
Can you suggest some perl script. My OS is HP-UX 11.11 I want to it into a cron job.
Every night it will backup the file with that day's date and open a dummy file.
Thanks
Ash (3 Replies)
Hi,
I'm using Linux 2.6 cross compiled for a embedded powerpc.
Earlier we were using busybox syslogd for logging the messages, but as the flexibilty is less, we decided to move to standard syslog.
Now I want to limit the size of the log file (/var/log/messages) to 128K.
I couldn't find... (4 Replies)
Hi,
I am using fetchmail in my application so as to download mails to the localhost where the application is hosted from the mailserver.Fetchmail is configured as as to run as a daemon polling mails during an interval of 1sec.
So my concern here is, during each 2sec it is writing two... (10 Replies)
Hi Can anyone assist me, I am trying to compate the size of a logfile to a maximum size 1000 and delete if exceeds the limit.
The problem I am getting is the command not found for the line
if ( $LOGNAME_SIZE >= $MAXSIZE); then
Appreciate your response.
Script:
LOGDIR="/home/... (6 Replies)
I am very new tothe shellscripting. I wrote a script as follows.
NOW=$(date +"%b-%d-%y")
LOGFILE="log-$NOW.log"
sqlplus -s username/password @select.sql <<EOF >>$LOGFILE
exit
EOF
the out put of the script file is around 20 columns.But it is not displaying in single line. Some part ofthe... (2 Replies)
:wall:Dear All.:p
How to check log size every 10min. by script (not crontab)
if log size not change with alert "Log not update"
My Path :: /usr/home/logical/mono/log/tplink/
My Log :: mono11_tplink.log , mono12_tplink.log , etc
I want oup put.
EX. if log not update.
.
.
.
Fri Jan ... (1 Reply)
Anyone knows how to use AWK to achieve the following
Sun Feb 12 00:41:01-00:41:59 Success:2 Fail:2
Sun Feb 12 00:42:01-00:42:59 Success:1 Fail:2
Sun Feb 12 01:20:01-01:20:59 Success:1 Fail:2
Mon Feb 13 22:41:01-22:41:59 Success:1 Fail:1
log file:
Success
Success
Fail
Fail
... (9 Replies)
Hi, I've been trying to develop a script that performs the parsing of a log every 1 minute and then generating some statistics. I'm fairly new to programming and this is why I come to ask if I can lend a hand.
this is my log:
xxxx 16/04/2012 17:00:52 - xxxx714 - E234 - Time= 119 ms.... (8 Replies)
Discussion started by: jockx
8 Replies
LEARN ABOUT DEBIAN
storebackupupdatebackup
STOREBACKUPUPDATEBACKUP(1) User Contributed Perl Documentation STOREBACKUPUPDATEBACKUP(1)NAME
storeBackupUpdateBackup.pl - updates / finalizes backups created by storeBackup.pl with option --lateLink, --lateCompress
SYNOPSIS
storeBackupUpdateBackup.pl -b backupDirectory [--autorepair]
[--print] [--verbose] [--debug] [--lockFile] [--noCompress]
[--progressReport number] [--checkOnly]
[--logFile
[--suppressTime] [-m maxFilelen]
[[-n noOfOldFiles] | [--saveLogs]]
[--compressWith compressprog]]
storeBackupUpdateBackup.pl --interactive --backupDir topLevlDir
[--autorepair] [--print]
WARNING
!!! USAGE IN PARALLEL WITH storeBackup.pl CAN DESTROY YOUR BACKUPS !!!
OPTIONS --interactive, -i
interactive mode for reparing / deleting currupted
backups created with option '--lateLinks'
--backupDir, -b
top level directory of all backups (must exist)
--autorepair, -a
repair simple inconsistencies automatically without
requesting the action
--print
print configuration read from configuration file and stop
--verbose, -v
verbose messages
--debug, -d
generate detailed information about the files
with the linking information in it
--lockFile, -L
lock file, if exist, new instances will finish if
an old is already running
If set to the same file as in storeBackup it will
prevent $prog from running in parallel
to storeBackup, default is $lockFile
--noCompress
maximal number of parallel compress operations,
default = chosen automatically
--checkOnly -c
do not perform any action, only check consistency
--progressReport
print progress report:
after each 'number' files when compressing
after each 'number * 1000' files when linking
after each 'number * 10000' files when performing chmod
--logFile, -l
logFile, Default: stdout
--suppressTime
suppress output of time in logfile
--maxFilelen, -m
maximal length of log file, default = 1e6
--noOfOldFiles, -n
number of old log files, default = 5
--saveLogs
save log files with date and time instead of deleting the
old (with [-noOldFiles])
--compressWith
compress saved log files (e.g. with 'gzip -9').
default is 'bzip2'
COPYRIGHT
Copyright (c) 2008,2012 by Heinz-Josef Claes. Published under the GNU General Public License v3 or any later version
perl v5.14.2 2012-06-16 STOREBACKUPUPDATEBACKUP(1)