05-07-2011
script to constantly read the last 500 new logs in a log file
Hello,
I would like to write a bash script that would monitor a log file for a certain number of logs, let's say 500 logs and when it reaches that number to write the last log to another file. For example, I want to watch the /var/adm/messages and everytime, there is 500 new logs that are generated in this file to write the last log into a new file lik /var/tmp/tmp1.log. So, if the logs keep popping up in the /var/adm/message, for every 500 of them there will one instance in /var/tmp/tmp1.log. But, it has to monitor for new logs. Let's say the script already read 500 logs, the count has to be for new logs occurence all the time. It should not consider previous logs that could have been in the /var/adm/messages before that it might have already read. Please give me some ideas.
Thank you,
10 More Discussions You Might Find Interesting
1. UNIX for Dummies Questions & Answers
I have a log file which is continuously added to, called log.file. I'd like to
monitor this file, and when certain lines are found, update some totals in
another file. I've played around with tail -f, grep, and awk, but can't seem
to hit the right note, so to speak.
The lines I'm... (0 Replies)
Discussion started by: nortonloaf
0 Replies
2. Shell Programming and Scripting
Hi,
I have gps receiver log..its giving readings .like below
Trying 127.0.0.1...
Connected to localhost.
Escape character is '^]'.
GPSD,R=1
$GPGSV,3,1,11,08,16,328,40,11,36,127,00,28,33,283,39,20,11,165,00*71... (3 Replies)
Discussion started by: gudivada213
3 Replies
3. Shell Programming and Scripting
What is wrong with my menu script? Do I need to continue with the read statements? All I want to do with option 4 is to cat some /var/log/files and awk out a few lines? How do I do that please?
$ cat menu.sh ... (11 Replies)
Discussion started by: taekwondo
11 Replies
4. Shell Programming and Scripting
Hi,
I am working on the script to parsing the specific message like "aaaa" in multiple log files like N1-***,N2-***,N3-***...
The script is to find the list of lof files which contains the message "aaaa" and export the list into excel filE.
Can anyone give help?
Thanks (2 Replies)
Discussion started by: shyork2001
2 Replies
5. Shell Programming and Scripting
Hi
I have a big log file
:08,936 DEBUG HttpConnectionManager.getConnection: config =
11:39:08,936 DEBUG Getting free connection,
11:39:08,989 DEBUG Freeing connection, hostConfig=HostConfiguration
11:39:08,989 DEBUG Notifying no-one, there are no waiting threads
11:39:09,046... (4 Replies)
Discussion started by: javaholics
4 Replies
6. Shell Programming and Scripting
Hello Team,
I need help to improve my script which is used to grep 500 error messages in the logs.
I am using following logic in the script to grep 500 error messages in the logs.
var1=`awk '$9 == "500"' access_log | tail -1`
The above logic is not useful if logs are not getting... (1 Reply)
Discussion started by: coolguyamy
1 Replies
7. UNIX for Advanced & Expert Users
Hi,
Im looking for a shell script which will search for a particular string in a log file as below scenario
1. I need to run URL http://localhost/client/update?feedid=200 in shell script at(eg)4:00 PM which will not take more than 15 mins to complete.
2. After 15 mins i need to... (6 Replies)
Discussion started by: Paulwintech
6 Replies
8. Shell Programming and Scripting
# cat /tmp/checkdate.log
SQL*Plus: Release 11.2.0.1.0 Production on Mon Sep 17 22:49:00 2012
Copyright (c) 1982, 2009, Oracle. All rights reserved.
Connected to:
Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
FIRST_TIME NEXT_TIME... (1 Reply)
Discussion started by: SarwalR
1 Replies
9. Shell Programming and Scripting
Hi All,
I want to read the log file for last 30mins logs with time stamps.
Am using below command but, it is not working for me
awk -F - -vDT="$(date --date="30 minutes ago" "+%b %_d %H:%M:%S")" ' DT < $1' log.file >tmp.txt
log file time format is 2016-09-27 14:00:25,192
Use code... (1 Reply)
Discussion started by: Prashanth.K
1 Replies
10. Shell Programming and Scripting
United States 1.2.3.4 80 10 1563790914 1 1932454179 1.2.3.6 55517 11.1.2.1 55517
Italy 1.2.3.4 80 10 1563790914 1 1932454179 1.2.3.6 55517 11.1.2.1 55517
India 1.2.3.4 80 10 1563790914 1 1932454179 1.2.3.6 55517 11.1.2.1 55517
south Africa 1.2.3.4 80 10 1563790914 1... (9 Replies)
Discussion started by: arm
9 Replies
LEARN ABOUT DEBIAN
save_binary_logs
SAVE_BINARY_LOGS(1p) User Contributed Perl Documentation SAVE_BINARY_LOGS(1p)
NAME
save_binary_logs - Concatenating binary or relay logs from the specified file/position to the end of the log. This command is automatically
executed from MHA Manager on failover, and manual execution should not be needed normally.
SYNOPSIS
# Test
$ save_binary_logs --command=test --binlog_dir=/var/lib/mysql --start_file=mysqld-bin.000002
# Saving binary logs
$ save_binary_logs --command=save --binlog_dir=/var/lib/mysql --start_file=mysqld-bin.000002 --start_pos=312
--output_file=/var/tmp/aggregate.binlog
# Saving relay logs
$ save_binary_logs --command=save --start_file=mysqld-relay-bin.000002 --start_pos=312 --relay_log_info=/var/lib/mysql/relay-log.info
--output_file=/var/tmp/aggregate.binlog
save_binary_logs concatenates binary or relay logs from the specified log file/position to the end of the log. This tool is intended to be
invoked from the master failover script(MHA Manager), and manual execution is normally not needed.
DESCRIPTION
Suppose that master is crashed and the latest slave server has received binary logs up to mysqld-bin.000002:312. It is likely that master
has more binary logs. If it is not sent to the slave, slaves will lose all binlogs from mysqld-bin.000002:312. The purpose of the
save_binary_logs is to save binary logs that are not replicated to slaves. If master is reachable through SSH and binary logs are readable,
saving binary logs is possible.
Here is an example:
$ save_binary_logs --command=save --start_file=mysqld-bin.000002 --start_pos=312 --output_file=/var/tmp/aggregate.binlog
Then all binary logs starting from mysqld-bin.000002:312 are concatenated and stored into /var/tmp/aggregate.binlog. If you have binary
logs up to mysqld-bin.000004, the following mysqlbinlog outputs are written.
mysqld-bin.000002:Format Description Event(FDE), plus from 312 to the tail mysqld-bin.000003:from 0 to the tail, excluding FDE
mysqld-bin.000004:from 0 to the tail, excluding FDE
perl v5.14.2 2012-01-08 SAVE_BINARY_LOGS(1p)