How to make awk command faster for large amount of data?


 
Thread Tools Search this Thread
Top Forums Shell Programming and Scripting How to make awk command faster for large amount of data?
# 8  
Old 10-02-2018
Assuming your input is sorted by date (which seems likely, given logfiles), processing files individually makes another really big optimization possible - once the date exceeds the cutoff, quit! You might skip entire files.

Code:
...
gunzip < "$FILE" | awk '$3 > "[20/Jun/2018:22:00:00" { exit } ; {...}' > /tmp/$$/$FNAME &
...

That may be worth trying even without multithreading, actually.

Code:
for FILE in nginx*
do
        gunzip < "$FILE" | awk '$3 > "[20/Jun/2018:22:00:00" { exit } ; {...}'
done > output

This User Gave Thanks to Corona688 For This Post:
# 9  
Old 10-02-2018
Corona688's reasoning is absolutely correct. I created a large file and timed several awk selection methods:




Code:
time awk '{if ($9 == "(200)" && $3 > "[13/Jul/2018:17:00:00" && $3 < "[13/Jul/2018:21:00:00") print}' file > /dev/null

real    0m0,794s
user    0m0,725s
sys    0m0,068s
time awk '$9 == "(200)" {if ($3 > "[13/Jul/2018:17:00:00" && $3 < "[13/Jul/2018:21:00:00") print}' file > /dev/null

real    0m0,787s
user    0m0,733s
sys    0m0,052s
time awk '$9 == "(200)" {X = substr ($3, 14); if (X > "17:00:00" && X < "21:00:00") print}' file > /dev/null

real    0m0,806s
user    0m0,732s
sys    0m0,072s
time awk '$9 == "(200)" {X = substr ($3, 14); if (X < "17:00:00") next; if (X > "21:00:00") exit; print}' file > /dev/null

real    0m0,775s
user    0m0,676s
sys    0m0,093s
time awk '$9 == "(200)" {X = substr ($3, 14); if (X ~ /1[7-9]|2[01]:[0-5][0-9]:[0-5][0-9]$/) print}' file > /dev/null

real    0m0,827s
user    0m0,727s
sys    0m0,078s

, and found very little variation in execution time. You'd better focus on the data supply / file access.

Last edited by RudiC; 10-02-2018 at 03:22 PM..
This User Gave Thanks to RudiC For This Post:
# 10  
Old 10-04-2018
Quote:
Originally Posted by Corona688
How big are your files, how long do they take, and how fast is your disk? If you're hitting throughput limits optimizing your programs won't help one iota. If you're not, however, you can process several files at once for large gains.

File sizes varies, for instance, some files have ~300MB, others ~1GB or ~2GB.
My disk is not operating in full capacity while running awk, so I think your solution to read multiple files at once will benefit me.

Although I can't tell what's the datetime just by file name, I can leverage the fact that log files are ordered to choose what files i'll be reading, like it was suggested here. I'll make an exampĺe to illustrate.

I have 832 files in one directory totalizing 100GB, let's say nginx1.gz, nginx2.gz, ..., nginx832.gz.
first line of nginx1.gz has [11/Jul/2018:18:00:01 and the last line [11/Jul/2018:21:00:01
first line of nginx2.gz also has [11/Jul/2018:18:00:01 and the last line [11/Jul/2018:21:00:01

The natural would be nginx2.gz start with the same time or later than nginx1.gz
I could do what you've suggested with this code:

Code:
        gunzip < "$FILE" | awk '$3 > "[20/Jun/2018:22:00:00" { exit } ; {...}'

, but I would avoid to read just 2 hours of logs. So, I thought of reading the first and last line of each file and decide whether I would read the file or not. For instance,
1st line datetime would be in: zcat file1.gz | head -n 1
last line datetime would be in: zcat file1.gz | tail -n 1

So, there are two cases when I could skip reading the files:
1 - When 1st and last line is before the time I want
2 - When 1st line is after the time I want (like you've suggested)

This way I think it'll be much faster. I've finish this modification and I'm testing it right now.
If everything is allright I'll test your solution of reading multiple files at once and give you the feedback of the results.

Thank you very much for your help, I didn't know about wait and it'll probably help me.
# 11  
Old 10-04-2018
Unfortunately you can't seek to the end of a compressed file without decompressing the entire thing, so it really doesn't save you time. Otherwise that'd be a really good idea.

How big is your data, and how long does it take to process? These numbers will tell us how much improvement is possible worst-case.
These 2 Users Gave Thanks to Corona688 For This Post:
# 12  
Old 10-04-2018
Expanding on what Corona688 said, your two commands:
Code:
zcat file1.gz | head -n 1
zcat file1.gz | tail -n 1

decompress the file twice (maybe not completing the first decompression), and if you find that there is some data in that file that you need, you'll then decompress it again for your awk script to process.

I would strongly suggest creating a separate text file that contains the timestamp of the first record in each compressed file and the name of that compressed file. (And, add a new entry to the end of that file each time you create a new compress log file.) Then you can look at that (uncompressed) text file to quickly determine which compressed file(s) you need to uncompress and feed to your awk script to get the records you want for any particular timestamp range.
This User Gave Thanks to Don Cragun For This Post:
# 13  
Old 10-04-2018
I have ~120GB of compressed files per day. It's taking many hours to process


real 217m15.559s
user 763m17.030s
sys 87m40.926s


I didn't know using tail -n 1 would descompress the entire file, but it shouldn't take less time because it doesn't compare the 3rd column of all lines? For instance, if I have 100000000 lines, it would be 100000000 comparisons that would me eliminate if the last line is before the time I want to get
# 14  
Old 10-04-2018
The comparison is taking a trivial amount of time compared to the time required to:
  1. read your compressed data,
  2. uncompress your data,
  3. write your uncompressed data into a pipe, and
  4. read your uncompressed data from the pipe.
Anything you can do to eliminate those four steps for data that can't match your desired time range will yield huge benefits in run-time reduction.

Increasing the number of times you perform those four steps will increase your run times; not decrease them.
This User Gave Thanks to Don Cragun For This Post:
Login or Register to Ask a Question

Previous Thread | Next Thread

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

How to make awk command faster?

I have the below command which is referring a large file and it is taking 3 hours to run. Can something be done to make this command faster. awk -F ',' '{OFS=","}{ if ($13 == "9999") print $1,$2,$3,$4,$5,$6,$7,$8,$9,$10,$11,$12 }' ${NLAP_TEMP}/hist1.out|sort -T ${NLAP_TEMP} |uniq>... (13 Replies)
Discussion started by: Peu Mukherjee
13 Replies

2. Shell Programming and Scripting

Perl : Large amount of data put into an array

This basic code works. I have a very long list, almost 10000 lines that I am building into the array. Each line has either 2 or 3 fields as shown in the code snippit. The array elements are static (for a few reasons that out of scope of this question) the list has to be "built in". It... (5 Replies)
Discussion started by: sumguy
5 Replies

3. Shell Programming and Scripting

awk changes to make it faster

I have script like below, who is picking number from one file and and searching in another file, and printing output. Bu is is very slow to be run on huge file.can we modify it with awk #! /bin/ksh while read line1 do echo "$line1" a=`echo $line1` if then echo "$num" cat file1|nawk... (6 Replies)
Discussion started by: mirwasim
6 Replies

4. Shell Programming and Scripting

Faster way to use this awk command

awk "/May 23, 2012 /,0" /var/tmp/datafile the above command pulls out information in the datafile. the information it pulls is from the date specified to the end of the file. now, how can i make this faster if the datafile is huge? even if it wasn't huge, i feel there's a better/faster way to... (8 Replies)
Discussion started by: SkySmart
8 Replies

5. Shell Programming and Scripting

Running rename command on large files and make it faster

Hi All, I have some 80,000 files in a directory which I need to rename. Below is the command which I am currently running and it seems, it is taking fore ever to run this command. This command seems too slow. Is there any way to speed up the command. I have have GNU Parallel installed on my... (6 Replies)
Discussion started by: shoaibjameel123
6 Replies

6. Emergency UNIX and Linux Support

Help to make awk script more efficient for large files

Hello, Error awk: Internal software error in the tostring function on TS1101?05044400?.0085498227?0?.0011041461?.0034752266?.00397045?0?0?0?0?0?0?11/02/10?09/23/10???10?no??0??no?sct_det3_10_20110516_143936.txt What it is It is a unix shell script that contains an awk program as well as... (4 Replies)
Discussion started by: script_op2a
4 Replies

7. Shell Programming and Scripting

How to tar large amount of files?

Hello I have the following files VOICE_hhhh SUBSCR_llll DEL_kkkk Consider that there are 1000 VOICE files+1000 SUBSCR files+1000DEL files When i try to tar these files using tar -cvf backup.tar VOICE* SUBSCR* DEL* i get the error: ksh: /usr/bin/tar: arg list too long How can i... (9 Replies)
Discussion started by: chriss_58
9 Replies

8. AIX

amount of memory allocated to large page

We just set up a system to use large pages. I want to know if there is a command to see how much of the memory is being used for large pages. For example if we have a system with 8GB of RAm assigned and it has been set to use 4GB for large pages is there a command to show that 4GB of the *GB is... (1 Reply)
Discussion started by: daveisme
1 Replies

9. Programming

Read/Write a fairly large amount of data to a file as fast as possible

Hi, I'm trying to figure out the best solution to the following problem, and I'm not yet that much experienced like you. :-) Basically I have to read a fairly large file, composed of "messages" , in order to display all of them through an user interface (made with QT). The messages that... (3 Replies)
Discussion started by: emitrax
3 Replies

10. Shell Programming and Scripting

awk help to make my work faster

hii everyone , i have a file in which i have line numbers.. file name is file1.txt aa bb cc "12" qw xx yy zz "23" we bb qw we "123249" jh here 12,23,123249. is the line number now according to this line numbers we have to print lines from other file named... (11 Replies)
Discussion started by: kumar_amit
11 Replies
Login or Register to Ask a Question