Any way to improve performance of this script


 
Thread Tools Search this Thread
Top Forums Shell Programming and Scripting Any way to improve performance of this script
# 1  
Old 03-16-2009
Any way to improve performance of this script

I have a data file of 2 gig

I need to do all these, but its taking hours, any where i can improve performance, thanks a lot


#!/usr/bin/ksh
echo TIMESTAMP="$(date +'_%y-%m-%d.%H-%M-%S')"
function showHelp {
cat << EOF >&2
syntax extreme.sh FILENAME
Specify filename to parse
EOF
echo "$*" >&2
exit 1;
}
if [ $# -gt 1 -o $# -eq 0 ]; then showHelp "Please Specify Only One File\n";fi
#if [ $# -eq 0 ]; then showHelp "Please Specify Only One File\n";fi
sed -e "/$UNKNOWN SEGMENT/d" $1 > $1.tmp
mv $1.tmp $1
for i in `cut -c216-231 $1 |uniq`
do
awk /$i/ $1 > $i.file
split -a 4 -l 20000 $i.file $i.file
rm $i.file
FILES=01
for y in `ls $i*`
do
NUM_OF_FILES="$(ls $i* |wc -l|tr -d " ")"
if [ $NUM_OF_FILES -lt 10 ];then
NUM_OF_FILES=0$NUM_OF_FILES
sed -e "s/./$NUM_OF_FILES/230" -e "s/.//232" -e "s/../$FILES/110" -e "s/../$FILES/225" -e "s/./$NUM_OF_FILES/468" -e "s/.//470" $y |nl -nrz -ba -w6 > $y.tmp
else
NUM_OF_FILES=$NUM_OF_FILES
sed -e "s/./$NUM_OF_FILES/230" -e "s/.//232" -e "s/../$FILES/110" -e "s/../$FILES/225" -e "s/./$NUM_OF_FILES/468" -e "s/.//470" $y |nl -nrz -ba -w6 > $y.tmp
fi
FILES=`expr $FILES + 1`
typeset -Z2 FILES
mv $y.tmp $y
done
cat $i* |nl -nrz -ba -w7 > $i.file
rm $i.*aa*
done
cat *.file > $1.new
cut -f2,1- $1.new |sed 's/ //1' > $1.new.tmp
mv $1.new.tmp $1.new
awk '{one=substr($0,1,13); two=substr($0,525,13);gsub(two,one); print}' $1.new |cut -f2- > $1.new.tmp
mv $1.new.tmp $1.new
rm -f *.file
echo "Please Check the Modified \"$1.new\" in the current directory"
echo TIMESTAMP="$(date +'_%y-%m-%d.%H-%M-%S')"
# 2  
Old 03-18-2009
run it on a faster box?

What exactly is the script doing?
# 3  
Old 03-19-2009
too much awk, sed, cats lying around. what exactly are you wanting to do.
# 4  
Old 03-19-2009
I have a big data file , these are things i am doing

I splitting the file with same pattern & then split into 20000 lines each

& adding count number of files(splitted) using sed & adding line numbers

again merging all same pattern files & again add line numbers.

Then swap first 2 columns, add substitute in another position in the file

Finally merging all files back.
Login or Register to Ask a Question

Previous Thread | Next Thread

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

Bash script search, improve performance with large files

Hello, For several of our scripts we are using awk to search patterns in files with data from other files. This works almost perfectly except that it takes ages to run on larger files. I am wondering if there is a way to speed up this process or have something else that is quicker with the... (15 Replies)
Discussion started by: SDohmen
15 Replies

2. Programming

Improve the performance of my C++ code

Hello, Attached is my very simple C++ code to remove any substrings (DNA sequence) of each other, i.e. any redundant sequence is removed to get unique sequences. Similar to sort | uniq command except there is reverse-complementary for DNA sequence. The program runs well with small dataset, but... (11 Replies)
Discussion started by: yifangt
11 Replies

3. UNIX for Dummies Questions & Answers

How to improve the performance of this script?

Hi , i wrote a script to convert dates to the formate i want .it works fine but the conversion is tkaing lot of time . Can some one help me tweek this script #!/bin/bash file=$1 ofile=$2 cp $file $ofile mydates=$(grep -Po '+/+/+' $ofile) # gets 8/1/13 mydates=$(echo "$mydates" | sort |... (5 Replies)
Discussion started by: vikatakavi
5 Replies

4. Programming

Help with improve the performance of grep

Input file: #content_1 12314345345 242467 #content_14 436677645 576577657 #content_100 3425546 56 #content_12 243254546 1232454 . . Reference file: content_100 (1 Reply)
Discussion started by: cpp_beginner
1 Replies

5. Shell Programming and Scripting

How to improve the performance of parsers in Perl?

Hi, I have around one lakh records. I have used XML for the creation of the data. I have used these 2 Perl modules. use XML::DOM; use XML::LibXML; The data will loo like this and most it is textual entries. <eid>19000</eid> <einfo>This is the ..........</einfo> ......... (3 Replies)
Discussion started by: vanitham
3 Replies

6. Shell Programming and Scripting

Want to improve the performance of script

Hi All, I have written a script as follows which is taking lot of time in executing/searching only 3500 records taken as input from one file in log file of 12 GB Approximately. Working of script is read the csv file as an input having 2 arguments which are transaction_id,mobile_number and search... (6 Replies)
Discussion started by: poweroflinux
6 Replies

7. Shell Programming and Scripting

Improve the performance of a shell script

Hi Friends, I wrote the below shell script to generate a report on alert messages recieved on a day. But i for processing around 4500 lines (alerts) the script is taking aorund 30 minutes to process. Please help me to make it faster and improve the performace of the script. i would be very... (10 Replies)
Discussion started by: apsprabhu
10 Replies

8. UNIX for Dummies Questions & Answers

Improve Performance

hi someone tell me which ways i can improve disk I/O and system process performance.kindly refer some commands so i can do it on my test machine.thanks, Mazhar (2 Replies)
Discussion started by: mazhar99
2 Replies

9. Shell Programming and Scripting

How to improve grep performance...

Hi All, I am using grep command to find string "abc" in one file . content of file is *********** abc = xyz def= lmn ************ i have given the below mentioned command to redirect the output to tmp file grep abc file | sort -u | awk '{print #3}' > out_file Then i am searching... (2 Replies)
Discussion started by: pooga17
2 Replies

10. UNIX for Advanced & Expert Users

improve performance by using ls better than find

Hi , i'm searching for files over many Aix servers with rsh command using this request : find /dir1 -name '*.' -exec ls {} \; and then count them with "wc" but i would improve this search because it's too long and replace directly find with ls command but "ls *. " doesn't work. and... (3 Replies)
Discussion started by: Nicol
3 Replies
Login or Register to Ask a Question