supports the find command traversing directory file trees - i.e., searching and locating files.
Assuming you want to keep the command you already have (and I am not sue that Rudi's suggested test is valid because of file and directory caching ):
A limiting factor is known to be the number of sub-directories in the file tree, and possibly the number of available open file descriptors - a per process limit.
If you can parallelize your code using several processes it may improve performance. I'm not sure this will help much because it depends on the number of sub-directories being large to gain any benefit. The developers who write system code try to maximize throughput.
What I'm saying is: performance enhancement work is subjective and often a misplaced resource and a waste of programmer time.
Suppose your command runs in one minute in production. Then you work hard and get it down to 35 seconds. The user perception of "slow" will still be there, so you have to get it down to maybe 6 seconds to make users happy and see it as "faster". In this case getting an order of magnitude improvement may not be possible.
And in this case you would have to do something about directory caching messing up testing because (you check this yourself) once you open a directory the system caches it for speedier access. Use the time command and rerun the command to see what I mean:
This User Gave Thanks to jim mcnamara For This Post:
I have a ksh shell script and i need to pass arguments which are generated by data pulled from a database.
When the argument to the shell script is too long (about 4000 charecters) the below is the issue observed.
I copy the command which is 4000 charecters long from the logs and paste it... (7 Replies)
Hello All,
We have a working script which identifies and kills ipcs resources which havent been correctly killed during normal shutdowns.
It is working fine and dandy however there are some issues now.
Environment:
SunOS 5.10 Generic_148888-03 sun4u sparc SUNW,SPARC-Enterprise
... (4 Replies)
Hi,
I have created a shell script for Server Log Automation Process. I have used
find xargs grep command to search the string.
for Example,
find -name | xargs grep "816995225" > test.txt .
Here my problem is,
We have lot of records and we want to grep the string... (4 Replies)
Hi,
I wanted to run a particlar script for every 20 minutes. I dont have crontab in my server. Hence i ran this script in a loop by providing the command sleep 1200
Now i wanted to know is there any performance issue if this job keeps on execute in the server.
Thanks,
Puni (1 Reply)
I'm using sed to do find and replace. But since the file is huge and i have more than 1000 files to be searched, the script is taking a lot of time. Can somebody help me with a better sed command. Below is the details.
Input:
1
1
2
3
3
4
5
5
Here I know the file is sorted.
... (4 Replies)
I have written a code using AWK & sed to compare two files.
The structure of the files is like this"
Format is this:
<bit code> <file code> <string>
Follwoed by any numbers of properties lines whic start with a "space"
10101010101111101 XX abcd a
AS sasa
BS kkk
1110000101010110 XX... (1 Reply)
Hi ,
i'm searching for files over many Aix servers with rsh command using this request :
find /dir1 -name '*.' -exec ls {} \;
and then count them with "wc"
but i would improve this search because it's too long and replace directly find with ls command but "ls *. " doesn't work.
and... (3 Replies)