Don't use a loop to get this done, your processing the 2.5GB details.txt file for each name in emp.txt. So if you had 2 names in emp.txt your processing 5GB of detail.txt. 10 names = 25GB. It doesn't scale well that way.
Try this:
Then you are only processing details.txt once, and of course however big emp.txt is.
Using -F might also save some time. If you don't have the '-F' option look for 'fgrep'.
But being on HP-UX the standard 'grep' should have the -F option available.
Last edited by rwuerth; 11-17-2011 at 01:59 PM..
Reason: I'm scatter brained today. Keep thinking of things to add, after the fact.
These 2 Users Gave Thanks to rwuerth For This Post:
111111111100000000001111111111
123232323200000010001114545454
232435424200000000001232131212
342354234301000000002323423443
232435424200000000001232131212
2390898994200000000001238908092
This is the record format.
From 11th position to 20th position in a record there are 0's occuring,and... (6 Replies)
Hi,
I have to find out the run time for 40-45 different componets. These components writes in to a genreric log file in a single directory.
eg.
directory is LOG and the log file name format is generic_log_<process_id>_<date YY_MM_DD_HH_MM_SS>.log
i am taking the run time using the time... (3 Replies)
I have file which contains around 5000 lines.
The lines are fixed legth but having no delimiter.Each line line contains nearly 3000 characters.
I want to delete the lines
a> if it starts with 1 and if 576th postion is a digit i,e 0-9
or
b> if it starts with 0 or 9(i,e header and footer)
... (4 Replies)
Background
-------------
The Unix flavor can be any amongst Solaris, AIX, HP-UX and Linux. I have below 2 flat files.
File-1
------
Contains 50,000 rows with 2 fields in each row, separated by pipe.
Row structure is like Object_Id|Object_Name, as following:
111|XXX
222|YYY
333|ZZZ
... (6 Replies)
Dear All,
Good Evening!!
I have a requirement to ftp a 220GB backup file to a remote backup server.
I wrote a script for this purpose.
But it takes more than 8 hours to transfer this file.
Is there any other method to do it in less time???
Thanks in Advance!!!
---------- Post updated... (5 Replies)
Hi Experts,
I had to edit (a particular value) in header line of a very huge file so for that i wanted to search & replace a particular value on a file which was of 24 GB in Size. I managed to do it but it took long time to complete. Can anyone please tell me how can we do it in a optimised... (7 Replies)
Hi,
I have created a shell script for Server Log Automation Process. I have used
find xargs grep command to search the string.
for Example,
find -name | xargs grep "816995225" > test.txt .
Here my problem is,
We have lot of records and we want to grep the string... (4 Replies)
I'm trying to remove duplicate data from an input file with unsorted data which is of size >50GB and write the unique records to a new file.
I'm trying and already tried out a variety of options posted in similar threads/forums. But no luck so far..
Any suggestions please ?
Thanks !! (9 Replies)
Hi All,
I am new to this forum and this is my first post.
My requirement is like to optimize the time taken to grep the file with 40000 lines.
There are two files FILEA(40000 lines) FILEB(40000 lines).
The requirement is like this, both the file will be in the format below... (11 Replies)
Hi All,
This query is regarding performance improvement of a command.
I have a list of IDs in a file (say file1 with single ID column) and file2 has the data rows.
I need to get the IDs from file1 and search in file2, matching rows from file2 should be written to a file3.
For this... (4 Replies)
Discussion started by: Tanu
4 Replies
LEARN ABOUT DEBIAN
pescetti
PESCETTI(1) General Commands Manual PESCETTI(1)NAME
pescetti -- Pseudo-Duplimate Generator
SYNOPSIS
pescetti
DESCRIPTION
This manual page documents briefly the pescetti command.
OPTIONS
Here are a list of the available options and what they do. You must specify exactly one from --demo, --generate or --load.
--help Prints the help text
--demo Demonstration mode. Generates one hand with permutations and the tutorial for how to use them.
--generate=N
Generate N random boards
--load=boards.txt
Load boards+analysis from boards.txt
--load-dds=boards.dds
Load boards from boards.dds in dds format
--load-analysis=tricks.txt
Load analysis from tricks.txt
--permutations=permutations.txt
Generate the permutations and save them to the given file
--curtains=curtains.txt
Save curtain cards to file curtains.txt
--save=boards.txt
Save the boards+analysis to boards.txt
--save-dds=boards.dds
Save the boards to boards.dds in dds format
--save-analysis=tricks.txt
Save the analysis to tricks.txt
--format=html|txt|pdf
Set the output mode to the given format
--title=title
Set the title for the output
--output=hands.txt
Print the hands to hands.txt, rather than to standard output
--stats Generate statistics about the set of boards; included in the hands output
--analyze Run the dds analyzer on the boards and print the resulting numberof tricks (warning SLOW)
--criteria=
A list of criteria to apply to each generated hand to generate specific hand types. The list should be space separated and each
item may be suffixed with a colon and a (fractional) probability value which can be used to weight the criteria.
E.g. --criteria="weaknt:0.8 strongnt:0.5"
Valid criteria are: unbalanced weaknt strongnt twont strongtwo weaktwo three twoclubs 4441 singlesuit twosuits partscore game
slam game-invite slam-invite jumpshift jumpfit splinter bacon weird
--probability=factor
Generate hands matching the criteria with only the given probability. Factor is in the range 0 to 1. On each attempt to generate
a board it is rejected if it doesn't match the criteria with the given probability. A factor of about 0.8 gives roughly half
matching boards
AUTHOR
This manual page was written by Matthew Johnson <debian@matthew.ath.cx>. Permission is granted to copy, distribute and/or modify this docu-
ment under the terms of the GNU General Public License, Version 2 as published by the Free Software Foundation.
On Debian systems, the complete text of the GNU General Public License can be found in /usr/share/common-licenses/GPL.
PESCETTI(1)