---------- Post updated at 05:54 PM ---------- Previous update was at 05:51 PM ----------
Hi,
This msg is intended to all who are all replied to help me out.
Hats off for your efforts to help me. Also i request each one of you to suggest me a link of good materials as you feel it was, for me to learn the SED & AWK atleast the basics.
Thanks.
[..]
You are welcome! Please note I updated my post and added the tildes to the search string (~$transnum~)which had fallen off before and which should make it it a bit more accurate which was also suggested by Don earlier...
Sorry, Since i am new to use blogging websites i am afraid of giving a banks transaction input structure in a public website. I am afraid since it might end up me in trouble and also apologize for a faulty input. I am learning 1 by 1 towards perfection.
I am going to try out your new suggestions will update you in 15 mins.
Thanks.
---------- Post updated at 05:40 PM ---------- Previous update was at 05:01 PM ----------
Hi Don
I am getting the below error after doing the changes what ever you have suggested.
The line 32042 is the EOT line of the particular transaction reference number. Please find the code below
This time i just directly gave the output file name rather than a variable.
Kindly let me know where i am missing something.
Thanks.
Realize that I have been up all night trying to help you (and it is now almost 6AM where I am), so I may not be thinking clearly. But, could you please explain why you chose to change the code I suggested:
to: FILENAME is an awk variable holding the name of the current input file. But, /tmp/remedixz.20160120_085021_41222370_1_new is an attempt to divide nothing by the contents of the variable tmp divided by contents of the variable remedixz followed by a syntax error. And since neither tmp nor remedixz have been defined in this awk script, both are treated as a division by zero.
Would you PLEASE just try the following script without changing it:
Note that this has a few changes to match your latest description of your transaction format, has a typo fixed, and has some minor performance improvements. It also now includes your filenames (which had not been provided before).
If /tmp/transnum contains the single line:
and there is a transaction in your big transaction file with that transaction number, it should produce a file named /tmp/remedixz.20160120_085021_41222370_1_ABC160120XYZ0983921 containing that transaction. And, as stated before, if /tmp/transnum contains multiple transaction numbers on separate lines, one invocation of this script will produce an output file for each transaction given.
If this all works, you could also add an END clause to print a list of any transaction numbers that were specified in your transaction numbers file that were not found in your big transactions file.
These 3 Users Gave Thanks to Don Cragun For This Post:
This is not working. I exported the value of transnum to variable t. The output file doesn,t have the required output.
Please find one of the existing inline perl we use. If you give me your command in the same format it will be helpful
If I understand you correctly, the command could have been:
Here's another script that will search a file with one transaction number per line and it will output a file ending in .transaction number per each find.
Save as mad_man.pl
Run as perl mad_man.pl trans_numbers data_with_trans
Or chmod +x mad_man.pl /path/to/mad_man.pl /path/to/trans_numbers /path/to/data_with_trans
It will save in /path/to/data_with_trans.<number>
Try this adaptation of RudiC's suggestion and Don's adaption for proper shell quoting on AIX:
---
Not so much 2047 bytes, in most implementations much higher or unlimited, and for some there is a much lower limit but unrelated to LINE_MAX, as I think we worked out before here: Sequence extraction
Hi sed code which Scrutinizer posted worked for a set of transaction which is actually 3455 characters
Thanks
---------- Post updated at 12:48 PM ---------- Previous update was at 12:36 PM ----------
Hi
I am going to try all of your new suggestions today and reply you back.
I just tried your suggestion and ran the script as it is you have given.
I ran three times totally, I will explain as it is what happened every time.
First Run:
Second run:
Third Run:
Note for all the runs only one output file created with extn 0000004646(I am not understanding from where this 4646 is coming from).
Can you please suggest.
Thanks.
Note for all the runs only one output file created with extn 0000004646(I am not understanding from where this 4646 is coming from).
Can you please suggest.
Thanks.
I sincerely apologize. In each case, the output file you got had a filename derived from the 2nd field (i.e., the data between the 1st and 2nd tildes which seems to be a constant for the transactions you selected to print) in a line that contained a transaction number you wanted to print, and the contents of that file was the transactions starting with the transaction after the next to the last transaction number you requested in the big input file through the last transaction number you requested from the big input file.
It comes from me not getting nearly enough sleep, you not providing sample data that matched the actual format of your data, and from me not getting nearly enough sleep. (There were three problems and I'm blaming two of them on not getting enough sleep.) Now that I have cleaned up my test data to match what I believe is your current data format, the following seems to work. Please try this replacement:
Hopefully, this will do what you want.
As stated before, if someone wants to try this on a Solaris/SunOS system, change awk to /usr/xpg4/bin/awk or nawk.
This User Gave Thanks to Don Cragun For This Post:
Hi all,
I have a file like this I want to extract only those regions which are big and continous
chr1 3280000 3440000
chr1 3440000 3920000
chr1 3600000 3920000 # region coming within the 3440000 3920000. so i don't want it to be printed in output
chr1 3920000 4800000
chr1 ... (2 Replies)
Dear all,
I have stuck with this problem for some days.
I have a very big file, this file can not open by vi command.
There are 200 loops in this file, in each loop will have one line like this:
GWA quasiparticle energy with Z factor (eV)
And I need 98 lines next after this line.
Is... (6 Replies)
The dataset I'm working on is about 450G, with about 7000 colums and 30,000,000 rows.
I want to extract about 2000 columns from the original file to form a new file.
I have the list of number of the columns I need, but don't know how to extract them.
Thanks! (14 Replies)
Hi all
I have a big file which I have attached here.
And, I have to fetch certain entries and arrange in 5 columns
Name Drug DAP ID disease approved or notIn the attached file data is arranged with tab separated columns in this way:
and other data is... (2 Replies)
Hi,
I need a unix command to delete first n (say 100) lines from a log file. I need to delete some lines from the file without using any temporary file. I found sed -i is an useful command for this but its not supported in my environment( AIX 6.1 ). File size is approx 100MB.
Thanks in... (18 Replies)
hi,
i have two files.
file1.sh
echo "unix"
echo "linux"
file2.sh
echo "unix linux forums"
now the output i need is
$./file2.sh
unix linux forums (3 Replies)
Hi,
I have a big (2.7 GB) text file. Each lines has '|' saperator to saperate each columns.
I want to delete those lines which has text like '|0|0|0|0|0'
I tried:
sed '/|0|0|0|0|0/d' test.txt
Unfortunately, it scans the file but does nothing.
file content sample:... (4 Replies)
I have a command which prints #lines after and before the search string in the huge file
nawk 'c-->0;$0~s{if(b)for(c=b+1;c>1;c--)print r;print;c=a}b{r=$0}' b=0 a=10 s="STRING1" FILE
The file is 5 gig big.
It works great and prints 10 lines after the lines which contains search string in... (8 Replies)
1 . Thanks everyone who read the post first.
2 . I have a log file which size is 143M , I can not use vi open it .I can not use xedit open it too.
How to view it ?
If I want to view 200-300 ,how can I implement it
3 . Thanks (3 Replies)