| Help | unix | grep | sort | uniq - Different output from what I thought would be the same
Hello,
I'm having an consistency issue....
1) This will highlight every 'a' in each word.
2) This will highlight 'a' if it occurs at least once in a sequence. So every 'a'.
Output of 1) I would think would be identical to 2)
I output the both results into seperate text files:
8647903 Oct 12 21:34 holding2a.txt this is where 1) went
8642625 Oct 12 21:34 holding2b.txt this is where 2) went
Why is there difference in file size?
more holding2a.txt | wc -l results 276975
more holding2a.txt | wc -l results 276975
this holds true for holding2b.txt
So I wanted to compare the text in each file.
more holding2a.txt | sort revealed this output
more holding2b.txt | sort revealed this output
I checked the word and line count on both files again and they were the same.
I finally I merged the two files:
The result is 798 for both -w and -l.
Sample output:
Top part.
Bottom few.
1) I don't know why uniq -u isn't removing the what appears to be duplicates.
2) I don't know why sort isn't sorting properly, I mean 'aa' is in two different places at the top of the list, I would think they would be together.
I tried sort -d and sort -s which resulted in what appeared to be the order. sort -d did take noticably longer to finish.
Hello all,
Need to pick your brains,
I have a 10Gb file where each row is a name, I am expecting about 50 names in total. So there are a lot of repetitions in clusters.
So I want to do a
sort -u file
Will it be considerably faster or slower to use a uniq before piping it to sort... (3 Replies)
Hi !
I am trying to remove doubbled entrys in a textfile only between delimiters.
Like that example but i dont know how to do that with sort or similar.
input:
{
aaa
aaa
}
{
aaa
aaa
}
output:
{
aaa
}
{ (8 Replies)
Hi again,
I have files with the following contents
datetime,ip1,port1,ip2,port2,number
How would I find out how many times ip1 field shows up a particular file? Then how would I find out how many time ip1 and port 2 shows up?
Please mind the file may contain 100k lines. (8 Replies)
I have a flatfile A.txt
2012/12/04 14:06:07 |trees|Boards 2, 3|denver|mekong|mekong12
2012/12/04 17:07:22 |trees|Boards 2, 3|denver|mekong|mekong12
2012/12/04 17:13:27 |trees|Boards 2, 3|denver|mekong|mekong12
2012/12/04 14:07:39 |rain|Boards 1|tampa|merced|merced11
How do i sort and get... (3 Replies)
Hi All,
I have a text file with the format shown below. Some of the records are duplicated with the only exception being date (Field 15). I want to compare all duplicate records using subscriber number (field 7) and keep only those records with greater date.
... (1 Reply)
Input File is :
-------------
25060008,0040,03,
25136437,0030,03,
25069457,0040,02,
80303438,0014,03,1st
80321837,0009,03,1st
80321977,0009,03,1st
80341345,0007,03,1st
84176527,0047,03,1st
84176527,0047,03,
20000735,0018,03,1st
25060008,0040,03,
I am using the following in the script... (5 Replies)
Does anyone have a quick and dirty way of performing a sort and uniq in perl?
How an array with data like:
this is bkupArr BOLADVICE_VN
this is bkupArr MLT6800PROD2A
this is bkupArr MLT6800PROD2A
this is bkupArr BOLADVICE_VN_7YR
this is bkupArr MLT6800PROD2A
I want to sort it... (4 Replies)
Using the last, uniq, sort and cut commands, determine how many times the different users have logged in.
I know how to use the last command and cut command...
i came up with last | cut -f1 -d" " | uniq
i dont know if this is right, can someone please help me... thanks (1 Reply)
I have a file:
Fred
Fred
Fred
Jim
Fred
Jim
Jim
If sort is executed on the listed file, shouldn't the output be?:
Fred
Fred
Fred
Fred
Jim
Jim
Jim (3 Replies)