Can any one give me command How to delete duplicate records with out sort.
Suppose if the records like below:
345,bcd,789
123,abc,456
234,abc,456
712,bcd,789
out tput should be
345,bcd,789
123,abc,456
Key for the records is 2nd and 3rd fields.fields are seperated by colon(,). (2 Replies)
Can any one give me command How to delete duplicate records with out sort.
Suppose if the records like below:
345,bcd,789
123,abc,456
234,abc,456
712,bcd,789
out tput should be
345,bcd,789
123,abc,456
Key for the records is 2nd and 3rd fields.fields are seperated by colon(,). (19 Replies)
I have many pdf's scattered across 4 machines. There is 1 location where I have other Pdf's maintained. But the issues it the 4 machines may have duplicate pdf's among themselves, but I want just 1 copy of each so that they can be transfered to that 1 location.
What I have thought is:
1) I have... (11 Replies)
I have a file a.txt having content like
deepak
ram
sham
deepram
sita
kumar
I Want to delete the first line containing "deep" ...
I tried using...
grep -i 'deep' a.txt
It gives me 2 rows...I want to delete the first one..
+ need to know the command to delete the line from... (5 Replies)
I have a csv file that I would like to remove duplicate lines based on field 1 and sort. I don't care about any of the other fields but I still wanna keep there data intact. I was thinking I could do something like this but I have no idea how to print the full line with this. Please show any method... (8 Replies)
if different branch code is available for same BIC code and one of the branch code is XXX.only one row will be stored and with branch code as XXX .rest of the rows for the BIC code will not be stored.
for example if $7 is BIC code and $8 is branch code
INPUT file are following... (9 Replies)
I want to delete partical duplicate file
>gma-miR156d Gm01,PACID=26323927 150.00 -18.28 2 18 17 35 16 75.00% 81.25%
>>gma-miR156d Gm01,PACID=26323927 150.00 -18.28 150.00 -18.28 1 21 119 17
I want to order by the second column and delete the... (1 Reply)
Hi, all
I have a csv file that I would like to remove duplicate lines based on 1st field and sort them by the 1st field. If there are more than 1 line which is same on the 1st field, I want to keep the first line of them and remove the rest. I think I have to use uniq or something, but I still... (8 Replies)
Discussion started by: refrain
8 Replies
LEARN ABOUT DEBIAN
tv_sort
TV_SORT(1p) User Contributed Perl Documentation TV_SORT(1p)NAME
tv_sort - Sort XMLTV listings files by date, and add stop times.
SYNOPSIS
tv_sort [--help] [--by-channel] [--output FILE] [FILE...]
DESCRIPTION
Read XMLTV data and write out the same data sorted in date order. Where stop times of programmes are missing, guess them from the start
time of the next programme on the same channel. For the last programme of a channel, no stop time can be added.
Tv_sort also performs some sanity checks such as making sure no two programmes on the same channel overlap.
--output FILE write to FILE rather than standard output
--by-channel sort first by channel id, then by date within each
channel.
--duplicate-error If the input contains the same programme more than once,
consider this as an error. Default is to silently
ignore duplicate entries.
The time sorting is by start time, then by stop time. Without --by-channel, if start times and stop times are equal then two programmes
are sorted by internal channel id. With --by-channel, channel id is compared first and then times.
You can think of tv_sort as converting XMLTV data into a canonical form, useful for diffing two files.
EXAMPLES
At a typical Unix shell or Windows command prompt:
tv_sort <in.xml >out.xml
tv_sort in.xml --output out.xml
These are different ways of saying the same thing.
AUTHOR
Ed Avis, ed@membled.com
perl v5.14.2 2006-03-02 TV_SORT(1p)