Sponsored Content
Top Forums Shell Programming and Scripting Help in modifying a PERL script to sort Singletons and Duplicates Post 302948886 by Aia on Friday 3rd of July 2015 01:22:31 PM
Old 07-03-2015
Use as starting point.

Code:
#!/usr/bin/perl

# segregate.pl

use strict;
use warnings;

# data structure as a hash
my %db;
# populate a data structure based on first field
while(my $line = <>) {
    # remove new line
    chomp $line;
    # get field one
    my ($id) = split /=/, $line;
    # add field1 => line to the structure
    push @{$db{$id}}, $line;
}

# sorted keys for singletons
my @single;
# sorted keys for duplicates
my @dup;
# iterate through db hash using sorted keys
for my $key (sort keys %db) {
    if(scalar @{$db{$key}} == 1) {
        # add to singles list if one item
        push @single, $key;
    }
    else {
        # more than one item, add to dup list
        push @dup, $key;
    }
}

# Display lists
print "SINGLETONS:\n";
# iterate only through sorted singles
for my $single (@single) {
    print @{$db{$single}}, "\n";
}

print "DUPES:\n";
# iterate only through sorted duplicates
for my $dup (@dup) {
    print join "\n", @{$db{$dup}};
    print "\n";
}

Usage:
perl segregate.pl datafile
This User Gave Thanks to Aia For This Post:
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

Removing duplicates [sort , uniq]

Hey Guys, I have file which looks like this, Contig201#numbPA Contig1452#nmdynD6PA dm022p15.r#CG6461PA dm005e16.f#SpatPA IGU001_0015_A06.f#CG17593PA I need to remove duplicates based on the chracter matching upto '#'. for example if we consider this.. Contig201#numbPA... (4 Replies)
Discussion started by: sharatz83
4 Replies

2. Shell Programming and Scripting

Sort, Uniq, Duplicates

Input File is : ------------- 25060008,0040,03, 25136437,0030,03, 25069457,0040,02, 80303438,0014,03,1st 80321837,0009,03,1st 80321977,0009,03,1st 80341345,0007,03,1st 84176527,0047,03,1st 84176527,0047,03, 20000735,0018,03,1st 25060008,0040,03, I am using the following in the script... (5 Replies)
Discussion started by: Amruta Pitkar
5 Replies

3. UNIX for Dummies Questions & Answers

removing duplicates and sort -k

Hello experts, I am trying to remove all lines in a csv file where the 2nd columns is a duplicate. I am try to use sort with the key parameter sort -u -k 2,2 File.csv > Output.csv File.csv File Name|Document Name|Document Title|Organization Word Doc 1.doc|Word Document|Sample... (3 Replies)
Discussion started by: orahi001
3 Replies

4. Shell Programming and Scripting

modifying perl script

Hi freinds I have a small problem I want u to help me in, I have a syslog server and configured it to send me email automatically, I get a small perl script to help me in, and tested it to send alerts to root and it worked successfully without any problems Now I want to send it outside, I... (4 Replies)
Discussion started by: reaky
4 Replies

5. Shell Programming and Scripting

need Shell script for Sort BASED ON FIRST FIELD and PRINT THE WHOLE FILE WITHOUT DUPLICATES

Can some one provide me a shell script. I have file with many columns and many rows. need to sort the first column and then remove the duplicates records if exists.. finally print the full data with first coulm as unique. Sort BASED ON FIRST FIELD and remove the duplicates if exists... (2 Replies)
Discussion started by: tuffEnuff
2 Replies

6. Shell Programming and Scripting

remove duplicates and sort

Hi, I'm using the below command to sort and remove duplicates in a file. But, i need to make this applied to the same file instead of directing it to another. Thanks (6 Replies)
Discussion started by: dvah
6 Replies

7. Shell Programming and Scripting

Help in modifying existing Perl Script to produce report of dupes

Hello, I have a large amount of data with the following structure: Word=Transliterated word I have written a Perl Script (reproduced below) which goes through the full file and identifies all dupes on the right hand side. It creates successfully a new file with two headers: Singletons and Dupes.... (5 Replies)
Discussion started by: gimley
5 Replies

8. Shell Programming and Scripting

Bash - remove duplicates without sort

I need to use bash to remove duplicates without using sort first. I can not use: cat file | sort | uniq But when I use only cat file | uniq some duplicates are not removed. (4 Replies)
Discussion started by: locoroco
4 Replies

9. Shell Programming and Scripting

Sort and Remove duplicates

Here is my task : I need to sort two input files and remove duplicates in the output files : Sort by 13 characters from 97 Ascending Sort by 1 characters from 96 Ascending If duplicates are found retain the first value in the file the input files are variable length, convert... (4 Replies)
Discussion started by: ysvsr1
4 Replies

10. Shell Programming and Scripting

Concatenate and sort to remove duplicates

Following is the input. 1st and 3rd block are same(block starts here with '*' and ends before blank line) , 2nd and 4th blocks are also the same: cat <file> * Wed Feb 24 2016 Tariq Saeed <tariq.x.saeed@mail.com> 2.0.7-1.0.7 - add vmcore dump support for ocfs2 * Mon Jun 8 2015 Brian Maly... (4 Replies)
Discussion started by: Paras Pandey
4 Replies
sortbib(1)						      General Commands Manual							sortbib(1)

Name
       sortbib - sort bibliographic database

Syntax
       sortbib [-sKEYS] database...

Description
       The  command sorts files of records containing refer key-letters by user-specified keys.  Records may be separated by blank lines, or by .[
       and .] delimiters, but the two styles may not be mixed together.  This program reads through each database and pulls out key fields,  which
       are  sorted  separately.   The sorted key fields contain the file pointer, byte offset, and length of corresponding records.  These records
       are delivered using disk seeks and reads, so may not be used in a pipeline to read standard input.

       By default, alphabetizes by the first %A and the %D fields, which contain the senior author and date.  The -s option is used to specify new
       KEYS.  For instance, -sATD will sort by author, title, and date, while -sA+D will sort by all authors, and date.  Sort keys past the fourth
       are not meaningful.  No more than 16 databases may be sorted together at one time.  Records longer than 4096 characters will be truncated.

       The command sorts on the last word on the %A line, which is assumed to be the author's last name.  A word in the final  position,  such	as
       ``jr.''	or  ``ed.'',  will be ignored if the name beforehand ends with a comma.  Authors with two-word last names or unusual constructions
       can be sorted correctly by using the convention ``'' in place of a blank.  A %Q field is considered to be the same as %A, except  sorting
       begins  with  the first, not the last, word.  The command sorts on the last word of the %D line, usually the year.  It also ignores leading
       articles (like ``A'' or ``The'') when sorting by titles in the %T or %J fields; it will ignore articles of any  modern  European  language.
       If a sort-significant field is absent from a record, places that record before other records containing that field.

Options
       -sKEYS
	  Specifies new sort KEYS.  For example, ATD sorts by author, title, and date.

See Also
       addbib(1), indxbib(1), lookbib(1), refer(1), roffbib(1)

																	sortbib(1)
All times are GMT -4. The time now is 03:15 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy