Hi,
How to output the duplicate record to another file. We say the record is duplicate based on a column whose position is from 2 and its length is 11 characters.
The file is a fixed width file.
ex of Record:
DTYU12333567opert tjhi kkklTRG9012
The data in bold is the key on which... (1 Reply)
Hi All,
I needs to fetch unique records based on a keycolumn(ie., first column1) and also I needs to get the records which are having max value on column2 in sorted manner... and duplicates have to store in another output file.
Input :
Input.txt
1234,0,x
1234,1,y
5678,10,z
9999,10,k... (7 Replies)
Hello,
I am new to shell scripting. I have a huge file with multiple columns for example:
I have 5 columns below.
HWUSI-EAS000_29:1:105 + chr5 76654650 AATTGGAA HHHHG
HWUSI-EAS000_29:1:106 + chr5 76654650 AATTGGAA B@HYL
HWUSI-EAS000_29:1:108 + ... (4 Replies)
Hi team,
I have 20 columns csv files. i want to find the duplicates in that file based on the column1 column10 column4 column6 coulnn8 coulunm2 . if those columns have same values . then it should be a duplicate record.
can one help me on finding the duplicates,
Thanks in advance.
... (2 Replies)
Hi,
I have a file with fields like below:
A;XYZ;102345;222
B;XYZ;123243;333
C;ABC;234234;444
D;MNO;103345;222
E;DEF;124243;333
desired output:
C;ABC;234234;444
D;MNO;103345;222
E;DEF;124243;333
ie, if the 4rth field is a duplicate.. i need only those records where... (5 Replies)
Hi All,
I have a text file with three columns. I would like a simple script that removes lines in which column 1 has duplicate entries, but use the largest value in column 3 to decide which one to keep. For example:
Input file:
12345a rerere.rerere len=23
11111c fsdfdf.dfsdfdsf len=33 ... (3 Replies)
Hi guys,Got a bit of a bind I'm in. I'm looking to remove duplicates from a pipe delimited file, but do so based on 2 columns. Sounds easy enough, but here's the kicker...
Column #1 is a simple ID, which is used to identify the duplicate.
Once dups are identified, I need to only keep the one... (2 Replies)
I have /tmp dir with filename as:
010020001_S-FOR-Sort-SYEXC_20160229_2212101.marker
010020001_S-FOR-Sort-SYEXC_20160229_2212102.marker
010020001-S-XOR-Sort-SYEXC_20160229_2212104.marker
010020001-S-XOR-Sort-SYEXC_20160229_2212105.marker
010020001_S-ZOR-Sort-SYEXC_20160229_2212106.marker... (4 Replies)
Discussion started by: gnnsprapa
4 Replies
LEARN ABOUT DEBIAN
sfood-checker
SFOOD-CHECKER(1) General Commands Manual SFOOD-CHECKER(1)NAME
sfood-checker - check for superfluous import statements in Python source code
SYNOPSIS
sfood-checker [options] files...
DESCRIPTION
This script is used to detect forgotten imports that are not used anymore. When writing Python code (which happens so fast), it is often
the case that we forget to remove useless imports.
This is implemented using a search in the AST, and as such we do not require to import the module in order to run the checks. This is a
major advantage over all the other lint/checker programs, and the main reason for taking the time to write it.
As inputs, it can receive either files or directories; in case no argument is passed, it parses the current directory recursively.
OPTIONS -h, --help
show the help message and exit
--debug
Debugging output.
-I IGNORES, --ignore=IGNORES
Add the given directory name to the list to be ignored.
-d, --disable-pragmas
Disable processing of pragma directives as strings after imports.
-D, --duplicates, --enable-duplicates
Enable experimental heuristic for finding duplicate imports.
-M, --missing, --enable-missing
Enable experimental heuristic for finding missing imports.
SEE ALSO sfood(1), sfood-cluster(1), sfood-copy(1), sfood-flatten(1), sfood-graph(1), sfood-imports(1).
AUTHOR
sfood-checker was written by Martin Blais <blais@furius.ca> and it's part of snakefood suite.
This manual page was written by Sandro Tosi <morph@debian.org>, for the Debian project (and may be used by others).
January 2, 2009 SFOOD-CHECKER(1)