Hi,
Let me explain the problem clearly:
Let the entries in my file be:
lion,tiger,bear
apple,mango,orange,apple,grape
unix,windows,solaris,windows,linux
red,blue,green,yellow
orange,maroon,pink,violet,orange,pink
Can we detect the lines in which one of the words(separated by field... (8 Replies)
Hi,
How to identify duplicate columns in a row?
Input data: may have 30 columns
9211480750 LK 120070417 920091030
9211480893 AZ 120070607
9205323621 O7 120090914 120090914 1420090914 2020090914 2020090914
9211479568 AZ 120070327 320090730
9211479571 MM 120070326
9211480892 MM 120070324... (3 Replies)
Hi all,
I have a tab-delimited file and want to remove identical lines, i.e. all of line 1,2,4 because the columns are the same as the columns in other lines. Any input is appreciated.
abc gi4597 9997 cgcgtgcg $%^&*()()*
abc gi4597 9997 cgcgtgcg $%^&*()()*
ttt ... (1 Reply)
Dear All,
I need to find the difference between two adjacent columns. The file is having 'i' columns and i need to find the difference between two adjacent columns (like $1 difference $2; $2 difference $3; .... and $(i-1) difference $i). I have used the following coding
awk '{ for (i=1; i<NF;... (7 Replies)
Hello experts,
I have a requirement where I have to implement two checks on a csv file:
1. Check to see if the value in first column is duplicate, if any value is duplicate script should exit.
2. Check to verify if the value at second column is between "yes" or "no", if it is anything else... (4 Replies)
Hi,
I have a file with 1M records
ABC 200 400 2.4 5.6
ABC 410 299 12 1.5
XYZ 4 5 6 7
MNO 22 40 30 70
MNO 47 55 80 150
What I want is for all the rows it should take the max value where there are duplicates
output
ABC 410 400 12 5.6
XYZ 4 5 6 7
MNO 47 55 80 150
How can i... (6 Replies)
I have this structure:
col1 col2 col3 col4 col5
27 xxx 38 aaa ttt
2 xxx 38 aaa yyy
1 xxx 38 aaa yyy
I need to collapse duplicate lines ignoring column 1 and add values of duplicate lines (col1) so it will look like this:
col1 col2 col3 col4 col5
27 xxx 38 aaa ttt ... (3 Replies)
I have a 13gb file. It has the following columns:
The 3rd column is basically correlation values. I want to delete those rows which are repeated between the columns:
A B 0.04
B C 0.56
B B 1
A A 1
C D 1
C C 1
Desired Output: (preferably in a .csv format
A,B,0.04
B,C,0.56
C,D,1... (3 Replies)
Input
1,ABCD,no
2,system,yes
3,ABCD,yes
4,XYZ,no
5,XYZ,yes
6,pc,noCode used to find duplicate with regard to 2nd column
awk 'NR == 1 {p=$2; next} p == $2 { print "Line" NR "$2 is duplicated"} {p=$2}' FS="," ./input.csv
Now is there a wise way to de-duplicate the entire line (remove... (4 Replies)
Discussion started by: deadyetagain
4 Replies
LEARN ABOUT DEBIAN
sary
SARY(1) Sary SARY(1)NAME
sary - manual page for sary 1.0.2
SYNOPSIS
sary [OPTION]... PATTERN FILE
DESCRIPTION -c, --count
only print a count of occurrences
-i, --ignore-case
ignore case distinctions
-l, --lexicographical
sort in lexicographical order
-A, --after-context=NUM
print NUM lines of trailing context
-B, --before-context=NUM
print NUM lines of leading context
-C, --context=[NUM],
print NUM (default 2) lines of output context
-s, --start=TAG,
print tagged region. set start tag to TAG
-e, --end=TAG,
print tagged region. set end tag to TAG
-h, --help
display this help and exit
COPYRIGHT
Copyright (C) 2000 Satoru Takabayashi All rights reserved.
This is free software; you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by
the Free Software Foundation; either version 2.1, or (at your option) any later version.
This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MER-
CHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more details.
sary 1.0.2 December 2000 SARY(1)