Nawk script to compare records of a file based on a particular column.
Hi Gurus,
I am struggling with nawk command where i am processing a file based on columns.
Here is the sample data file.
In this file, the first column is the primary key.But there are some duplicate entries in first column.
in this case :
Now my requirement is..
I want to search the file according to the first column and if there are any duplicate entries in first column then i will search the 6th column and print that row in which the 6th column is not empty.
My desired output is as below:
I have till now tried this script but somehow its not giving me the desired output.
Can anyone help me understand what i am doing wrong in my awk script.
Hi All,
I need to select only those records having a non zero record in the first column of a comma delimited file.
Suppose my input file is having data like:
"0","01/08/2005 07:11:15",1,1,"Created",,"01/08/2005"
"0","01/08/2005 07:12:40",1,1,"Created",,"01/08/2005"... (2 Replies)
Hi all,
I would like to extract records of a file based on a condition. The file contains 47 fields, and I would like to extract only those records that match a certain value in one of the columns, e.g.
COL1 COL2 COL3 ............... COL47
1 XX 45 ... (4 Replies)
Here is my situation. I need to compare two tab separated files (diff is not useful since there could be known difference between files).
I have found similar posts , but not fully matching.I was thinking of writing a shell script using cut and grep and while loop but after going thru posts it... (2 Replies)
Hi,
I want to remove duplicate records including the first line based on column1. For example
inputfile(filer.txt):
-------------
1,3000,5000
1,4000,6000
2,4000,600
2,5000,700
3,60000,4000
4,7000,7777
5,999,8888
expected output:
----------------
3,60000,4000
4,7000,7777... (5 Replies)
Hi guys,
I need help in making a command to find some data.
I have multiple files in which multiple records are present.. Each record is separated with a carriage return and in each record there are multiple fields with each field separated by "|"
what i want is that I want to extract... (1 Reply)
I was reading this thread. It looks like a simpler way to say this is to only keep uniq lines based on field or column 1.
https://www.unix.com/shell-programming-scripting/165717-removing-duplicate-records-file-based-single-column.html
Can someone explain this command please? How are there no... (5 Replies)
Hi Experts,
I want to compare 2 text files based on their column values text1 is like
prd-1234 yes no yes yes
prd-2345 no no no yes
prd-6475 yes yes yes no
and test 2 is
prd-1234 no no no yes
prd-2345 yes no no no
desired out put as follows
prd-1234 1 3
prd-235 1 4
basically it shows... (5 Replies)
Hi, I have two files roughly 1200 fields in length for each row, sorted on the 2nd field. I need to compare based on that 2nd column between file1 and file2 and print lines that exist in both files into separate files (I can't guarantee that every line in file1 is in file2).
Example:
File1: ... (1 Reply)
Hi,
I am having below two CSV's
col_1,col_2,col_3
1,2,4
1,3,6
col_1,col_3,col2,col_5,col_6
1,2,3,4,5
1,6,3,,,
I need to compare based on the columns where the mismatch is
expected output
col_1,col_2,col_3
1,2,4 (3 Replies)
Hi,
I tried filtering the records in a csv file using "awk" command listed below.
awk -F"~" '$4 ~ /Active/{print }' inputfile > outputfile
The output always has all the entries.
The same command worked for different users from one of the forum links.
content of file I was... (3 Replies)