I want rearrange about 5 million rows (with 300 columns) into groups.
Data looks like the following: where there were various experiments (column 2) conducted at different locations (column headers in top row column 4 onwards) in different years (column 1) using instruments (column 3). The numbers in the matrix ( row 2 onwards, column 4 onwards) indicate how many instances of experiments were successful.
What I want is to rearrange the rows
Input
What I would like to have are the rows rearranged into groups such that
1) Within the same year
2) Using the same instrument
create groups such that,
each group has at least 3 locations in common, each of which has at least 20 successful experiments.
Requested Output
The group number assignment is in the first column, it doesn't matter to me what number is assigned to which group as long as the members are assigned to the correct group. Also if experiments cant be assigned to groups, they should be left without a group number. Assignment should be with the group with maximum common locations, if equal then it can be assigned to multiple groups.
Please assist, I am learning some very basic awk, this seems beyond me..
I tried this but I dont think this will get anywhere near to what I require
I have a string which has IP and data of files for that ip which is printed using the below code: Sample string code below:
str1="10.9.11.128\n-rwxr-xr-x user1 2019-12-29 17:53 /var/branch/custom/tg.xml 286030210\n10.9.12.129\n-rwxr-xr-x user1 2019-12-29 17:53 /app/branch/custom/tg.xml... (6 Replies)
As part of some report generation, I've written a script to fetch the values from DB. But, unluckily, for certain Time ranges(1-9.99,10-19.99 etc), I don't have data in DB.
In such cases, I would like to write zero (0) instead of empty. The desired output will be exported to csv file.
... (1 Reply)
Hi, some help is highly appreciated, I want to de-group my data for statistical analysis. I made up some sample data, there shouldnt be repeated lines in the output. My data is in excel but I can make it tab-delimited text.
A B,C
A B,D,E
X Y
X Y,ZExpected output
A B
A C
A D
A E
X Y... (3 Replies)
Hi all,
I have 2 files containing data like this:
so if there is same entry repeated in the column like1,2,3,4
I have to check if there is different entries column like 2,4
but similar entries for duplicatein column 2 like1,3
the output shuld be like this for first file
... (5 Replies)
Hi folks,
I have the following data.Any help is greatly appreciated.
order File_name
7222245 7222245.pdf
7222245 7222245a.pdf
7222245 7222245b.pdf
7222245 7222245c.pdf
7222245 7222245d.pdf
7222250 ... (1 Reply)
Hi gurus,
I have data coming in as shown below. And in each case, I need to pick the data in the last group as shown below.
Data Set 1:
DC | 18161621
LA | 15730880
NY | 16143237
DC | 18161621
LA | 17316397
NY | 17915905
DC | 18161621
LA | 17993534
NY | 18161621
DC | 18161621... (11 Replies)
I am familiar with scripting, but I am trying to see if there is an easy way to append files from similar files into one file. For example, if there is file1_20121201, file1_20121202, file1_20121203,
file2_20121201, file2_20121202, file2_20121203
I want to be able to combine all the data from... (3 Replies)
Hello,
I have a file and i want to sort by third column and extract the three top lines of each group, it is determined by the second column (144, 89, 55, etc).
Could you please help me with the appropiate awk shell script
XLY-XLP 144 0.592772 XLY-XLE 144 0.798121 ... (3 Replies)
Hi,
I am writing just to share my appreciation for help I have received from this site in the past.
In a previous post Split File by Data Group I received a lot of help with a troublesome awk script to reformat some complicated data blocks. What I learned really came in hand recently when I... (1 Reply)
Hi all,
I'm having a little trouble solving a file split I need to get done.
I have the following data:
1. Light
1A. Light Soft
texture: it's soft
color: the color value is that of something light
vital statistics: srm: 23 og: 1.035 sp: 1.065
comment: this is nice if you like... (8 Replies)