Sponsored Content
Full Discussion: Columns to rows
Top Forums UNIX for Beginners Questions & Answers Columns to rows Post 303040599 by Roopensingh on Friday 1st of November 2019 12:11:37 PM
Old 11-01-2019
Columns to rows

HI UNIX Gurus,
Not sure if this was already asked and an UNIX Guru has replied but I could not find what i wanted. I have linux environment and need help on this. I have several files like this.


Code:
a,1
b,1
utc,10/12/2019
local,10/12/2018
name,xxxy
deg,feh
10,12
20,8
30,50
32,64
46,65
49,66
50,70
51.2,71
54,73.6

and so on


i want the first 5 rows to be in columns repeated like below.



Code:
deg,feh,a,b,utc,local,name10,12,1,2,10/12/2019,10/12/2018,xxxy
20,8,1,2,10/12/2019,10/12/2018,xxxy
30,50,1,2,10/12/2019,10/12/2018,xxxy
32,64,1,2,10/12/2019,10/12/2018,xxxy
46,65,1,2,10/12/2019,10/12/2018,xxxy
49,66,1,2,10/12/2019,10/12/2018,xxxy
50,70,1,2,10/12/2019,10/12/2018,xxxy
51.2,71,1,2,10/12/2019,10/12/2018,xxxy
54,73.6,1,2,10/12/2019,10/12/2018,xxxy

Moderator's Comments:
Mod Comment
Please use code tags when posting data and code samples!

Last edited by vgersh99; 11-01-2019 at 01:25 PM.. Reason: code tags, please!
 

10 More Discussions You Might Find Interesting

1. Shell Programming and Scripting

# of rows and columns

Hi, Does anyone know the command to know the # of rows and columns for a file? thanks (3 Replies)
Discussion started by: kylle345
3 Replies

2. Shell Programming and Scripting

Columns into rows

Hi, Let me know how to achieve the below requirment Input: ======== BEGIN DSSUBRECORD Name "DOC_NO_2" SqlType "-5" Precision "0" Scale "0" Nullable "0" END DSSUBRECORD BEGIN DSSUBRECORD Name "FROM_LOC" ... (1 Reply)
Discussion started by: Ramya_1104
1 Replies

3. Shell Programming and Scripting

Rows into columns?

I have a file thats space delimited that looks something like this: Joe Smith jsmith 43234 bill1;bill2;read;read2;schedule Andy Summers asummers 11232 bill1;read Beth McConnel bmconnel 43443 read;read2;schedule;bill Susan Fowler sfowler 09332 bill1;read;schedule I need to... (8 Replies)
Discussion started by: regexnub
8 Replies

4. Shell Programming and Scripting

rows to columns

Hi Friends, I have a input file as below. how to convert rows to columns? Friday:recharge 3861140 Monday:recharge 4036228 Saturday:recharge 3996376 Sunday:recharge 3777749 Thursday:recharge 3858537 Tuesday:recharge 4047045 Wednesday:recharge 3954798 desinred output Sunday ... (3 Replies)
Discussion started by: suresh3566
3 Replies

5. Shell Programming and Scripting

Rows into Columns

Input File vCenter Servers: 172.28.173.207: vCenter Server connectivity status: Accessible ESX servers: Name: nyp-vhst1001-at.hq.nt.life.com IP address: 10.34.36.11 Virtual machines: Name:nyp-bbmds-at Ip address: 172.28.173.139 ... (1 Reply)
Discussion started by: greycells
1 Replies

6. Shell Programming and Scripting

Rows to Columns

Hi Guru's, I have a requirement where i need to convert rows to column based on a key column. Input: Account_id|Trip_Org|Trip_Dest|City|Hotel_Nm 123|DFW|CHI|Dallas|Hyatt 123|LAS|LPA|Vegas|Hyatt Palace Output:... (6 Replies)
Discussion started by: rakesh5300
6 Replies

7. Shell Programming and Scripting

Deleting all the fields(columns) from a .csv file if all rows in that columns are blanks

Hi Friends, I have come across some files where some of the columns don not have data. Key, Data1,Data2,Data3,Data4,Data5 A,5,6,,10,, A,3,4,,3,, B,1,,4,5,, B,2,,3,4,, If we see the above data on Data5 column do not have any row got filled. So remove only that column(Here Data5) and... (4 Replies)
Discussion started by: ks_reddy
4 Replies

8. Shell Programming and Scripting

Evaluate 2 columns, add sum IF two columns match on two rows

Hi all, I know this sounds suspiciously like a homework course; but, it is not. My goal is to take a file, and match my "ID" column to the "Date" column, if those conditions are true, add the total number of minutes worked and place it in this file, while not printing the original rows that I... (6 Replies)
Discussion started by: mtucker6784
6 Replies

9. Shell Programming and Scripting

Compare 2 csv files by columns, then extract certain columns of matcing rows

Hi all, I'm pretty much a newbie to UNIX. I would appreciate any help with UNIX coding on comparing two large csv files (greater than 10 GB in size), and output a file with matching columns. I want to compare file1 and file2 by 'id' and 'chain' columns, then extract exact matching rows'... (5 Replies)
Discussion started by: bkane3
5 Replies

10. Shell Programming and Scripting

Rows to columns

Hi, I have a text file with records as below Service Contract: Account Type: Client Number: Group Number: Account Currency: I want to print 2nd,3rd and 5th as a separate column, like -> Account Type: ,Client Number: ,Account Currency: How can I do that? (1 Reply)
Discussion started by: dsid
1 Replies
4S-IMPORT(1J)							      4store							     4S-IMPORT(1J)

NAME
4s-import -- Import RDF into a 4store KB SYNOPSIS
4s-import kbname [-v] [-a] [-M default-model] [-m model] [-f format] [rdf-file] ... -v, --verbose Increase verbosity (by default success is silent) -a, --add Ordinarily importing RDF to a model replaces any previous contents of that model, by using the --add flag this behaviour is overridden and any triples which were previously in the model are retained. -M, --model-default Set a model (graph) URI which is used by default for all imported files -m, --model Set a model (graph) URI for the next named file only (overrides -M if it has been used) -f, --format Tell the RDF parser the format of the files (if not specified the parser will guess) SEE ALSO
4s-query(1), 4s-size(1), 4s-httpd(1), 4s-backend(1), 4s-delete-model(1) EXAMPLES
$ 4s-import demo one.ttl two.xrdf Imports two files into the demo KB, each as a separate model (graph) guessing the RDF format used for each file $ 4s-import demo -M http://example.com/menu starters.ttl mains.ttl sweets.ttl Imports three files into a single graph named <http://example.com/menu> 4store May 31, 2019 4store
All times are GMT -4. The time now is 05:31 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy