Sponsored Content
Top Forums Shell Programming and Scripting Verify large file with Unix function Post 302340499 by duke0001 on Monday 3rd of August 2009 03:50:02 PM
Old 08-03-2009
Verify large file with Unix function

I am seeking help on one UNIX function writting. Please help.

I have a large file, named 'MyFile', It was tab-delmited. I am told that each line in 'MyFile' has 7 columns, and that the values in the 5th column are integers. How do I use shell functions (and standard LINUX/UNIX filters) to verify that these conditions were satisfied in 'MyFile'?
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Unix File System performance with large directories

Hi, how does the Unix File System perform with large directories (containing ~30.000 files)? What kind of structure is used for the organization of a directory's content, linear lists, (binary) trees? I hope the description 'Unix File System' is exact enough, I don't know more about the file... (3 Replies)
Discussion started by: dive
3 Replies

2. HP-UX

Need to split a large data file using a Unix script

Greetings all: I am still new to Unix environment and I need help with the following requirement. I have a large sequential file sorted on a field (say store#) that is being split into several smaller files, one for each store. That means if there are 500 stores, there will be 500 files. This... (1 Reply)
Discussion started by: SAIK
1 Replies

3. Shell Programming and Scripting

how to edit large file in unix

hi All, Plz let me know how to edit a file with 2000000 records. each record contains with 40 field seperated by |. i want modify 455487 record, but i am uable to edit this large file using vi editor in unix. plz let me know how to modify this file. Thanks in advance. -Bali Reddy (3 Replies)
Discussion started by: balireddy_77
3 Replies

4. Shell Programming and Scripting

Performance issue in UNIX while generating .dat file from large text file

Hello Gurus, We are facing some performance issue in UNIX. If someone had faced such kind of issue in past please provide your suggestions on this . Problem Definition: /Few of load processes of our Finance Application are facing issue in UNIX when they uses a shell script having below... (19 Replies)
Discussion started by: KRAMA
19 Replies

5. UNIX for Dummies Questions & Answers

Copying large file problem on SVR4 Unix

We have 3 Unix servers all running SVR4 Unix 1.4. I have no problems copying files to and from 2 of the servers using either the rcp command or ftp but when i come to transfer large files to the third server the copy gives up part way through and crashes this server. Copying smaller files using RCP... (7 Replies)
Discussion started by: coatesd
7 Replies

6. UNIX for Dummies Questions & Answers

Verify the data type in a file with UNIX function

I am seeking help on this UNIX function, please help. Thanks in advance. I have a large file, named as 'MyFile'. It was tab-delmited, I am told that each record in column 1 is unique. How would I verify this with UNIX function or command? (1 Reply)
Discussion started by: duke0001
1 Replies

7. UNIX for Dummies Questions & Answers

Verify a flat file with UNIX function or script.

I re-post my question to seek your help and critique. I was required to verify a very large and tab-delimited file, named 'MyFile'. 1. The each line in 'MyFile' has 7 columns, and that the values in the 5th column are integers. I need to use shell functions (and standard LINUX/UNIX filters) to... (1 Reply)
Discussion started by: duke0001
1 Replies

8. AIX

How to send large file through E-mail IN Unix AIX

hI ALL, I am newbie to Unix.There are 2 GB files in UNIX that should be sent to one of my friend who is working on Windows Environment.Is there any way to send these file to him? By Compressing like that? Can anyone help on this? With thanks, Aravindh (4 Replies)
Discussion started by: Aravindhbabu
4 Replies

9. Shell Programming and Scripting

Can someone verify the code of two function

Input=D123,S1234,D345 | kESTREL PRODUCTION SUPPORT echo -en "Enter the logmsg=" read logmsg logmsg1=${logmsg%%|*}; echo "$logmsg1"|tr ',' '\n' | sed 's/*$ | sed '/^$/d'//'>pre-commit.config Char() { while read line do if ] || ];then echo "Success" else exit 1;... (1 Reply)
Discussion started by: rohit22hamirpur
1 Replies

10. UNIX for Beginners Questions & Answers

Command to extract empty field in a large UNIX file?

Hi All, I have records in unix file like below. In this file, we have empty fields from 4th Column to 22nd Column. I have some 200000 records in a file. I want to extract records only which have empty fields from 4th field to 22nd filed. This file is comma separated file. what is the unix... (2 Replies)
Discussion started by: rakeshp
2 Replies
STAG-FLATTEN(1p)					User Contributed Perl Documentation					  STAG-FLATTEN(1p)

NAME
stag-flatten - turns stag data into a flat table SYNOPSIS
stag-flatten -c name -c person/name dept MyFile.xml DESCRIPTION
reads in a file in a stag format, and 'flattens' it to a tab-delimited table format. given this data: (company (dept (name "special-operations") (person (name "james-bond")) (person (name "fred")))) the above command will return a two column table special-operations james-bond special-operations fred If there are multiple values for the columns within the node, then the cartesian product will be calculated USAGE
stag-flatten [-p PARSER] [-c COLS] [-c COLS] NODE <file> ARGUMENTS
-p|parser FORMAT FORMAT is one of xml, sxpr or itext xml assumed as default -c|column COL1,COL2,COL3,.. the name of the columns/elements to write out this can be specified either with multiple -c arguments, or with a comma-seperated (no spaces) list of column (terminal node) names after a single -c -n|nest if set, then the output will be a compress repeating values into the same row; each cell in the table will be enclosed by {}, and will contain a comma-delimited set of values SEE ALSO
Data::Stag perl v5.10.0 2008-12-23 STAG-FLATTEN(1p)
All times are GMT -4. The time now is 03:15 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy