03-07-2008
Parsing out records from one huge record
Hi,
I have one huge record and know that each record in the file is 550 bytes long. How do I parse out individual records from the single huge record.
Thanks,
10 More Discussions You Might Find Interesting
1. Shell Programming and Scripting
I have a very big file of 5gb size and there are about 50 million records in there. I have to delete the records based on recrord number that I know fromoutside with out opening the file. The record numbers are very random like 5000678, 7890005 etc.
Can somebody let me know how i can... (5 Replies)
Discussion started by: dsravan
5 Replies
2. UNIX for Advanced & Expert Users
Hi,
I got a file which is one huge record. I know each record should be 550 bytes long. How do I parse out the records from the one huge record. (1 Reply)
Discussion started by: bwrynz1
1 Replies
3. Shell Programming and Scripting
111111111100000000001111111111
123232323200000010001114545454
232435424200000000001232131212
342354234301000000002323423443
232435424200000000001232131212
2390898994200000000001238908092
This is the record format.
From 11th position to 20th position in a record there are 0's occuring,and... (6 Replies)
Discussion started by: mjkreddy
6 Replies
4. Shell Programming and Scripting
Hi, I am trying to parse a very long record in a text file into multiple records by checking ADD, DELETE, or MODIFY field value in a shell script.
Input
# File name xyz.txt
ADD|N000|8015662|DELETE|N001|9915662|MODIFY|N999|85678
Output
ADD|N000|8015662|
DELETE|N001|9915662|... (8 Replies)
Discussion started by: naveed
8 Replies
5. Shell Programming and Scripting
Hi All,
I have a *.csv files in a die /pro/lif/dow, (pipe delimiter file), these files are having 8 columns and 6 column(CDR_LOGIC) records are populated as below, I need to incorporate the below logic in all the *.csv files.
11||:ColumnA||:ColumnB
123||:ColumnA
IIF(:ColumnA = :ColumnC then... (6 Replies)
Discussion started by: shruthidwh
6 Replies
6. Shell Programming and Scripting
I was given a data file that I need to split into multiple lines/records based on a key word. The problem is that it is 2.5GB or bigger and everything I try in perl or sed causes a Segmentation fault. Can someone give me some other ideas.
The data is of the form:... (5 Replies)
Discussion started by: leolson
5 Replies
7. Shell Programming and Scripting
I have not get much answer/solution for the posting. Here I break down the question and hope to get some help.
1. How can I use AWK to read in two records at the same time and keep loop to next two when the condition is meet?
position 1-10 --> Unique to identity whether... (4 Replies)
Discussion started by: menglm
4 Replies
8. Shell Programming and Scripting
I need to make one record to multiple records based on occurence column in the record and change the date.For example below first record has 5 ,so need to create 5 records from one and change the date to 5 months.Occurence can be any number.
I am unable to come with a script.Can some one help
... (5 Replies)
Discussion started by: traininfa
5 Replies
9. Shell Programming and Scripting
Hi i want to fetch 100k record from a file which is looking like as below.
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
... (17 Replies)
Discussion started by: lathigara
17 Replies
10. UNIX for Beginners Questions & Answers
Hi,
I have a backup report that unfortunately has some kind of hanging indent thing where the first line contains one column more than the others
I managed to get the output that I wanted using awk, but just wanting to know if there is short way of doing it using the same awk
Below is what... (2 Replies)
Discussion started by: newbie_01
2 Replies
LEARN ABOUT MOJAVE
cap_mkdb
CAP_MKDB(1) BSD General Commands Manual CAP_MKDB(1)
NAME
cap_mkdb -- create capability database
SYNOPSIS
cap_mkdb [-v] [-f outfile] file1 [file2 ...]
DESCRIPTION
Cap_mkdb builds a hashed database out of the getcap(3) logical database constructed by the concatenation of the specified files .
The database is named by the basename of the first file argument and the string ``.db''. The getcap(3) routines can access the database in
this form much more quickly than they can the original text file(s).
The ``tc'' capabilities of the records are expanded before the record is stored into the database.
The options as as follows:
-f outfile
Specify a different database basename.
-v Print out the number of capability records in the database.
FORMAT
Each record is stored in the database using two different types of keys.
The first type is a key which consists of the first capability of the record (not including the trailing colon (``:'')) with a data field
consisting of a special byte followed by the rest of the record. The special byte is either a 0 or 1, where a 0 means that the record is
okay, and a 1 means that there was a ``tc'' capability in the record that couldn't be expanded.
The second type is a key which consists of one of the names from the first capability of the record with a data field consisting a special
byte followed by the the first capability of the record. The special byte is a 2.
In normal operation names are looked up in the database, resulting in a key/data pair of the second type. The data field of this key/data
pair is used to look up a key/data pair of the first type which has the real data associated with the name.
RETURN VALUE
The cap_mkdb utility exits 0 on success and >0 if an error occurs.
SEE ALSO
dbopen(3), getcap(3), termcap(5)
BSD
May 31, 2019 BSD