Those commands are trying to load the entire line into memory at once, which won't work on a 64 32-bit system since there's a 4GB limit on per-process memory and sufficient clutter in the way that you probably can't get an entire 2.5G in one chunk.
How big are the individual records? You can tell awk to use things other than \n as a record separator, changing its definition of lines:
Use nawk or gawk if you have it.
Last edited by Corona688; 12-02-2011 at 04:39 PM..
Hi,
I am new to UNIX scripting and woiuld appreicate your help...
Input file contains only one (but long) record:
aaaaabbbbbcccccddddd.....
Desired file:
NEW RECORD #new record (hardcoded) added as first record - its length is irrelevant#
aaaaa
bbbbb
ccccc
ddddd
...
...
... (1 Reply)
All,
We receive a file with a large no of records (records can vary) and we have to split it into two files based on another file. e.g.
File1:
UHDR 2008112
"25187","00000022","00",21-APR-1991,"" ,"D",-000000519,+0000000000,"C", ,+000000000,+000000000,000000000,"2","" ,21-APR-1991... (7 Replies)
Hi ,
i have files coming in my system which are very huge in MB and GBs, all these files are in a single line, there is no newline character.
I need to get only last 700 bytes of these files, of this i am splitting the files by "split -b 700 filename" but this gives all the splitted... (2 Replies)
Hi
I have to write a script to split the huge file into several pieces. The file columns is | pipe delimited. The data sample is as:
6625060|1420215|07308806|N|20100120|5572477081|+0002.79|+0000.00|0004|0001|......... (3 Replies)
Hi,
I have a requiement where in i will get a single file but there will be mutiple headers.
Suppose say for eg:
Header1
Data...
Data...
Header2
Data..
Data..
Header3
Data..
Data..
I want to split each with the corresponding data into a single file.
Please let me know how... (1 Reply)
I have a bif text file with the following format:
d1_03 fr:23
d1_03 fr:56
d1_03 fr:67
d1_03 fr:78
d1_01 fr:35
d1_01 fr:29
d1_01 fr:45
d2_09 fr:34
d2_09 fr:78
d3_98 fr:90
d3_98 fr:104
d3_98 fr:360
I have like thousands of such lines
I want to reformat this file based on column 1... (3 Replies)
Hi,
I have a huge file with a single line.
But I want to break that line into lines of with each line having five columns.
My file is like this:
code:
"hi","there","how","are","you?","It","was","great","working","with","you.","hope","to","work","you."
I want it like this:
code:... (1 Reply)
Hi all,
I am new to scripting and I have a requirement
we have source file as
HEADER 01.10.2010 14:32:37 NAYA
TA0022
TA0000
20000001;20060612;99991231;K4;02;3
20000008;20080624;99991231;K4;02;3
20000026;19840724;99991231;KK;01;3
20000027;19840724;99991231;KK;01;3... (6 Replies)
Hi i want to fetch 100k record from a file which is looking like as below.
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
... (17 Replies)
Hi Friends ,
Please guide me with the code to extract multiple files from one file .
The File Looks like ( Suppose a file has 2 tables list ,column length may vary )
H..- > File Header....
H....- >Table 1 Header....
D....- > Table 1 Data....
T....- >Table 1 Trailer....
H..-> Table 2... (1 Reply)
Discussion started by: AspiringD
1 Replies
LEARN ABOUT DEBIAN
american-english-huge
american-english-huge(5) Users' Manual american-english-huge(5)NAME
american-english-huge - a list of English words
DESCRIPTION
/usr/share/dict/american-english-huge is an ASCII file which contains an alphabetic list of words, one per line.
FILES
There may be any number of word lists in /usr/share/dict/. /etc/dictionaries-common/words is a symbolic link to the currently-chosen
/usr/share/dict/<language> file. /usr/share/dict/words is a symbolic link to /etc/dictionaries-common/words, and is the name by which
other software should refer to the system word list. See select-default-wordlist(8) for more information, and/or to change the currently-
chosen word list.
The directory /usr/share/dict can contain word lists for many languages, with name of the language in English, e.g., /usr/share/dict/french
and /usr/share/dict/danish contain respectively lists of French and Danish words if they exist. Such lists should be coded using the ISO
8859-1 character set encoding.
SEE ALSO ispell(1), select-default-wordlist(8), and the Filesystem Hierarchy Standard.
HISTORY
The words lists are not specific, and may be generated from any number of sources.
The system word list used to be /usr/dict/words. For compatibility, software should check that location if /usr/share/dict/words does not
exist.
AUTHOR
Word lists are collected and maintained by various authors. The Debian English word lists are built from the SCOWL (Spell- Checker Ori-
ented Word Lists) package, whose upstream editor is Kevin Atkinson <kevina@users.sourceforge.net>.
Debian 16 June 2003 american-english-huge(5)