05-14-2008
sorting huge file
Hi
I was able to run successfully when i set the variable TMPDIR to a location where the space is available.
Thanks for the information.
Regards
Dhana
10 More Discussions You Might Find Interesting
1. UNIX for Dummies Questions & Answers
I have a file {filename} which contains 65000 records
I need to split into 6 smaller files roughly 11000 records each.
Can someone advise me of the Unix command to do so ?
Many thanks (2 Replies)
Discussion started by: grinder182533
2 Replies
2. Solaris
Hi,
I have files with records of 40,00,000& 39,00,000 and i want to find out the
content
1.which is existing in file1 and not in file2.
2.Which is exisitng in file2 and not in file1.
The format of the file will be like
404ABCDEFGHIJK|CDEFGHIJK|1234567890|1
If its a smaller one i... (1 Reply)
Discussion started by: salaathi
1 Replies
3. Shell Programming and Scripting
I have a file with data extracted, and need to insert a header with a constant string, say: H|PayerDataExtract
if i use sed, i have to redirect the output to a seperate file like
sed ' sed commands' ExtractDataFile.dat > ExtractDataFileWithHeader.dat
the same is true for awk
and... (10 Replies)
Discussion started by: deepaktanna
10 Replies
4. Shell Programming and Scripting
Hi ,
i have files coming in my system which are very huge in MB and GBs, all these files are in a single line, there is no newline character.
I need to get only last 700 bytes of these files, of this i am splitting the files by "split -b 700 filename" but this gives all the splitted... (2 Replies)
Discussion started by: Prateek007
2 Replies
5. Shell Programming and Scripting
Hi i need to compare two fixed length files and produce the differences if any to a seperate file. I have to capture each and every differneces line by line. Ideally my files should not have any differences but if there are any then it should be captured without any miss. Also my files sizes are... (4 Replies)
Discussion started by: naveenn08
4 Replies
6. Shell Programming and Scripting
Dear All,
Good Evening!!
I have a requirement to ftp a 220GB backup file to a remote backup server.
I wrote a script for this purpose.
But it takes more than 8 hours to transfer this file.
Is there any other method to do it in less time???
Thanks in Advance!!!
---------- Post updated... (5 Replies)
Discussion started by: Naga06
5 Replies
7. Shell Programming and Scripting
Hi Experts,
I had to edit (a particular value) in header line of a very huge file so for that i wanted to search & replace a particular value on a file which was of 24 GB in Size. I managed to do it but it took long time to complete. Can anyone please tell me how can we do it in a optimised... (7 Replies)
Discussion started by: manishkomar007
7 Replies
8. AIX
Hi,
I need to transfer 2000 files from one host to another..
I modified /etc/security/limits to -1 and ulimit -f, ulimit -s, ulimit -a..
Even then only 700 files are transferred.
Could You please help me to sort out this issue..
I think some configuration related to memory is... (3 Replies)
Discussion started by: Priya Amaresh
3 Replies
9. UNIX for Dummies Questions & Answers
Please help. My file system is 100%, I can't seem to find what is taking so much space. The total hard drive space is 150Gig free but I got nothing now.
I did to this to find the big file but it's taking so much time. Is there any other way?
du -ah / | more
find ./ -size +200M... (3 Replies)
Discussion started by: samnyc
3 Replies
10. Shell Programming and Scripting
Hi.
Pardon me if I'm posting a duplicate thread but..
I have a text file with over 150 Million records, file size is in the range if MB(close to GB).
The requirement is to read ALL the lines excepting the FIRST LINE which is the file header and the LAST LINE which is it's trailer record.
... (8 Replies)
Discussion started by: kumarjt
8 Replies
LEARN ABOUT CENTOS
gethugepagesizes
GETHUGEPAGESIZES(3) Library Functions Manual GETHUGEPAGESIZES(3)
NAME
gethugepagesizes - Get the system supported huge page sizes
SYNOPSIS
#include <hugetlbfs.h>
int gethugepagesizes(long pagesizes[], int n_elem);
DESCRIPTION
The gethugepagesizes() function returns either the number of system supported huge page sizes or the sizes themselves. If pagesizes is
NULL and n_elem is 0, then the number of huge pages the system supports is returned. Otherwise, pagesizes is filled with at most n_elem
page sizes.
RETURN VALUE
On success, either the number of huge page sizes supported by the system or the number of huge page sizes stored in pagesizes is returned.
On failure, -1 is returned and errno is set appropriately.
ERRORS
EINVAL n_elem is less than zero or n_elem is greater than zero and pagesizes is NULL.
Also see opendir(3) for other possible values for errno. This error occurs when the sysfs directory exists but cannot be opened.
NOTES
This call will return all huge page sizes as reported by the kernel. Not all of these sizes may be usable by the programmer since mount
points may not be available for all sizes. To test whether a size will be usable by libhugetlbfs, hugetlbfs_find_path_for_size() can be
called on a specific size to see if a mount point is configured.
SEE ALSO
oprofile(1), opendir(3), hugetlbfs_find_path_for_size(3), libhugetlbfs(7)
AUTHORS
libhugetlbfs was written by various people on the libhugetlbfs-devel mailing list.
October 10, 2008 GETHUGEPAGESIZES(3)