Hi methyl, we had considered that option too but, since we are dealing with huge amount of data, so there is a possibility that it wud hamper the database.
your file has a unique key (like some sort of account id, order id etc.)
it is sorted by that key (in ascending order)
Here is an approach that I use when working with flat files and comparing data (its psuedo code):
The performance boost comes from the fact that each file is traversed exactly once in this approach.
You will have to handle boundary conditions (especially end-of-file and other errors properly in the code.
friends
good morning
FTP works perfect but I have a doubt
if I want to transport 10 files, I imagine that I should not open 10 connections as I can transfer more than 1 file?
ftp -n <<!EOF
open caburga
user ephfact ephfact
cd /users/efactura/docONE/entrada
bin
mput EPH`date... (16 Replies)
I have a script that I am using to copy around 40-70k files to a NFS NAS.
I have posted my code below in hopes that someone can help me figure out a faster way of achieving this.
At the end of the script i need to have all the files in the list, copied over to the nas with source directory... (8 Replies)
Hi all! I need help to do a few things with a .txt file using egrep.
1. I need to list all sequences where the vowel letters 'a, e, i, o, u' occur in that order, possibly separated by characters other than a, e, i, o, u; consisting of one or more complete words, possibly including punctuation.
... (1 Reply)
I have a txt file with several columns and i want to peform an operation on two columns and output it to a new txt file .
file.txt
900.00000 1 1 1
500.00000
500.00000
100000.000
4
4
1.45257346E-07 899.10834 ... (4 Replies)
- I am looking for different kind of awk solution which I don't think is mentioned before in these forums.
Number of rows in the file are fixed
Their are two columns in file1.txt
1 1
2 2
3 3
4 4
5 5
6 6
7 7
8 8
9 9
10 10
I am looking for 3... (1 Reply)