I have a file which has a list of 200 tables e.g: table.txt
I need to do a count for each table and store it in a file.
So I did something like this:
The above script does the job, but it goes through each table one at a time.
Is there a way to make this process more parallel, so that I can query multiple tables at a time?
I was thinking of doing a split. Something like
The above line would create 10 splits, so I could query 10 tables at a time.
But I'm not sure what my next step is so that 10 Select clauses are issued at smae time.
I bet you'd get much better speedup by connecting to the database once instead of 200 times, and accomplish this without DDOS-ing your database server.
I am writing a script to kick off a process to gather logs on multiple nodes in parallel using "&". These processes create individual log files. Which I would like to filter and convert in CSV format after they are complete. I am facing following issues:
1. Monitor all Processes parallelly.... (5 Replies)
I have few very huge files ~ 2 Billion rows of 130 column(CDR data) in a folder, I have written shell script need to read on each file in a folder and will create a new files based on some logic.
But problem is it's taking time to create a new file due to the size , So i dont want to corrupt... (6 Replies)
HI All,
I have scenerio where I need to call sub modules through for loop
for (i=0; i<8000 ;i++)
{
..
BLOCKA
}
BLOCKA
{
..
..
subroutine 1;
subroutine 2;
}
I want this to be run in parallel
process1 BLOCKA
{ (6 Replies)
Hi All,
I have two ksh script. 1st script calls the 2nd script and the second script calls an 'C' program.
I want 1st script to wait until the 'C' program completes.
I cant able to get the process id for the 'C' program (child process) to make the 1st script to wait for the second... (7 Replies)
Hi,
Is there any way to run parallel jobs using make command?
I am using non-GNU make utility on AIX 4.3.
I want to run 2 jobs symultaneously using make utility.
Thanks.
Suman (0 Replies)
Hi
I need to split a huge file into multiple smaller files using split command.
After that i need to process each file in the back ground with sql loader .Sql loader is a utlity to load CSV files into oracle .
Check the status of each of these sqlloaders and then after sucessfull... (6 Replies)
Hi guys:
I have a an oracle job which uses 10 parallel hints and would like to killit when it hangs. I want to kill all the processes that have been spawned. what I do right now is get the pid of the scheduler process which initiated theis job and the do a ps -ef| grep 'pid' and trace through... (1 Reply)