Data::Stream::Bulk::DBI(3pm) User Contributed Perl Documentation Data::Stream::Bulk::DBI(3pm)NAME
Data::Stream::Bulk::DBI - N-at-a-time iteration of DBI statement results.
VERSION
version 0.11
SYNOPSIS
use Data::Stream::Bulk::DBI;
my $sth = $dbh->prepare("SELECT hate FROM sql"); # very big resultset
$sth->execute;
return Data::Stream::Bulk::DBI->new(
sth => $sth,
max_rows => $n, # how many at a time
slice => [ ... ], # if you want to pass the first param to fetchall_arrayref
);
DESCRIPTION
This implementation of Data::Stream::Bulk api works with DBI statement handles, using "fetchall_arrayref" in DBI.
It fetches "max_rows" at a time (defaults to 500).
ATTRIBUTES
sth The statement handle to call "fetchall_arrayref" on.
slice
Passed verbatim as the first param to "fetchall_arrayref". Should usually be "undef", provided for completetness.
max_rows
The second param to "fetchall_arrayref". Controls the size of each buffer.
Defaults to 500.
METHODS
get_more
See Data::Stream::Bulk::DoneFlag.
Calls "fetchall_arrayref" to get the next chunk of rows.
all Calls "fetchall_arrayref" to get the raminder of the data (without specifying "max_rows").
AUTHOR
Yuval Kogman <nothingmuch@woobling.org>
COPYRIGHT AND LICENSE
This software is copyright (c) 2012 by Yuval Kogman.
This is free software; you can redistribute it and/or modify it under the same terms as the Perl 5 programming language system itself.
perl v5.14.2 2012-02-14 Data::Stream::Bulk::DBI(3pm)
Check Out this Related Man Page
Data::Stream::Bulk::Util(3pm) User Contributed Perl Documentation Data::Stream::Bulk::Util(3pm)NAME
Data::Stream::Bulk::Util - Utility functions for Data::Stream::Bulk
VERSION
version 0.11
SYNOPSIS
use Data::Stream::Bulk::Util qw(array);
use namespace::clean;
# Wrap a list in L<Data::Stream::Bulk::Array>
return bulk(qw(foo bar gorch baz));
# return an empty resultset
return nil();
DESCRIPTION
This module exports convenience functions for use with Data::Stream::Bulk.
EXPORTS
Sub::Exporter is used to create the "import" routine, and all of its aliasing/currying goodness is of course supported.
nil Creates a new Data::Stream::Bulk::Nil object.
Takes no arguments.
bulk @items
Creates a new Data::Stream::Bulk::Array wrapping @items.
cat @streams
Concatenate several streams together.
Returns "nil" if no arguments are provided.
filter { ... } $stream
Calls "filter" on $stream with the provided filter.
unique $stream
Filter the stream to remove duplicates.
Note that memory use may potentially scale to O(k) where k is the number of distinct items, because this is implemented in terms of a
seen hash.
In the future this will be optimized to be iterative for sorted streams.
References are keyed by their refaddr (see "id" in Hash::Util::FieldHash).
AUTHOR
Yuval Kogman <nothingmuch@woobling.org>
COPYRIGHT AND LICENSE
This software is copyright (c) 2012 by Yuval Kogman.
This is free software; you can redistribute it and/or modify it under the same terms as the Perl 5 programming language system itself.
perl v5.14.2 2012-02-14 Data::Stream::Bulk::Util(3pm)
Is there any possibility that a Stream Read and Write queues will interchange messages of any kind. If so what are the different possiblites and under what circumstances ?
Thanks in advance. (4 Replies)
Hi Yogesh,
Lucky that i caught you online. Yeah i read about DBI and the WriteExcel module. But the server is not supporting these modules. It said..."Cannot locate DBI"..."Cannot locate Spreadsheet::WriteExcel"
I tried creating a simple text file to get the query output, but the... (1 Reply)
Hi all,
I am using the following code in order to access the data of a table:
my $sql = qq{ SELECT * FROM cc_test_cases};
$sth = $dbh->prepare( $sql );
$sth->execute( );
while(@row1 = $sth->fetchrow_array()) {
# print "$row1: $row1\n";
print "@row1\n";
#print("THE VALUE... (1 Reply)
I try to prepare a sufficient function to execute a sql-statement, getting back ONLY retrieved results!
What I can't figured out how to make the sqlplus not printing the 'Connected to ...', 'Disconnected from...' and the executed statements after 'SQL> '.
I am under impression that having... (5 Replies)
Hi there, I wonder if somebody could help me out, i have a written a DBI perl snippet using fetchall_arrayref() that i want to use to populate a simple standard array with the results of a query
#!/usr/bin/perl -w
use DBI;
my $dbh =... (3 Replies)
Under smit, one has to manually select each fix with F7. there 9000 fixes left to be marked. How Can I manually install/Mark all of these without SMIT.
---------- Post updated at 02:29 PM ---------- Previous update was at 01:15 PM ----------
From the command line
instfix -T -d... (4 Replies)
here's my statement:
sed -i '4s/$/ #/' sed_test_file.txt
I want to edit the n'th line (the '4' is for the example) of the file (but you can't just put the n instead of the 4)
sed_test_file is a simple text file containig text only
I tried to use $urlCounter instead of 4, as urlCounter... (4 Replies)
A report needs to come some what similar to this
No of elements Stream Batch No Load time
A B C D
A,B,C im able to get quite easily
wc -l /usr/local/intranet/areas/prod/output/SRGW_0?/*/MESSAGE_T.dat
O/P of above command.
A B C ... (1 Reply)
I'm trying to write a script that can compile my students' homework submissions in bulk. My students' c code is buried in a file path that looks like this:
./Homework\ X/Doe, John/Submission\ Attachments
Where I'm struggling is determining how to navigate to each of the submission attachment... (11 Replies)
Hello,
I am trying to spool data from database into a file on solaris through ksh.
Data is getting fetched but the problem is record is getting split in to multiple lines.
excerpt from sql is
whenever sqlerror exit 1;
set define on
set echo off
set feed off
set head off
set... (1 Reply)
Hi All,
I have same file with like 1xyz_1 ,1xyz_2 , 2ghj_1, 2ghj_2, 4wer_1, 4wer1_2 etc with lots of file having similar type of name. I want to creat folder and copy similar file to that folder. for example folder 1xyz should contain only 1xyz_1 ,1xyz_2. :mad:
Thank you in Advance.
... (7 Replies)
Hi All, :D
Actullay I am looking for a smart way :b: to parse files in a directory whose count is around 2000000 :eek: in a single day.
Find is working with me but taking a lot of times :confused:, sometimes even a day which is not helping me.:wall:
So anyone can help me know a smart... (5 Replies)