Hi Guy's can someone help me in converting the following
I have a flat text file which has several thousand lines which I need to convert to a csv it's got a consistent format but basically want every time it hit's txt to create a new line with the subsequent lines comma delimited for example
... (6 Replies)
thanks for allowing me join your forum
i have an output of linux command "who"
which provides following details.....
CURRENT USER/ACCT INFO
17:31:36 up 4:49, 4 users, load average: 0.03, 0.04, 0.00
USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT
root :0 - 12:59 ?xdm? 4:54 0.02s /bin/sh /usr/bi... (1 Reply)
INput file attached in thread : Column widths at 24,73,82,87,121 characters (sed 's/./,/24;s/./,/73;s/./,/81;s/./,/87;s/./,/121;s/ *, */,/g' fixedinputfile >output.csv
). The client wants instead of hard coding the column widths as they are not fixed .he has given the hint stating that ( ... (3 Replies)
thanks for allowing me to join your forum
i have a html file with three columns ------------Last visit date , URL and link,,,,,,,, how can i convert the same into csv so that i can output into database... the mechine is linux...i made a little googling and got idea that there is ways for... (2 Replies)
I have a CSV file in the below format.
while generating CSV file from excel sheet , date in excel sheet(Format :Mon 8/28/2012) got converted into the below format with numbers 41148,41149 so on.
Could anyone please let me know how to the convert the numbers(41148,41149 so on.) to its actual... (2 Replies)
Hello,
I have a binary csv file that was created on 'Red Hat Enterprise Linux Server release 6.6'. Now we have transferred all files on Ubuntu 16.04.2 LTS/xenial
On opening the file in Ubuntu, there are special characters
... (8 Replies)
Discussion started by: nans
8 Replies
LEARN ABOUT DEBIAN
data::stream::bulk::dbi
Data::Stream::Bulk::DBI(3pm) User Contributed Perl Documentation Data::Stream::Bulk::DBI(3pm)NAME
Data::Stream::Bulk::DBI - N-at-a-time iteration of DBI statement results.
VERSION
version 0.11
SYNOPSIS
use Data::Stream::Bulk::DBI;
my $sth = $dbh->prepare("SELECT hate FROM sql"); # very big resultset
$sth->execute;
return Data::Stream::Bulk::DBI->new(
sth => $sth,
max_rows => $n, # how many at a time
slice => [ ... ], # if you want to pass the first param to fetchall_arrayref
);
DESCRIPTION
This implementation of Data::Stream::Bulk api works with DBI statement handles, using "fetchall_arrayref" in DBI.
It fetches "max_rows" at a time (defaults to 500).
ATTRIBUTES
sth The statement handle to call "fetchall_arrayref" on.
slice
Passed verbatim as the first param to "fetchall_arrayref". Should usually be "undef", provided for completetness.
max_rows
The second param to "fetchall_arrayref". Controls the size of each buffer.
Defaults to 500.
METHODS
get_more
See Data::Stream::Bulk::DoneFlag.
Calls "fetchall_arrayref" to get the next chunk of rows.
all Calls "fetchall_arrayref" to get the raminder of the data (without specifying "max_rows").
AUTHOR
Yuval Kogman <nothingmuch@woobling.org>
COPYRIGHT AND LICENSE
This software is copyright (c) 2012 by Yuval Kogman.
This is free software; you can redistribute it and/or modify it under the same terms as the Perl 5 programming language system itself.
perl v5.14.2 2012-02-14 Data::Stream::Bulk::DBI(3pm)