---------- Post updated at 08:08 AM ---------- Previous update was at 08:01 AM ----------
Does this mean Perl 5.8.6 doesnt support LWP ... I dont think so , but I'm not a computer expert ... please could you provide more info to solve this problem
Cheers
my host run on a Free bsd server and i have a cgi script that requires LWP module, but i my host say that that module is installed, so i would like to know if the is a command because i have telnet access to know if the module or which modules are installed on my account (itīs that the script donīt... (4 Replies)
My company has a product that is running on JBoss on Solaris against Oracle 8.1.7. We are having an issue with the server process and high CPU utilization. During this time, and only during this time, we are experiencing database locks that will not let go. A 'ps -L' on the server process... (5 Replies)
Hi all,
Without oracle tunning I understand that oracle processes always run on a single CPU at a time (the prstat output always show 1 CPU number) even on multi-processor hardware.
Is it correct?
Regards, (1 Reply)
Hi,
I've a issue in writing a perl script that will automatically monitor the site availability.
There are two different cookies are set in two consecutive flows to a URL and this second cookie has to be passed to the third step which actually gives permission to access based upon the cookie. ... (1 Reply)
Hi,
Could some one tell me how to login to any web site and get that page using perl LWP. I heard that we can login to the site using LWP.
I dont want to use WWW:Mechanize as I dont have that module installed on the server.
Appreciate your early response.
Thanks... (8 Replies)
Hello all.. i am new to perl scripting.. i wanted to parse a text file, encode the parsed text and attach in url.. please point me to right resources if you know any..This is my major problem. Now i try to get a url running and save it in a text file using LWP module in perl, I used following... (0 Replies)
Hi;
Can i have ne sample examples of of using " lwp-request" in shell script.
Is it necessary to have perl installed already in linux box for using this;
Thnks; (2 Replies)
I am requesting for the text parsing section below. Any helps are highly appreciated.
<tr valign="top"><td nowrap>Source name</td>
<td style="text-align: justify">Sample Name<br></td>
I want Sample Name from above.
In the same file, I have to search for another pattern like this
<td><a... (1 Reply)
Hi,
I would like to read content from a https site.
I have decided to use LWP module in perl. but it throwed 401 Authorization required error.
i dont want to hard code the password in my perl code. Is there any way to achieve the authentication without hardcoding the password.
Thanks,... (1 Reply)
Discussion started by: pandeesh
1 Replies
LEARN ABOUT OSX
html::linkextor
HTML::LinkExtor(3) User Contributed Perl Documentation HTML::LinkExtor(3)NAME
HTML::LinkExtor - Extract links from an HTML document
SYNOPSIS
require HTML::LinkExtor;
$p = HTML::LinkExtor->new(&cb, "http://www.perl.org/");
sub cb {
my($tag, %links) = @_;
print "$tag @{[%links]}
";
}
$p->parse_file("index.html");
DESCRIPTION
HTML::LinkExtor is an HTML parser that extracts links from an HTML document. The HTML::LinkExtor is a subclass of HTML::Parser. This means
that the document should be given to the parser by calling the $p->parse() or $p->parse_file() methods.
$p = HTML::LinkExtor->new
$p = HTML::LinkExtor->new( $callback )
$p = HTML::LinkExtor->new( $callback, $base )
The constructor takes two optional arguments. The first is a reference to a callback routine. It will be called as links are found. If
a callback is not provided, then links are just accumulated internally and can be retrieved by calling the $p->links() method.
The $base argument is an optional base URL used to absolutize all URLs found. You need to have the URI module installed if you provide
$base.
The callback is called with the lowercase tag name as first argument, and then all link attributes as separate key/value pairs. All
non-link attributes are removed.
$p->links
Returns a list of all links found in the document. The returned values will be anonymous arrays with the following elements:
[$tag, $attr => $url1, $attr2 => $url2,...]
The $p->links method will also truncate the internal link list. This means that if the method is called twice without any parsing
between them the second call will return an empty list.
Also note that $p->links will always be empty if a callback routine was provided when the HTML::LinkExtor was created.
EXAMPLE
This is an example showing how you can extract links from a document received using LWP:
use LWP::UserAgent;
use HTML::LinkExtor;
use URI::URL;
$url = "http://www.perl.org/"; # for instance
$ua = LWP::UserAgent->new;
# Set up a callback that collect image links
my @imgs = ();
sub callback {
my($tag, %attr) = @_;
return if $tag ne 'img'; # we only look closer at <img ...>
push(@imgs, values %attr);
}
# Make the parser. Unfortunately, we don't know the base yet
# (it might be different from $url)
$p = HTML::LinkExtor->new(&callback);
# Request document and parse it as it arrives
$res = $ua->request(HTTP::Request->new(GET => $url),
sub {$p->parse($_[0])});
# Expand all image URLs to absolute ones
my $base = $res->base;
@imgs = map { $_ = url($_, $base)->abs; } @imgs;
# Print them out
print join("
", @imgs), "
";
SEE ALSO
HTML::Parser, HTML::Tagset, LWP, URI::URL
COPYRIGHT
Copyright 1996-2001 Gisle Aas.
This library is free software; you can redistribute it and/or modify it under the same terms as Perl itself.
perl v5.16.2 2011-10-15 HTML::LinkExtor(3)