04-01-2008
HTML5 will incorporate high definition extensions
Tue, 01 Apr 2008 10:30:00 GMT
PILTDOWN, UK - High definition HTML is coming to a site near you -- whether you will see it is another matter. An addition to the HTML5 working draft specification was approved by the
World Wide Web Consortium (W3C) working group over the weekend, including several new HTML elements and attributes designed to support output on high definition displays. But due to patent protection, open source Web browsers may not be allowed to implement a significant portion of the standard.
Source...
8 More Discussions You Might Find Interesting
1. UNIX for Advanced & Expert Users
Hi,
I am seeing very high kernel usage and very high load averages on my system (Although we are not loading much data to our database). Here is the output of top...does anyone know what i should be looking at?
Thanks,
Lorraine
last pid: 13144; load averages: 22.32, 19.81, 16.78 ... (4 Replies)
Discussion started by: lorrainenineill
4 Replies
2. OS X (Apple)
Would anyone here know how modify the appleHDA.kext for use with a Sigmatel HD Audio STAC9200? Help would be much appreciated. Thank you (0 Replies)
Discussion started by: Heather
0 Replies
3. UNIX for Dummies Questions & Answers
Could someone please direct me to a link that gives the definitions for each of the letters from the results of the $- environment variable? It would be nice to know what shell options each of the letters represents, but I am specifically looking for the shell option for 'c' (lowercase c). Thank... (12 Replies)
Discussion started by: sszd
12 Replies
4. Shell Programming and Scripting
Hi,
I am having a problem running multiple commands in a autosys command line. For example I would like to run a perl script first and then add sleep command after that in the same autosys job like below
insert_job: xxxxxxxxx
command: `perlscript.pl ; sleep 180`
Perlscript.pl... (0 Replies)
Discussion started by: waavman
0 Replies
5. Red Hat
i have a Intel Quad Core Xeon X3440 (4 x 2.53GHz, 8MB Cache, Hyper Threaded) with 16gig and 1tb harddrive with a 1gb port and my apache is causing my cpu to go up to 100% on all four cores heres my http.config
<IfModule prefork.c>
StartServers 10
MinSpareServers 10
MaxSpareServers 15... (4 Replies)
Discussion started by: awww
4 Replies
6. Shell Programming and Scripting
Hi ,
I am trying to :wall: my head while scripting ..I am really new to this stuff , never did it before :( .
how to find cpu's system high time and user time high in a script??
thanks , help would be appreciated !
:) (9 Replies)
Discussion started by: sushwey
9 Replies
7. What is on Your Mind?
A byte is the smallest unit of storage which can be accessed in a computer's memory- either in RAM or ROM.It also holds exactly 8 bits.But its old view one byte was sufficient to hold one 8 bit character.Modern days especially on .NET or international versions of Win 32, 16 bits is needed.
... (2 Replies)
Discussion started by: stoudtLion
2 Replies
8. What is on Your Mind?
You may notice some new headers and section labels on existing pages and sections.
I have been updating the site to be fully HTML5 compliant and have started to wrap various HTML templates for the site with HTML 5 elements like the ones below:
<article> Defines an article in a document... (7 Replies)
Discussion started by: Neo
7 Replies
LEARN ABOUT REDHAT
html::linkextor
HTML::LinkExtor(3) User Contributed Perl Documentation HTML::LinkExtor(3)
NAME
HTML::LinkExtor - Extract links from an HTML document
SYNOPSIS
require HTML::LinkExtor;
$p = HTML::LinkExtor->new(&cb, "http://www.perl.org/");
sub cb {
my($tag, %links) = @_;
print "$tag @{[%links]}
";
}
$p->parse_file("index.html");
DESCRIPTION
HTML::LinkExtor is an HTML parser that extracts links from an HTML document. The HTML::LinkExtor is a subclass of HTML::Parser. This means
that the document should be given to the parser by calling the $p->parse() or $p->parse_file() methods.
$p = HTML::LinkExtor->new([$callback[, $base]])
The constructor takes two optional arguments. The first is a reference to a callback routine. It will be called as links are found. If
a callback is not provided, then links are just accumulated internally and can be retrieved by calling the $p->links() method.
The $base argument is an optional base URL used to absolutize all URLs found. You need to have the URI module installed if you provide
$base.
The callback is called with the lowercase tag name as first argument, and then all link attributes as separate key/value pairs. All
non-link attributes are removed.
$p->links
Returns a list of all links found in the document. The returned values will be anonymous arrays with the follwing elements:
[$tag, $attr => $url1, $attr2 => $url2,...]
The $p->links method will also truncate the internal link list. This means that if the method is called twice without any parsing
between them the second call will return an empty list.
Also note that $p->links will always be empty if a callback routine was provided when the HTML::LinkExtor was created.
EXAMPLE
This is an example showing how you can extract links from a document received using LWP:
use LWP::UserAgent;
use HTML::LinkExtor;
use URI::URL;
$url = "http://www.perl.org/"; # for instance
$ua = LWP::UserAgent->new;
# Set up a callback that collect image links
my @imgs = ();
sub callback {
my($tag, %attr) = @_;
return if $tag ne 'img'; # we only look closer at <img ...>
push(@imgs, values %attr);
}
# Make the parser. Unfortunately, we don't know the base yet
# (it might be diffent from $url)
$p = HTML::LinkExtor->new(&callback);
# Request document and parse it as it arrives
$res = $ua->request(HTTP::Request->new(GET => $url),
sub {$p->parse($_[0])});
# Expand all image URLs to absolute ones
my $base = $res->base;
@imgs = map { $_ = url($_, $base)->abs; } @imgs;
# Print them out
print join("
", @imgs), "
";
SEE ALSO
HTML::Parser, HTML::Tagset, LWP, URI::URL
COPYRIGHT
Copyright 1996-2001 Gisle Aas.
This library is free software; you can redistribute it and/or modify it under the same terms as Perl itself.
perl v5.8.0 2001-04-10 HTML::LinkExtor(3)