12-18-2006
How to get the page size (of a url) using wget
Hi ,
I am trying to get page size of a url(e.g.,
www.example.com) using wget command.Any thoughts what are the parameters i need to send with wget to get the size alone?
Regards,
Raj
10 More Discussions You Might Find Interesting
1. Shell Programming and Scripting
I all,
I wrote an script which starts a Weblogic server and waits until its loaded to deploy several apps. The way I checked was something like:
while ; do
wget --spider <URL>:<port>/console > /dev/null 2>&1
rc=$?
done
This works perfectly because it's an HTML site and when server is... (2 Replies)
Discussion started by: AlbertGM
2 Replies
2. UNIX for Dummies Questions & Answers
So, I'd like to wget a webpage, as its not going to stick around forever - but the problem is the webpage has a semicolon in it.
wget http://example.com/stuff/asdf;asdf obviously doesn't get the right webpage.
Any good way around this? (2 Replies)
Discussion started by: Julolidine
2 Replies
3. Shell Programming and Scripting
Hello,
I am experiencing an issue while downloading a few pages using wget. All of them work without a problem except one which is a page that does a tail on the log and as a result is constantly getting updated.
wget here seems to run endlessly and needs to be manually killed. I wanted to... (0 Replies)
Discussion started by: prafulnama
0 Replies
4. Shell Programming and Scripting
for example, I have an html file, contain
<a href="http://awebsite" id="awebsite" class="first">website</a>and sometime a line contains more then one link, for example
<a href="http://awebsite" id="awebsite" class="first">website</a><a href="http://bwebsite" id="bwebsite"... (36 Replies)
Discussion started by: 14th
36 Replies
5. UNIX for Dummies Questions & Answers
Hi All,
I want to launch "ex: http://gmail.com" from the cmd window and validate the credentials with username and password, is it possible?
I have found something like this
"wget --http-user=USER' --http-password=PASSWORD http://gmail.com" am new to this and unable to find a solution, i... (0 Replies)
Discussion started by: harsha85
0 Replies
6. Shell Programming and Scripting
Good evening to all!!
I'm trying to become familiar with wget.
I would like to download a page from Wikipedia with all images and CSSs but without going down to all links present in the page. It should be named index.html.
I would like also to save it to /mnt/us inside a new folder.
This is... (5 Replies)
Discussion started by: silver18
5 Replies
7. UNIX for Dummies Questions & Answers
Hi Experts,
Problem statement :
We have an URL for which we need to read the data and get parsed inside the shell scripts.
My Aix has very limited perl utility, i cant install any utility as well.
Precisely, wget,cURL,Lynx,w3m and Lwp cant be used as i got these utilities only when i googled... (0 Replies)
Discussion started by: scott_cog
0 Replies
8. Shell Programming and Scripting
Hi Experts,
Problem statement :
We have an URL for which we need to read the data and get parsed inside the shell scripts.My Aix has very limited perl utility, i cant install any utility as well.
Precisely, wget,cURL,Lynx,w3m and Lwp cant be used as i got these utilities only when i googled it.... (12 Replies)
Discussion started by: scott_cog
12 Replies
9. Post Here to Contact Site Administrators and Moderators
Hi
I just tried to post following link while answering, its not parsing properly, just try on your browser
Tried to paste while answering :
https://www.unix.com/302873559-post2.htmlNot operator is not coming with HTML/PHP tags so attaching file (2 Replies)
Discussion started by: Akshay Hegde
2 Replies
10. Shell Programming and Scripting
Wget Error Codes:
0 No problems occurred.
1 Generic error code.
2 Parse error—for instance, when parsing command-line options, the .wgetrc or .netrc…
3 File I/O error.
4 Network failure.
5 SSL verification failure.
6 Username/password authentication failure.
... (3 Replies)
Discussion started by: mohtashims
3 Replies
LEARN ABOUT CENTOS
uri::url
URI::URL(3) User Contributed Perl Documentation URI::URL(3)
NAME
URI::URL - Uniform Resource Locators
SYNOPSIS
$u1 = URI::URL->new($str, $base);
$u2 = $u1->abs;
DESCRIPTION
This module is provided for backwards compatibility with modules that depend on the interface provided by the "URI::URL" class that used to
be distributed with the libwww-perl library.
The following differences exist compared to the "URI" class interface:
o The URI::URL module exports the url() function as an alternate constructor interface.
o The constructor takes an optional $base argument. The "URI::URL" class is a subclass of "URI::WithBase".
o The URI::URL->newlocal class method is the same as URI::file->new_abs.
o URI::URL::strict(1)
o $url->print_on method
o $url->crack method
o $url->full_path: same as ($uri->abs_path || "/")
o $url->netloc: same as $uri->authority
o $url->epath, $url->equery: same as $uri->path, $uri->query
o $url->path and $url->query pass unescaped strings.
o $url->path_components: same as $uri->path_segments (if you don't consider path segment parameters)
o $url->params and $url->eparams methods
o $url->base method. See URI::WithBase.
o $url->abs and $url->rel have an optional $base argument. See URI::WithBase.
o $url->frag: same as $uri->fragment
o $url->keywords: same as $uri->query_keywords
o $url->localpath and friends map to $uri->file.
o $url->address and $url->encoded822addr: same as $uri->to for mailto URI
o $url->groupart method for news URI
o $url->article: same as $uri->message
SEE ALSO
URI, URI::WithBase
COPYRIGHT
Copyright 1998-2000 Gisle Aas.
perl v5.16.3 2012-02-11 URI::URL(3)