11-24-2009
There are various ways ..
lynx -dump 'URL'
or
links -dump 'URL'
or
wget -dump 'URL'
or curl or telnet port 80 or ....
8 More Discussions You Might Find Interesting
1. UNIX for Advanced & Expert Users
Hi all,
I need to write a unix script in which need to call a url.
Then need to pass parameters to that url.
please help.
Regards,
gander_ss (1 Reply)
Discussion started by: gander_ss
1 Replies
2. Shell Programming and Scripting
Hi all,
I need to write a unix script in which need to call a url.
Then need to pass parameters to that url.
please help.
Regards,
gander_ss (1 Reply)
Discussion started by: gander_ss
1 Replies
3. UNIX for Dummies Questions & Answers
Hello,
I need to redirect an existing URL, how can i do that?
There's a current web address to a GUI that I have to redirect to another webaddress. Does anyone know how to do this?
This is on Unix boxes Linux.
example:
https://m45.testing.address.net/host.php
make it so the... (3 Replies)
Discussion started by: SkySmart
3 Replies
4. Web Development
I am trying to find a way to test some code, but I need to rewrite a specific URL only from a specific HTTP_HOST
The call goes out to
http://SUB.DOMAIN.COM/showAssignment/7bde10b45efdd7a97629ef2fe01f7303/jsmodule/Nevow.Athena
The ID in the middle is always random due to the cookie.
I... (5 Replies)
Discussion started by: EXT3FSCK
5 Replies
5. Red Hat
HI,
When ever I try to access a URL containing '%26' in my application, it is not rendered.
If I replace the '%26' with '&' character, it works fine.
I am using Apache as my webserver for my application.
Could you please help me with a rewrite rule that resolves the issue?
Thanks in... (1 Reply)
Discussion started by: BSrikanthB
1 Replies
6. UNIX for Advanced & Expert Users
Hi,
I am trying to invoke an service using URL. I want to know how to call that url with nohup.
nohup links "__http://Administrator:assword@ServName:8080/invoke/wm.server.admin/shutdown?bounce=no&option=force&timeout=0" &
I am trying to run this command on unix command prompt, But in... (8 Replies)
Discussion started by: swap27
8 Replies
7. UNIX for Dummies Questions & Answers
Here is what I have so far:
find . -name "*php*" -or -name "*htm*" | xargs grep -i iframe | awk -F'"' '/<iframe*/{gsub(/.\*iframe>/,"\"");print $2}'
Here is an example content of a PHP or HTM(HTML) file:
<iframe src="http://ADDRESS_1/?click=5BBB08\" width=1 height=1... (18 Replies)
Discussion started by: striker4o
18 Replies
8. Shell Programming and Scripting
Hello,
Am very new to perl , please help me here !!
I need help in reading a URL from command line using PERL:: Mechanize and needs all the contents from the URL to get into a file.
below is the script which i have written so far ,
#!/usr/bin/perl
use LWP::UserAgent;
use... (2 Replies)
Discussion started by: scott_cog
2 Replies
LEARN ABOUT DEBIAN
hxcopy
HXCOPY(1) HTML-XML-utils HXCOPY(1)
NAME
hxcopy - copy an HTML file and update its relative links
SYNOPSIS
hxcopy [ -i old-URL ] [ -o new-URL ] [ file-or-URL [ file-or-URL ] ]
DESCRIPTION
The hxcopy command copies its first argument to its second argument, while updating relative links. The input is assumed to be HTML or
XHTML and may be slightly reformatted in the process.
If the second argument is omitted, hxcopy writes to standard output. In this case the option -o is required. If the first argument is also
omitted, hxcopy reads from standard input. In this case the option -i is required.
OPTIONS
The following options are supported:
-i old-URL
For the purposes of updating relative links, act as if old-URL is the location from which the input is copied. If this option is
omitted, the actual location of the first argument is used for calculating relative links.
-o new-URL
For the purposed of updating relative links, act as if new-URL is the location to which the input is copied. If this option is
omitted, the actual location of the second argument is used for calculating relative links.
ENVIRONMENT
To use a proxy to retrieve remote files, set the environment variables http_proxy and ftp_proxy. E.g., http_proxy="http://localhost:8080/"
BUGS
Unlike the last argument of cp(1), the last argument of hxcopy must be a file, not a directory.
The second argument must be a local file. Writing to a URL is not yet implemented. To work around this, replace hxcopy file.html
http://example.org/file.html by hxcopy -o http://example.org/file.html file.html tmp.html and then upload tmp.html to the given URL with
some other command, such as curl(1). The first argument, however, may be a URL. hxcopy will download the given file. (Currently only HTTP
is supported.)
EXAMPLE
Assume the HTML file foo.html contains a relative link to "../bar.html". Here are some examples of commands:
hxcopy foo.html bar/foo.html
The file foo.html is copied to ../bar/foo.html and the relative link to "../bar.html" becomes "../../bar.html".
hxcopy foo.html ../foo.html
The file foo.html is copied to ../foo.html and the relative link to "../bar.html" is rewritten as "bar.html".
hxcopy -i http://my.org/dir1/foo.html -o http://my.org/foo.html file1.html file2.html
The file file1.html is copied to file2.html and the relative link to "../bar.html" is rewritten as "bar.html". A command like this
may be useful to update files that are later uploaded to a server.
SEE ALSO
cp(1), curl(1), hxwls(1)
6.x 9 Dec 2008 HXCOPY(1)