Sponsored Content
Operating Systems Linux Red Hat Getting the response code and to know whether web site is loaded Post 302906204 by cnamejj on Wednesday 18th of June 2014 02:51:52 AM
Old 06-18-2014
Other people might have the same question I do, which makes it difficult to suggestion a solution. I'm not sure what "fully loaded" means?

If you just want to know if the website is up and responding to HTTP requests, you can just use "curl" (or another HTTP client command) to pull a URL that indicates that everything is working OK.

But if "fully loaded" means something else, like some initialization has completed, then unless the URL you pick somehow verifies that the initialization is done, then it may not give you the info you want.
This User Gave Thanks to cnamejj For This Post:
 

8 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Web site setup

Hi, I have a m/c loaded with redhat linux, static IP address and this m/c intern connected to the internet thro leaseline. Now i need to setup a company website has www.abc.xyz.com. What necessary configuration i need to do so that others can view our website. Thanks Bache Gowda (5 Replies)
Discussion started by: bache_gowda
5 Replies

2. UNIX for Dummies Questions & Answers

Web site mirroring

I have a unix server running RedHat Linux 9.0. I want to mirror my site on a different server with the same configuration so if the main server goes down, the other server will take over. How is this best accomplished? (1 Reply)
Discussion started by: wvmlt
1 Replies

3. AIX

intranet web site

hello I would like to create an internal web site, but how do i make it with Aix ? I must to install apache and send my http pages in the declared repertory ? There are some configuration files to modify ?? thank you (2 Replies)
Discussion started by: pascalbout
2 Replies

4. Shell Programming and Scripting

I need help in creating my first web site?

I have always wanted to create my own site, I've only had one class about the basics in creating web pages but know i would like to get into more and more detailed info about doing so anyone with alot of experience please help me out.:b: (0 Replies)
Discussion started by: kprescod4158
0 Replies

5. Shell Programming and Scripting

Checking response of the web page?

Hi, I need to check the response of the web page without downloading any content. I used the below command. _website=http://xxx.xxx.com wget --spider --user=username --password=password $_website > /home/username/temp.txt 2>&1; I getting the below error. The web page require... (3 Replies)
Discussion started by: ahamed
3 Replies

6. UNIX for Advanced & Expert Users

Checking response from a web page

wget --spider --user=xxxx --password=xxxx "http://xxx.xxxx.com" > /dev/null 2>&1; I am using the above command in if loop to check the response of the page without downloading the page. I just want to check whether the page is up and running. But when i execute the command i am getting HTTP... (2 Replies)
Discussion started by: ahamed
2 Replies

7. Web Development

Apache Web Server - Invalid Response

Hi, I have a SCO Unix Openserver V6 server which is hosting a website with Apache V1.3 as the http server. The web site has an initial login screen which re-directs to another page once the user name and password has been verified. When connecting to the website and trying to login, it times... (0 Replies)
Discussion started by: Martyn
0 Replies

8. IP Networking

Tshark/pcap and web-server response time

Hi everyone! How can I get response time difference between GET and HTTP/1.0 200 OK (i mean time latency of web-server) with using of tshark&shell or something else for each hostname from pcap file? What can you recommend me to do that? (1 Reply)
Discussion started by: lepetal
1 Replies
PHTTPGET(8)						    BSD System Manager's Manual 					       PHTTPGET(8)

NAME
phttpget -- retrieve multiple files via pipelined HTTP SYNOPSIS
phttpget server file ... DESCRIPTION
The phttpget utility is a minimalist pipelined HTTP client, which is used to retrieve multiple files from one server, and saves the down- loaded files in the current working directory, using the last portion of their download path as file names. By making several "in flight" HTTP requests, it can dramatically increase performance when a large number of small files need to be down- loaded. The freebsd-update(8) and portnap(8) tools use phttpget to download binary patch files. ENVIRONMENT
HTTP_PROXY URL of the proxy to use for HTTP requests. HTTP_PROXY_AUTH Authorization parameters for the HTTP proxy. HTTP_USER_AGENT The User-Agent string to use for HTTP requests. The default is ``phttpget/0.1''. HTTP_TIMEOUT Timeout for HTTP request in seconds. SEE ALSO
fetch(1), freebsd-update(8), portsnap(8) AUTHORS
The phttpget utility was written by Colin Percival <cperciva@FreeBSD.org> for use with portsnap(8) and later with freebsd-update(8). This manual page was written by Xin LI <delphij@FreeBSD.org>. BSD
January 3, 2015 BSD
All times are GMT -4. The time now is 08:34 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy