Sponsored Content
Full Discussion: Website is up and running
Top Forums Shell Programming and Scripting Website is up and running Post 303009457 by looney on Thursday 14th of December 2017 02:59:00 PM
Old 12-14-2017
Website is up and running

I want to check if web url is up and running through command or script. I will use that output in one job to make it fail or succeed. I searched wget curl but they have lot of description. I don't know much on networking side.
Kindly help on this.
 

8 More Discussions You Might Find Interesting

1. Filesystems, Disks and Memory

website

HELLO FELLOW GEEKS. PLZ CHECK OUT MY FRIENDS SITE AT http://isunshine.dhs.org or u can also join the message board at http://isunshine.dhs.org/scripts/ikonboard.cgi wixifer (1 Reply)
Discussion started by: wixifer
1 Replies

2. UNIX for Dummies Questions & Answers

Website

Hey guys I know you probably get this question a lot but, I want to make a website, and I don't have any experience doing this. I have a iMac and i was wondering if there is someone you could refer me to or a site that will show me how to do it. Thanks. (2 Replies)
Discussion started by: mmecca21
2 Replies

3. Web Development

my website.please. help me.

hello!! well, i am planning to make my own virtual pet site like that of a neopets. unfortunately i don't have any idea on how to do it.. i've tried searching in the net, but the result is really complicated. i don't know where to begin.i have already drawn some that i think would help... (2 Replies)
Discussion started by: ackiemae
2 Replies

4. UNIX for Dummies Questions & Answers

website directory

hye ,first timer,anyways have a question,how can someone access a website root directory,i mean if i have a website www.xyz.com ,how can i access it directory, i mean how can i use /root /home /hom to access my website, what is the proper syntax ?? thx (1 Reply)
Discussion started by: waelkd
1 Replies

5. Shell Programming and Scripting

Website crawler

Hi, I want to build a crawler that seeks for a keyword on certain websites. This is what the website looks like: website.com/xxxxAA11xxxx I want that the crawler automatically changes the letters alphanumerically and if a certain keyword is found, the website got to be logged. But... (12 Replies)
Discussion started by: yaylol
12 Replies

6. UNIX for Dummies Questions & Answers

Getting the source of website

Hi MAy i know how i can get html source of a given site in a file using bash script? I want to store the html source of the site in a file so that i can parse that file. Thanks, (1 Reply)
Discussion started by: pandeesh
1 Replies

7. Web Development

Help with ht5boilerplate - website.

need a little help with this below MAVEN DAZZ what i need is ?!! a) To make sure, that logo stays on top (left hand side) -fixed in this navigation bar. on the same navigation bar (MENU) should stay on the other right hand side. Can anyone explain me how will this work. i am new to... (0 Replies)
Discussion started by: dazdseg
0 Replies

8. Web Development

Which technology is used on this website?

Hi, Home ~ InfoBeans I want to know technology framework using which we can achieve that kind of look and feel Thanks (6 Replies)
Discussion started by: ezee
6 Replies
CURLOPT_POSTFIELDS(3)					     curl_easy_setopt options					     CURLOPT_POSTFIELDS(3)

NAME
CURLOPT_POSTFIELDS - specify data to POST to server SYNOPSIS
#include <curl/curl.h> CURLcode curl_easy_setopt(CURL *handle, CURLOPT_POSTFIELDS, char *postdata); DESCRIPTION
Pass a char * as parameter, pointing to the full data to send in a HTTP POST operation. You must make sure that the data is formatted the way you want the server to receive it. libcurl will not convert or encode it for you in any way. For example, the web server may assume that this data is url-encoded. The data pointed to is NOT copied by the library: as a consequence, it must be preserved by the calling application until the associated transfer finishes. This behaviour can be changed (so libcurl does copy the data) by setting the CURLOPT_COPYPOSTFIELDS(3) option. This POST is a normal application/x-www-form-urlencoded kind (and libcurl will set that Content-Type by default when this option is used), which is commonly used by HTML forms. Change Content-Type with CURLOPT_HTTPHEADER(3). You can use curl_easy_escape(3) to url-encode your data, if necessary. It returns a pointer to an encoded string that can be passed as postdata. Using CURLOPT_POSTFIELDS(3) implies CURLOPT_POST(3). If CURLOPT_POSTFIELDS(3) is explicitly set to NULL then libcurl will get the POST data from the read callback. If you want to send a zero- byte POST set CURLOPT_POSTFIELDS(3) to an empty string, or set CURLOPT_POST(3) to 1 and CURLOPT_POSTFIELDSIZE(3) to 0. Using POST with HTTP 1.1 implies the use of a "Expect: 100-continue" header. You can disable this header with CURLOPT_HTTPHEADER(3) as usual. To make multipart/formdata posts (aka RFC2388-posts), check out the CURLOPT_HTTPPOST(3) option combined with curl_formadd(3). DEFAULT
NULL PROTOCOLS
HTTP EXAMPLE
CURL *curl = curl_easy_init(); if(curl) { const char *data = "data to send"; curl_easy_setopt(curl, CURLOPT_URL, "http://example.com"); /* size of the POST data */ curl_easy_setopt(curl, CURLOPT_POSTFIELDSIZE, 12L); /* pass in a pointer to the data - libcurl will not copy */ curl_easy_setopt(curl, CURLOPT_POSTFIELDS, data); curl_easy_perform(curl); } AVAILABILITY
Always RETURN VALUE
Returns CURLE_OK SEE ALSO
CURLOPT_POSTFIELDSIZE(3), CURLOPT_READFUNCTION(3), libcurl 7.54.0 June 11, 2016 CURLOPT_POSTFIELDS(3)
All times are GMT -4. The time now is 01:31 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy