CURLOPT_HTTPGET(3) curl_easy_setopt options CURLOPT_HTTPGET(3)NAME
CURLOPT_HTTPGET - ask for a HTTP GET request
SYNOPSIS
#include <curl/curl.h>
CURLcode curl_easy_setopt(CURL *handle, CURLOPT_HTTPGET, long useget);
DESCRIPTION
Pass a long. If useget is 1, this forces the HTTP request to get back to using GET. Usable if a POST, HEAD, PUT, etc has been used previ-
ously using the same curl handle.
When setting CURLOPT_HTTPGET(3) to 1, it will automatically set CURLOPT_NOBODY(3) to 0 and CURLOPT_UPLOAD(3) to 0.
DEFAULT
0
PROTOCOLS
HTTP(S)
EXAMPLE
curl = curl_easy_init();
if(curl) {
curl_easy_setopt(curl, CURLOPT_URL, "http://example.com");
/* use a GET to fetch this */
curl_easy_setopt(curl, CURLOPT_HTTPGET, 1L);
/* Perform the request */
curl_easy_perform(curl);
}
AVAILABILITY
Along with HTTP
RETURN VALUE
Returns CURLE_OK if HTTP is supported, and CURLE_UNKNOWN_OPTION if not.
SEE ALSO CURLOPT_NOBODY(3), CURLOPT_UPLOAD(3),
libcurl 7.54.0 February 03, 2016 CURLOPT_HTTPGET(3)
Check Out this Related Man Page
CURLOPT_NOBODY(3) curl_easy_setopt options CURLOPT_NOBODY(3)NAME
CURLOPT_NOBODY - do the download request without getting the body
SYNOPSIS
#include <curl/curl.h>
CURLcode curl_easy_setopt(CURL *handle, CURLOPT_NOBODY, long opt);
DESCRIPTION
A long parameter set to 1 tells libcurl to not include the body-part in the output when doing what would otherwise be a download. For
HTTP(S), this makes libcurl do a HEAD request. For most other protocols it means just not asking to transfer the body data.
Enabling this option means asking for a download but without a body.
DEFAULT
0, the body is transferred
PROTOCOLS
Most
EXAMPLE
curl = curl_easy_init();
if(curl) {
curl_easy_setopt(curl, CURLOPT_URL, "http://example.com");
/* get us the resource without a body! */
curl_easy_setopt(curl, CURLOPT_NOBODY, 1L);
/* Perform the request */
curl_easy_perform(curl);
}
AVAILABILITY
Always
RETURN VALUE
Returns CURLE_OK
SEE ALSO CURLOPT_HTTPGET(3), CURLOPT_POST(3),
libcurl 7.54.0 February 03, 2016 CURLOPT_NOBODY(3)
I'm trying to upload a file to a page using curl and after uploading that file i want to store the redirecting page so i can download the results.
I'm using the command:
curl "http://apps.gdgps.net/kag_upload.php?kag_type=static&kag_frequency=dual&kag_latency=accurate&... (8 Replies)
I'm writing a program which uses curl to be run on Linux PCs which will be used by a number of different users. I cannot make the users all install curl on their individual machines, so I have tried to link curl in statically, rather than using libcurl.so. I downloaded the source and created a... (8 Replies)
hi!! this is my first post!! our university student search has this form which uses POST method(the website is swd.bits-goa.ac.in/searchform1)..now i tried sending post data from cURL...
curl -o this.html --data "name=rohan&id=&hostel=&room=&branch=&search=Search"... (9 Replies)
I am trying to get text from a webpage, in terminal form. So far I am:
1. getting the html for the page printed using curl (curl -s webpage.com), which is then
2. piped to awk, printing line number 29 (awk NR==29), then
3. this is where I am sort of lost. I know where in the printed line I... (7 Replies)
I have a script which uses cli curl to download the source code of a webpage and then tests if a specific string exists in the source.
The problem is that the website has a slow response, so the eval expression hasn't completed when the test starts.
The test returns a negative, and the curl... (8 Replies)
Greetings!!
am trying to retrieve a particular section from the url as in url.txt..
aim is to get the 83.8 MB as output, but somehow this is not happening!, please suggest what might be wrong.. attached is the screenshot and text file of the page source.
Best Regards,
Vinu (14 Replies)
Heyas
Trying to parse a tempfile, but somehow i mess up.
To my understand, this should work...
Plain:
tail -n1 out.tmp
1 81.5M 1 1066k 0 0 359k 0 0:03:52 0:00:02 0:03:50 359k
I want to get the 81.5M, so i'd assume it'll be $2 for awk....
tail -n1 out.tmp | awk... (24 Replies)
Hello.
I use curl to fetch a website, then, I want to extract the URLs from this curls output.
I tried both sed and grep, but couldnt figure it out.
Ive tried : sed -n 's/href="\(*\).*/\1/p' results.txt
and grep -o
grep -o '<a href="http://*.*.*/*">' results.txt.
What pattern... (6 Replies)
Hi
My goal is to fill an HTML form and submit.
What I have managed to do:
1. curl command to fill up the form and submit
2. a file which has the input
curl command:
curl -v -b cookie.txt -d __CSRFToken__=dc23d5da47953b3b390ec68d972af10380908b14 -d do=create -d a=open -d... (10 Replies)
I am trying to use REST API and curl in a bash script to generate a http report. The curl command at the end of the script should generate a html file but instead I get an error "HTTP/1.1 503 Service Unavailable".
This is the script
#!/bin/bash
export... (7 Replies)
Hello,
My question is about curl command. (ubuntu14.04)
In terminal, I am able to download my mainfile with:
curl -u user1:pass1 http://11.22.33.44/*******
When I convert it into bash script like this:
#!/bin/bash
cd /root/scripts
computer_ip=11.22.33.44
curl -u $1:$2... (8 Replies)
Hello:
I have created a REST service that is currently running on a Unix server. I issued a cURL command to call the REST service and verified it's working as designed; curl --request POST command and the url I used to test it is localhost:8090/someservice/somename. I have a teammate who now... (6 Replies)