CURLOPT_NOBODY(3) curl_easy_setopt options CURLOPT_NOBODY(3)NAME
CURLOPT_NOBODY - do the download request without getting the body
SYNOPSIS
#include <curl/curl.h>
CURLcode curl_easy_setopt(CURL *handle, CURLOPT_NOBODY, long opt);
DESCRIPTION
A long parameter set to 1 tells libcurl to not include the body-part in the output when doing what would otherwise be a download. For
HTTP(S), this makes libcurl do a HEAD request. For most other protocols it means just not asking to transfer the body data.
Enabling this option means asking for a download but without a body.
DEFAULT
0, the body is transferred
PROTOCOLS
Most
EXAMPLE
curl = curl_easy_init();
if(curl) {
curl_easy_setopt(curl, CURLOPT_URL, "http://example.com");
/* get us the resource without a body! */
curl_easy_setopt(curl, CURLOPT_NOBODY, 1L);
/* Perform the request */
curl_easy_perform(curl);
}
AVAILABILITY
Always
RETURN VALUE
Returns CURLE_OK
SEE ALSO CURLOPT_HTTPGET(3), CURLOPT_POST(3),
libcurl 7.54.0 February 03, 2016 CURLOPT_NOBODY(3)
Check Out this Related Man Page
CURLOPT_NOPROGRESS(3) curl_easy_setopt options CURLOPT_NOPROGRESS(3)NAME
CURLOPT_NOPROGRESS - switch off the progress meter
SYNOPSIS
#include <curl/curl.h>
CURLcode curl_easy_setopt(CURL *handle, CURLOPT_NOPROGRESS, long onoff);
DESCRIPTION
If onoff is to 1, it tells the library to shut off the progress meter completely for requests done with this handle. It will also prevent
the CURLOPT_PROGRESSFUNCTION(3) from getting called.
Future versions of libcurl are likely to not have any built-in progress meter at all.
DEFAULT
1, meaning it normally runs without a progress meter.
PROTOCOLS
All
EXAMPLE
CURL *curl = curl_easy_init();
if(curl) {
curl_easy_setopt(curl, CURLOPT_URL, "http://example.com");
/* enable progress meter */
curl_easy_setopt(curl, CURLOPT_NOPROGRESS, 0L);
/* Perform the request */
curl_easy_perform(curl);
}
AVAILABILITY
Always
RETURN VALUE
Returns CURLE_OK.
SEE ALSO CURLOPT_PROGRESSFUNCTION(3),
libcurl 7.54.0 March 06, 2016 CURLOPT_NOPROGRESS(3)
Aren't there any way to download files as below?
For example, I want to download all .html files under the root directory of unix.com/
curl -O https://www.unix.com/*.html
This won't work, but please tell me the way to do this.
Well, the best way is to get the file list of the directory, but i... (6 Replies)
I'm trying to upload a file to a page using curl and after uploading that file i want to store the redirecting page so i can download the results.
I'm using the command:
curl "http://apps.gdgps.net/kag_upload.php?kag_type=static&kag_frequency=dual&kag_latency=accurate&... (8 Replies)
I'm writing a program which uses curl to be run on Linux PCs which will be used by a number of different users. I cannot make the users all install curl on their individual machines, so I have tried to link curl in statically, rather than using libcurl.so. I downloaded the source and created a... (8 Replies)
I am trying to get text from a webpage, in terminal form. So far I am:
1. getting the html for the page printed using curl (curl -s webpage.com), which is then
2. piped to awk, printing line number 29 (awk NR==29), then
3. this is where I am sort of lost. I know where in the printed line I... (7 Replies)
I have a script which uses cli curl to download the source code of a webpage and then tests if a specific string exists in the source.
The problem is that the website has a slow response, so the eval expression hasn't completed when the test starts.
The test returns a negative, and the curl... (8 Replies)
Greetings!!
am trying to retrieve a particular section from the url as in url.txt..
aim is to get the 83.8 MB as output, but somehow this is not happening!, please suggest what might be wrong.. attached is the screenshot and text file of the page source.
Best Regards,
Vinu (14 Replies)
Hi,
For an order I requested, the provider has uploaded a tar file in public FTP site which internally has tons of files (compressed) and I need to download files that follows particular pattern which would be few hundreds.
Note: The order can't be requested for files that follows the... (7 Replies)
Heyas
Trying to parse a tempfile, but somehow i mess up.
To my understand, this should work...
Plain:
tail -n1 out.tmp
1 81.5M 1 1066k 0 0 359k 0 0:03:52 0:00:02 0:03:50 359k
I want to get the 81.5M, so i'd assume it'll be $2 for awk....
tail -n1 out.tmp | awk... (24 Replies)
Hello.
I use curl to fetch a website, then, I want to extract the URLs from this curls output.
I tried both sed and grep, but couldnt figure it out.
Ive tried : sed -n 's/href="\(*\).*/\1/p' results.txt
and grep -o
grep -o '<a href="http://*.*.*/*">' results.txt.
What pattern... (6 Replies)
Hello guys, first post sorry if I did some mess here =)
Using Ubuntu 14.04lts 64bits server version.
I have a list (url.list) with only URLs to download, one per line, that looks like this:
http://domain.com/teste.php?a=2&b=3&name=1
http://domain.com/teste.php?a=2&b=3&name=2
...... (6 Replies)
I am trying to use REST API and curl in a bash script to generate a http report. The curl command at the end of the script should generate a html file but instead I get an error "HTTP/1.1 503 Service Unavailable".
This is the script
#!/bin/bash
export... (7 Replies)
Hello,
My question is about curl command. (ubuntu14.04)
In terminal, I am able to download my mainfile with:
curl -u user1:pass1 http://11.22.33.44/*******
When I convert it into bash script like this:
#!/bin/bash
cd /root/scripts
computer_ip=11.22.33.44
curl -u $1:$2... (8 Replies)