wcurl: import v2025.09.27

Closes #18754
This commit is contained in:
Samuel Henrique 2025-09-27 09:42:08 +01:00 committed by Daniel Stenberg
parent 205758d7ea
commit a5a17b8ddb
No known key found for this signature in database
GPG Key ID: 5CC908FDB71E12C2
2 changed files with 36 additions and 27 deletions

View File

@ -45,6 +45,9 @@ By default, **wcurl** does:
## * Download multiple URLs in parallel
if the installed curl's version is \>= 7.66.0 (--parallel);
## * Use a total number of 5 parallel connections to the same protocol + hostname + port number target
if the installed curl's version is \>= 8.16.0 (--parallel-max-host);
## * Follow redirects;
## * Automatically choose a filename as output;
@ -124,10 +127,13 @@ Download a file passing the **--progress-bar** and **--http2** flags to curl:
**wcurl --curl-options="--progress-bar --http2" example.com/filename.txt**
Resume from an interrupted download (if more options are used, this needs to
be the last one in the list):
* Resume from an interrupted download. The options necessary to resume the download (`--clobber --continue-at -`) must be the **last** options specified in `--curl-options`. Note that the only way to resume interrupted downloads is to allow wcurl to overwrite the destination file:
**wcurl --curl-options="--continue-at -" example.com/filename.txt**
**wcurl --curl-options="--clobber --continue-at -" example.com/filename.txt**
Download multiple files without a limit of concurrent connections per host (the default limit is 5):
**wcurl --curl-options="--parallel-max-host 0" example.com/filename1.txt example.com/filename2.txt**
# AUTHORS

View File

@ -29,7 +29,7 @@
# Stop on errors and on usage of unset variables.
set -eu
VERSION="2025.05.26"
VERSION="2025.09.27"
PROGRAM_NAME="$(basename "$0")"
readonly PROGRAM_NAME
@ -91,7 +91,7 @@ error()
# Extra curl options provided by the user.
# This is set per-URL for every URL provided.
# Some options are global, but we are erroring on the side of needlesly setting
# Some options are global, but we are erroring on the side of needlessly setting
# them multiple times instead of causing issues with parameters that needs to
# be set per-URL.
CURL_OPTIONS=""
@ -133,7 +133,7 @@ sanitize()
is_subset_of()
{
case "${1}" in
*[!${2}]*|'') return 1;;
*[!${2}]* | '') return 1 ;;
esac
}
@ -151,9 +151,9 @@ percent_decode()
decode_out="${decode_out}${decode_hex2}"
# Skip decoding if this is a control character (00-1F).
# Skip decoding if DECODE_FILENAME is not "true".
if is_subset_of "${decode_hex1}" "23456789abcdefABCDEF" && \
is_subset_of "${decode_hex2}" "0123456789abcdefABCDEF" && \
[ "${DECODE_FILENAME}" = "true" ]; then
if is_subset_of "${decode_hex1}" "23456789abcdefABCDEF" \
&& is_subset_of "${decode_hex2}" "0123456789abcdefABCDEF" \
&& [ "${DECODE_FILENAME}" = "true" ]; then
# Use printf to decode it into octal and then decode it to the final format.
decode_out="$(printf "%b" "\\$(printf %o "0x${decode_hex1}${decode_hex2}")")"
fi
@ -171,7 +171,7 @@ get_url_filename()
# If what remains contains a slash, there's a path; return it percent-decoded.
case "${hostname_and_path}" in
# sed to remove everything preceding the last '/', e.g.: "example/something" becomes "something"
*/*) percent_decode "$(printf %s "${hostname_and_path}" | sed -e 's,^.*/,,')";;
*/*) percent_decode "$(printf %s "${hostname_and_path}" | sed -e 's,^.*/,,')" ;;
esac
# No slash means there was just a hostname and no path; return empty string.
}
@ -181,35 +181,38 @@ exec_curl()
{
CMD="curl "
# Store version to check if it supports --no-clobber and --parallel.
# Store version to check if it supports --no-clobber, --parallel and --parallel-max-host.
curl_version=$($CMD --version | cut -f2 -d' ' | head -n1)
curl_version_major=$(echo "$curl_version" | cut -f1 -d.)
curl_version_minor=$(echo "$curl_version" | cut -f2 -d.)
CURL_HAS_NO_CLOBBER=""
CURL_HAS_PARALLEL=""
CURL_NO_CLOBBER=""
CURL_PARALLEL=""
# --no-clobber is only supported since 7.83.0.
# --parallel is only supported since 7.66.0.
# --parallel-max-host is only supported since 8.16.0.
if [ "${curl_version_major}" -ge 8 ]; then
CURL_HAS_NO_CLOBBER="--no-clobber"
CURL_HAS_PARALLEL="--parallel"
elif [ "${curl_version_major}" -eq 7 ];then
CURL_NO_CLOBBER="--no-clobber"
CURL_PARALLEL="--parallel"
if [ "${curl_version_minor}" -ge 16 ]; then
CURL_PARALLEL="--parallel --parallel-max-host 5"
fi
elif [ "${curl_version_major}" -eq 7 ]; then
if [ "${curl_version_minor}" -ge 83 ]; then
CURL_HAS_NO_CLOBBER="--no-clobber"
CURL_NO_CLOBBER="--no-clobber"
fi
if [ "${curl_version_minor}" -ge 66 ]; then
CURL_HAS_PARALLEL="--parallel"
CURL_PARALLEL="--parallel"
fi
fi
# Detecting whether we need --parallel. It's easier to rely on
# Detecting whether we need --parallel. It's easier to rely on
# the shell's argument parsing.
# shellcheck disable=SC2086
set -- $URLS
if [ "$#" -gt 1 ]; then
CURL_PARALLEL="$CURL_HAS_PARALLEL"
else
# If there are less than two URLs, don't set the parallel flag.
if [ "$#" -lt 2 ]; then
CURL_PARALLEL=""
fi
@ -231,7 +234,7 @@ exec_curl()
[ -z "${OUTPUT_PATH}" ] && OUTPUT_PATH=index.html
fi
# shellcheck disable=SC2086
set -- "$@" ${NEXT_PARAMETER} ${PER_URL_PARAMETERS} ${CURL_HAS_NO_CLOBBER} ${CURL_OPTIONS} --output "${OUTPUT_PATH}" "${url}"
set -- "$@" ${NEXT_PARAMETER} ${PER_URL_PARAMETERS} ${CURL_NO_CLOBBER} --output "${OUTPUT_PATH}" ${CURL_OPTIONS} "${url}"
NEXT_PARAMETER="--next"
done
@ -268,13 +271,13 @@ while [ -n "${1-}" ]; do
OUTPUT_PATH="${opt}"
;;
-o|-O|--output)
-o | -O | --output)
shift
HAS_USER_SET_OUTPUT="true"
OUTPUT_PATH="${1}"
;;
-o*|-O*)
-o* | -O*)
opt=$(printf "%s\n" "${1}" | sed 's/^-[oO]//')
HAS_USER_SET_OUTPUT="true"
OUTPUT_PATH="${opt}"
@ -284,12 +287,12 @@ while [ -n "${1-}" ]; do
DECODE_FILENAME="false"
;;
-h|--help)
-h | --help)
usage
exit 0
;;
-V|--version)
-V | --version)
print_version
exit 0
;;