From ba039cfb2e1ba2522ee0a0fa2a84a1a6579e4877 Mon Sep 17 00:00:00 2001 From: Unit 193 Date: Mon, 20 Jul 2020 01:03:23 -0400 Subject: New upstream version 1.14.3. --- data/completion/_gallery-dl | 56 +++++++++++++++++++++++++++++++++++++++++++++ data/man/gallery-dl.1 | 2 +- data/man/gallery-dl.conf.5 | 36 ++++++++++++++++++++++++++++- 3 files changed, 92 insertions(+), 2 deletions(-) create mode 100644 data/completion/_gallery-dl (limited to 'data') diff --git a/data/completion/_gallery-dl b/data/completion/_gallery-dl new file mode 100644 index 0000000..b5074d2 --- /dev/null +++ b/data/completion/_gallery-dl @@ -0,0 +1,56 @@ +#compdef gallery-dl + +local curcontext="$curcontext" +typeset -A opt_args + +local rc=1 +_arguments -C -S \ +{-h,--help}'[Print this help message and exit]' \ +--version'[Print program version and exit]' \ +{-d,--dest}'[Destination directory]':'':_files \ +{-i,--input-file}'[Download URLs found in FILE ("-" for stdin)]':'':_files \ +--cookies'[File to load additional cookies from]':'':_files \ +--proxy'[Use the specified proxy]':'' \ +--clear-cache'[Delete all cached login sessions, cookies, etc.]' \ +{-q,--quiet}'[Activate quiet mode]' \ +{-v,--verbose}'[Print various debugging information]' \ +{-g,--get-urls}'[Print URLs instead of downloading]' \ +{-j,--dump-json}'[Print JSON information]' \ +{-s,--simulate}'[Simulate data extraction; do not download anything]' \ +{-K,--list-keywords}'[Print a list of available keywords and example values for the given URLs]' \ +--list-modules'[Print a list of available extractor modules]' \ +--list-extractors'[Print a list of extractor classes with description, (sub)category and example URL]' \ +--write-log'[Write logging output to FILE]':'':_files \ +--write-unsupported'[Write URLs, which get emitted by other extractors but cannot be handled, to FILE]':'':_files \ +--write-pages'[Write downloaded intermediary pages to files in the current directory to debug problems]' \ +{-r,--limit-rate}'[Maximum download rate (e.g. 500k or 2.5M)]':'' \ +{-R,--retries}'[Maximum number of retries for failed HTTP requests or -1 for infinite retries (default: 4)]':'' \ +{-A,--abort}'[Abort extractor run after N consecutive file downloads have been skipped, e.g. if files with the same filename already exist]':'' \ +--http-timeout'[Timeout for HTTP connections (default: 30.0)]':'' \ +--sleep'[Number of seconds to sleep before each download]':'' \ +--no-part'[Do not use .part files]' \ +--no-mtime'[Do not set file modification times according to Last-Modified HTTP response headers]' \ +--no-download'[Do not download any files]' \ +--no-check-certificate'[Disable HTTPS certificate validation]' \ +{-c,--config}'[Additional configuration files]':'':_files \ +--config-yaml'[==SUPPRESS==]':'':_files \ +{-o,--option}'[Additional "=" option values]':'' \ +--ignore-config'[Do not read the default configuration files]' \ +{-u,--username}'[Username to login with]':'' \ +{-p,--password}'[Password belonging to the given username]':'' \ +--netrc'[Enable .netrc authentication data]' \ +--download-archive'[Record all downloaded files in the archive file and skip downloading any file already in it.]':'':_files \ +--range'[Index-range(s) specifying which images to download. For example "5-10" or "1,3-5,10-"]':'' \ +--chapter-range'[Like "--range", but applies to manga-chapters and other delegated URLs]':'' \ +--filter'[Python expression controlling which images to download. Files for which the expression evaluates to False are ignored. Available keys are the filename-specific ones listed by "-K". Example: --filter "image_width >= 1000 and rating in ("s", "q")"]':'' \ +--chapter-filter'[Like "--filter", but applies to manga-chapters and other delegated URLs]':'' \ +--zip'[Store downloaded files in a ZIP archive]' \ +--ugoira-conv'[Convert Pixiv Ugoira to WebM (requires FFmpeg)]' \ +--ugoira-conv-lossless'[Convert Pixiv Ugoira to WebM in VP9 lossless mode]' \ +--write-metadata'[Write metadata to separate JSON files]' \ +--write-tags'[Write image tags to separate text files]' \ +--mtime-from-date'[Set file modification times according to "date" metadata]' \ +--exec'[Execute CMD for each downloaded file. Example: --exec "convert {} {}.png && rm {}"]':'' \ +--exec-after'[Execute CMD after all files were downloaded successfully. Example: --exec-after "cd {} && convert * ../doc.pdf"]':'' && rc=0 + +return rc diff --git a/data/man/gallery-dl.1 b/data/man/gallery-dl.1 index 21055ca..f05f2e8 100644 --- a/data/man/gallery-dl.1 +++ b/data/man/gallery-dl.1 @@ -1,4 +1,4 @@ -.TH "GALLERY-DL" "1" "2020-06-27" "1.14.2" "gallery-dl Manual" +.TH "GALLERY-DL" "1" "2020-07-18" "1.14.3" "gallery-dl Manual" .\" disable hyphenation .nh diff --git a/data/man/gallery-dl.conf.5 b/data/man/gallery-dl.conf.5 index 7060751..039e750 100644 --- a/data/man/gallery-dl.conf.5 +++ b/data/man/gallery-dl.conf.5 @@ -1,4 +1,4 @@ -.TH "GALLERY-DL.CONF" "5" "2020-06-27" "1.14.2" "gallery-dl Manual" +.TH "GALLERY-DL.CONF" "5" "2020-07-18" "1.14.3" "gallery-dl Manual" .\" disable hyphenation .nh .\" disable justification (adjust text to left margin only) @@ -290,6 +290,8 @@ and optional for .br * \f[I]sankaku\f[] .br +* \f[I]subscribestar\f[] +.br * \f[I]tsumino\f[] .br * \f[I]twitter\f[] @@ -836,6 +838,22 @@ Minimum wait time in seconds before API requests. Note: This value will internally be rounded up to the next power of 2. +.SS extractor.exhentai.domain +.IP "Type:" 6 +\f[I]string\f[] + +.IP "Default:" 9 +\f[I]"auto"\f[] + +.IP "Description:" 4 +.br +* \f[I]"auto"\f[]: Use \f[I]e-hentai.org\f[] or \f[I]exhentai.org\f[] +depending on the input URL +.br +* \f[I]"e-hentai.org"\f[]: Use \f[I]e-hentai.org\f[] for all URLs +.br +* \f[I]"exhentai.org"\f[]: Use \f[I]exhentai.org\f[] for all URLs + .SS extractor.exhentai.limits .IP "Type:" 6 \f[I]bool\f[] @@ -1009,6 +1027,22 @@ Include *Story Highlights* when downloading a user profile. .IP "Description:" 4 Download video files. +.SS extractor.khinsider.format +.IP "Type:" 6 +\f[I]string\f[] + +.IP "Default:" 9 +\f[I]"mp3"\f[] + +.IP "Description:" 4 +The name of the preferred file format to download. + +Use \f[I]"all"\f[] to download all available formats, +or a (comma-separated) list to select multiple formats. + +If the selected format is not available, +the first in the list gets chosen (usually mp3). + .SS extractor.kissmanga.captcha .IP "Type:" 6 \f[I]string\f[] -- cgit v1.2.3