From ba039cfb2e1ba2522ee0a0fa2a84a1a6579e4877 Mon Sep 17 00:00:00 2001 From: Unit 193 Date: Mon, 20 Jul 2020 01:03:23 -0400 Subject: New upstream version 1.14.3. --- data/completion/_gallery-dl | 56 +++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 56 insertions(+) create mode 100644 data/completion/_gallery-dl (limited to 'data/completion') diff --git a/data/completion/_gallery-dl b/data/completion/_gallery-dl new file mode 100644 index 0000000..b5074d2 --- /dev/null +++ b/data/completion/_gallery-dl @@ -0,0 +1,56 @@ +#compdef gallery-dl + +local curcontext="$curcontext" +typeset -A opt_args + +local rc=1 +_arguments -C -S \ +{-h,--help}'[Print this help message and exit]' \ +--version'[Print program version and exit]' \ +{-d,--dest}'[Destination directory]':'':_files \ +{-i,--input-file}'[Download URLs found in FILE ("-" for stdin)]':'':_files \ +--cookies'[File to load additional cookies from]':'':_files \ +--proxy'[Use the specified proxy]':'' \ +--clear-cache'[Delete all cached login sessions, cookies, etc.]' \ +{-q,--quiet}'[Activate quiet mode]' \ +{-v,--verbose}'[Print various debugging information]' \ +{-g,--get-urls}'[Print URLs instead of downloading]' \ +{-j,--dump-json}'[Print JSON information]' \ +{-s,--simulate}'[Simulate data extraction; do not download anything]' \ +{-K,--list-keywords}'[Print a list of available keywords and example values for the given URLs]' \ +--list-modules'[Print a list of available extractor modules]' \ +--list-extractors'[Print a list of extractor classes with description, (sub)category and example URL]' \ +--write-log'[Write logging output to FILE]':'':_files \ +--write-unsupported'[Write URLs, which get emitted by other extractors but cannot be handled, to FILE]':'':_files \ +--write-pages'[Write downloaded intermediary pages to files in the current directory to debug problems]' \ +{-r,--limit-rate}'[Maximum download rate (e.g. 500k or 2.5M)]':'' \ +{-R,--retries}'[Maximum number of retries for failed HTTP requests or -1 for infinite retries (default: 4)]':'' \ +{-A,--abort}'[Abort extractor run after N consecutive file downloads have been skipped, e.g. if files with the same filename already exist]':'' \ +--http-timeout'[Timeout for HTTP connections (default: 30.0)]':'' \ +--sleep'[Number of seconds to sleep before each download]':'' \ +--no-part'[Do not use .part files]' \ +--no-mtime'[Do not set file modification times according to Last-Modified HTTP response headers]' \ +--no-download'[Do not download any files]' \ +--no-check-certificate'[Disable HTTPS certificate validation]' \ +{-c,--config}'[Additional configuration files]':'':_files \ +--config-yaml'[==SUPPRESS==]':'':_files \ +{-o,--option}'[Additional "=" option values]':'' \ +--ignore-config'[Do not read the default configuration files]' \ +{-u,--username}'[Username to login with]':'' \ +{-p,--password}'[Password belonging to the given username]':'' \ +--netrc'[Enable .netrc authentication data]' \ +--download-archive'[Record all downloaded files in the archive file and skip downloading any file already in it.]':'':_files \ +--range'[Index-range(s) specifying which images to download. For example "5-10" or "1,3-5,10-"]':'' \ +--chapter-range'[Like "--range", but applies to manga-chapters and other delegated URLs]':'' \ +--filter'[Python expression controlling which images to download. Files for which the expression evaluates to False are ignored. Available keys are the filename-specific ones listed by "-K". Example: --filter "image_width >= 1000 and rating in ("s", "q")"]':'' \ +--chapter-filter'[Like "--filter", but applies to manga-chapters and other delegated URLs]':'' \ +--zip'[Store downloaded files in a ZIP archive]' \ +--ugoira-conv'[Convert Pixiv Ugoira to WebM (requires FFmpeg)]' \ +--ugoira-conv-lossless'[Convert Pixiv Ugoira to WebM in VP9 lossless mode]' \ +--write-metadata'[Write metadata to separate JSON files]' \ +--write-tags'[Write image tags to separate text files]' \ +--mtime-from-date'[Set file modification times according to "date" metadata]' \ +--exec'[Execute CMD for each downloaded file. Example: --exec "convert {} {}.png && rm {}"]':'' \ +--exec-after'[Execute CMD after all files were downloaded successfully. Example: --exec-after "cd {} && convert * ../doc.pdf"]':'' && rc=0 + +return rc -- cgit v1.2.3