diff options
| author | 2019-07-02 04:33:45 -0400 | |
|---|---|---|
| committer | 2019-07-02 04:33:45 -0400 | |
| commit | 195c45911e79c33cf0bb986721365fb06df5a153 (patch) | |
| tree | ac0c9b6ef40bea7aa7ab0c5c3cb500eb510668fa /gallery_dl/job.py | |
Import Upstream version 1.8.7upstream/1.8.7
Diffstat (limited to 'gallery_dl/job.py')
| -rw-r--r-- | gallery_dl/job.py | 492 |
1 files changed, 492 insertions, 0 deletions
diff --git a/gallery_dl/job.py b/gallery_dl/job.py new file mode 100644 index 0000000..667b9b3 --- /dev/null +++ b/gallery_dl/job.py @@ -0,0 +1,492 @@ +# -*- coding: utf-8 -*- + +# Copyright 2015-2019 Mike Fährmann +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License version 2 as +# published by the Free Software Foundation. + +import sys +import time +import logging +from . import extractor, downloader, postprocessor +from . import config, text, util, output, exception +from .extractor.message import Message + + +class Job(): + """Base class for Job-types""" + ulog = None + + def __init__(self, extr, parent=None): + if isinstance(extr, str): + extr = extractor.find(extr) + if not extr: + raise exception.NoExtractorError() + + self.extractor = extr + extr.log.extractor = extr + extr.log.job = self + extr.log.debug("Using %s for '%s'", extr.__class__.__name__, extr.url) + + # url predicates + self.pred_url = self._prepare_predicates( + "image", [util.UniquePredicate()], True) + + # queue predicates + self.pred_queue = self._prepare_predicates( + "chapter", [], False) + + # category transfer + if parent and parent.extractor.config( + "category-transfer", parent.extractor.categorytransfer): + self.extractor.category = parent.extractor.category + self.extractor.subcategory = parent.extractor.subcategory + + # user-supplied metadata + self.userkwds = self.extractor.config("keywords") + + def run(self): + """Execute or run the job""" + try: + log = self.extractor.log + for msg in self.extractor: + self.dispatch(msg) + except exception.AuthenticationError as exc: + msg = str(exc) or "Please provide a valid username/password pair." + log.error("Authentication failed: %s", msg) + except exception.AuthorizationError: + log.error("You do not have permission to access the resource " + "at '%s'", self.extractor.url) + except exception.NotFoundError as exc: + res = str(exc) or "resource (gallery/image/user)" + log.error("The %s at '%s' does not exist", res, self.extractor.url) + except exception.HttpError as exc: + err = exc.args[0] + if isinstance(err, Exception): + err = "{}: {}".format(err.__class__.__name__, err) + log.error("HTTP request failed: %s", err) + except exception.FormatError as exc: + err, obj = exc.args + log.error("Applying %s format string failed: %s: %s", + obj, err.__class__.__name__, err) + except exception.FilterError as exc: + err = exc.args[0] + log.error("Evaluating filter expression failed: %s: %s", + err.__class__.__name__, err) + except exception.StopExtraction: + pass + except OSError as exc: + log.error("Unable to download data: %s: %s", + exc.__class__.__name__, exc) + log.debug("", exc_info=True) + except Exception as exc: + log.error(("An unexpected error occurred: %s - %s. " + "Please run gallery-dl again with the --verbose flag, " + "copy its output and report this issue on " + "https://github.com/mikf/gallery-dl/issues ."), + exc.__class__.__name__, exc) + log.debug("", exc_info=True) + self.handle_finalize() + + def dispatch(self, msg): + """Call the appropriate message handler""" + if msg[0] == Message.Url: + _, url, kwds = msg + if self.pred_url(url, kwds): + self.update_kwdict(kwds) + self.handle_url(url, kwds) + + elif msg[0] == Message.Directory: + self.update_kwdict(msg[1]) + self.handle_directory(msg[1]) + + elif msg[0] == Message.Queue: + _, url, kwds = msg + if self.pred_queue(url, kwds): + self.handle_queue(url, kwds) + + elif msg[0] == Message.Urllist: + _, urls, kwds = msg + if self.pred_url(urls[0], kwds): + self.update_kwdict(kwds) + self.handle_urllist(urls, kwds) + + elif msg[0] == Message.Version: + if msg[1] != 1: + raise "unsupported message-version ({}, {})".format( + self.extractor.category, msg[1] + ) + # TODO: support for multiple message versions + + def handle_url(self, url, keywords): + """Handle Message.Url""" + + def handle_urllist(self, urls, keywords): + """Handle Message.Urllist""" + self.handle_url(urls[0], keywords) + + def handle_directory(self, keywords): + """Handle Message.Directory""" + + def handle_queue(self, url, keywords): + """Handle Message.Queue""" + + def handle_finalize(self): + """Handle job finalization""" + + def update_kwdict(self, kwdict): + """Update 'kwdict' with additional metadata""" + kwdict["category"] = self.extractor.category + kwdict["subcategory"] = self.extractor.subcategory + if self.userkwds: + kwdict.update(self.userkwds) + + def _prepare_predicates(self, target, predicates, skip=True): + pfilter = self.extractor.config(target + "-filter") + if pfilter: + try: + pred = util.FilterPredicate(pfilter, target) + except (SyntaxError, ValueError, TypeError) as exc: + self.extractor.log.warning(exc) + else: + predicates.append(pred) + + prange = self.extractor.config(target + "-range") + if prange: + try: + pred = util.RangePredicate(prange) + except ValueError as exc: + self.extractor.log.warning( + "invalid %s range: %s", target, exc) + else: + if skip and pred.lower > 1 and not pfilter: + pred.index += self.extractor.skip(pred.lower - 1) + predicates.append(pred) + + return util.build_predicate(predicates) + + def _write_unsupported(self, url): + if self.ulog: + self.ulog.info(url) + + @staticmethod + def _filter(kwdict): + """Return a copy of 'kwdict' without "private" entries""" + return {k: v for k, v in kwdict.items() if k[0] != "_"} + + +class DownloadJob(Job): + """Download images into appropriate directory/filename locations""" + + def __init__(self, url, parent=None): + Job.__init__(self, url, parent) + self.log = logging.getLogger("download") + self.pathfmt = None + self.archive = None + self.sleep = None + self.downloaders = {} + self.postprocessors = None + self.out = output.select() + + def handle_url(self, url, keywords, fallback=None): + """Download the resource specified in 'url'""" + # prepare download + self.pathfmt.set_keywords(keywords) + + if self.postprocessors: + for pp in self.postprocessors: + pp.prepare(self.pathfmt) + + if self.pathfmt.exists(self.archive): + self.handle_skip() + return + + if self.sleep: + time.sleep(self.sleep) + + # download from URL + if not self.download(url): + + # use fallback URLs if available + for num, url in enumerate(fallback or (), 1): + self.log.info("Trying fallback URL #%d", num) + if self.download(url): + break + else: + # download failed + self.log.error( + "Failed to download %s", self.pathfmt.filename or url) + return + + if not self.pathfmt.temppath: + self.handle_skip() + return + + # run post processors + if self.postprocessors: + for pp in self.postprocessors: + pp.run(self.pathfmt) + + # download succeeded + self.pathfmt.finalize() + self.out.success(self.pathfmt.path, 0) + if self.archive: + self.archive.add(keywords) + self._skipcnt = 0 + + def handle_urllist(self, urls, keywords): + """Download the resource specified in 'url'""" + fallback = iter(urls) + url = next(fallback) + self.handle_url(url, keywords, fallback) + + def handle_directory(self, keywords): + """Set and create the target directory for downloads""" + if not self.pathfmt: + self.initialize(keywords) + else: + self.pathfmt.set_directory(keywords) + + def handle_queue(self, url, keywords): + if "_extractor" in keywords: + extr = keywords["_extractor"].from_url(url) + else: + extr = extractor.find(url) + if extr: + self.__class__(extr, self).run() + else: + self._write_unsupported(url) + + def handle_finalize(self): + if self.postprocessors: + for pp in self.postprocessors: + pp.finalize() + + def handle_skip(self): + self.out.skip(self.pathfmt.path) + if self._skipexc: + self._skipcnt += 1 + if self._skipcnt >= self._skipmax: + raise self._skipexc() + + def download(self, url): + """Download 'url'""" + scheme = url.partition(":")[0] + downloader = self.get_downloader(scheme) + if downloader: + return downloader.download(url, self.pathfmt) + self._write_unsupported(url) + return False + + def get_downloader(self, scheme): + """Return a downloader suitable for 'scheme'""" + if scheme == "https": + scheme = "http" + try: + return self.downloaders[scheme] + except KeyError: + pass + + klass = downloader.find(scheme) + if klass and config.get(("downloader", scheme, "enabled"), True): + instance = klass(self.extractor, self.out) + else: + instance = None + self.log.error("'%s:' URLs are not supported/enabled", scheme) + self.downloaders[scheme] = instance + return instance + + def initialize(self, keywords=None): + """Delayed initialization of PathFormat, etc.""" + self.pathfmt = util.PathFormat(self.extractor) + if keywords: + self.pathfmt.set_directory(keywords) + self.sleep = self.extractor.config("sleep") + + skip = self.extractor.config("skip", True) + if skip: + self._skipexc = None + if isinstance(skip, str): + skip, _, smax = skip.partition(":") + if skip == "abort": + self._skipexc = exception.StopExtraction + elif skip == "exit": + self._skipexc = sys.exit + self._skipcnt = 0 + self._skipmax = text.parse_int(smax) + else: + self.pathfmt.exists = lambda x=None: False + + archive = self.extractor.config("archive") + if archive: + path = util.expand_path(archive) + self.archive = util.DownloadArchive(path, self.extractor) + + postprocessors = self.extractor.config("postprocessors") + if postprocessors: + self.postprocessors = [] + for pp_dict in postprocessors: + whitelist = pp_dict.get("whitelist") + blacklist = pp_dict.get("blacklist") + if (whitelist and self.extractor.category not in whitelist or + blacklist and self.extractor.category in blacklist): + continue + name = pp_dict.get("name") + pp_cls = postprocessor.find(name) + if not pp_cls: + postprocessor.log.warning("module '%s' not found", name) + continue + try: + pp_obj = pp_cls(self.pathfmt, pp_dict) + except Exception as exc: + postprocessor.log.error( + "'%s' initialization failed: %s: %s", + name, exc.__class__.__name__, exc) + else: + self.postprocessors.append(pp_obj) + self.extractor.log.debug( + "Active postprocessor modules: %s", self.postprocessors) + + +class SimulationJob(DownloadJob): + """Simulate the extraction process without downloading anything""" + + def handle_url(self, url, keywords, fallback=None): + self.pathfmt.set_keywords(keywords) + self.out.skip(self.pathfmt.path) + if self.sleep: + time.sleep(self.sleep) + if self.archive: + self.archive.add(keywords) + + def handle_directory(self, keywords): + if not self.pathfmt: + self.initialize() + + +class KeywordJob(Job): + """Print available keywords""" + + def handle_url(self, url, keywords): + print("\nKeywords for filenames and --filter:") + print("------------------------------------") + self.print_keywords(keywords) + raise exception.StopExtraction() + + def handle_directory(self, keywords): + print("Keywords for directory names:") + print("-----------------------------") + self.print_keywords(keywords) + + def handle_queue(self, url, keywords): + if not keywords: + self.extractor.log.info( + "This extractor delegates work to other extractors " + "and does not provide any keywords on its own. Try " + "'gallery-dl -K \"%s\"' instead.", url) + else: + print("Keywords for --chapter-filter:") + print("------------------------------") + self.print_keywords(keywords) + if self.extractor.categorytransfer: + print() + KeywordJob(url, self).run() + raise exception.StopExtraction() + + @staticmethod + def print_keywords(keywords, prefix=""): + """Print key-value pairs with formatting""" + suffix = "]" if prefix else "" + for key, value in sorted(keywords.items()): + if key[0] == "_": + continue + key = prefix + key + suffix + + if isinstance(value, dict): + KeywordJob.print_keywords(value, key + "[") + + elif isinstance(value, list): + if value and isinstance(value[0], dict): + KeywordJob.print_keywords(value[0], key + "[][") + else: + print(key, "[]", sep="") + for val in value: + print(" -", val) + + else: + # string or number + print(key, "\n ", value, sep="") + + +class UrlJob(Job): + """Print download urls""" + maxdepth = 1 + + def __init__(self, url, parent=None, depth=1): + Job.__init__(self, url, parent) + self.depth = depth + if depth >= self.maxdepth: + self.handle_queue = self.handle_url + + @staticmethod + def handle_url(url, _): + print(url) + + @staticmethod + def handle_urllist(urls, _): + prefix = "" + for url in urls: + print(prefix, url, sep="") + prefix = "| " + + def handle_queue(self, url, _): + try: + UrlJob(url, self, self.depth + 1).run() + except exception.NoExtractorError: + self._write_unsupported(url) + + +class DataJob(Job): + """Collect extractor results and dump them""" + + def __init__(self, url, parent=None, file=sys.stdout, ensure_ascii=True): + Job.__init__(self, url, parent) + self.file = file + self.data = [] + self.ascii = config.get(("output", "ascii"), ensure_ascii) + + def run(self): + # collect data + try: + for msg in self.extractor: + self.dispatch(msg) + except exception.StopExtraction: + pass + except Exception as exc: + self.data.append((exc.__class__.__name__, str(exc))) + except BaseException: + pass + + # convert numbers to string + if config.get(("output", "num-to-str"), False): + for msg in self.data: + util.transform_dict(msg[-1], util.number_to_string) + + # dump to 'file' + util.dump_json(self.data, self.file, self.ascii, 2) + + def handle_url(self, url, kwdict): + self.data.append((Message.Url, url, self._filter(kwdict))) + + def handle_urllist(self, urls, kwdict): + self.data.append((Message.Urllist, list(urls), self._filter(kwdict))) + + def handle_directory(self, kwdict): + self.data.append((Message.Directory, self._filter(kwdict))) + + def handle_queue(self, url, kwdict): + self.data.append((Message.Queue, url, self._filter(kwdict))) + + def handle_finalize(self): + self.file.close() |
