summaryrefslogtreecommitdiffstats
path: root/gallery_dl/__init__.py
diff options
context:
space:
mode:
Diffstat (limited to 'gallery_dl/__init__.py')
-rw-r--r--gallery_dl/__init__.py255
1 files changed, 255 insertions, 0 deletions
diff --git a/gallery_dl/__init__.py b/gallery_dl/__init__.py
new file mode 100644
index 0000000..3643a5c
--- /dev/null
+++ b/gallery_dl/__init__.py
@@ -0,0 +1,255 @@
+# -*- coding: utf-8 -*-
+
+# Copyright 2014-2019 Mike Fährmann
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License version 2 as
+# published by the Free Software Foundation.
+
+from __future__ import unicode_literals, print_function
+
+__author__ = "Mike Fährmann"
+__copyright__ = "Copyright 2014-2018 Mike Fährmann"
+__license__ = "GPLv2"
+__maintainer__ = "Mike Fährmann"
+__email__ = "mike_faehrmann@web.de"
+
+import sys
+
+if sys.hexversion < 0x3040000:
+ sys.exit("Python 3.4+ required")
+
+import json
+import logging
+from . import version, config, option, output, extractor, job, util, exception
+
+__version__ = version.__version__
+
+
+def progress(urls, pformat):
+ """Wrapper around urls to output a simple progress indicator"""
+ if pformat is True:
+ pformat = "[{current}/{total}] {url}"
+ pinfo = {"total": len(urls)}
+ for pinfo["current"], pinfo["url"] in enumerate(urls, 1):
+ print(pformat.format_map(pinfo), file=sys.stderr)
+ yield pinfo["url"]
+
+
+def parse_inputfile(file, log):
+ """Filter and process strings from an input file.
+
+ Lines starting with '#' and empty lines will be ignored.
+ Lines starting with '-' will be interpreted as a key-value pair separated
+ by an '='. where 'key' is a dot-separated option name and 'value' is a
+ JSON-parsable value for it. These config options will be applied while
+ processing the next URL.
+ Lines starting with '-G' are the same as above, except these options will
+ be valid for all following URLs, i.e. they are Global.
+ Everything else will be used as potential URL.
+
+ Example input file:
+
+ # settings global options
+ -G base-directory = "/tmp/"
+ -G skip = false
+
+ # setting local options for the next URL
+ -filename="spaces_are_optional.jpg"
+ -skip = true
+
+ https://example.org/
+
+ # next URL uses default filename and 'skip' is false.
+ https://example.com/index.htm
+ """
+ gconf = []
+ lconf = []
+
+ for line in file:
+ line = line.strip()
+
+ if not line or line[0] == "#":
+ # empty line or comment
+ continue
+
+ elif line[0] == "-":
+ # config spec
+ if len(line) >= 2 and line[1] == "G":
+ conf = gconf
+ line = line[2:]
+ else:
+ conf = lconf
+ line = line[1:]
+
+ key, sep, value = line.partition("=")
+ if not sep:
+ log.warning("input file: invalid <key>=<value> pair: %s", line)
+ continue
+
+ try:
+ value = json.loads(value.strip())
+ except ValueError as exc:
+ log.warning("input file: unable to parse '%s': %s", value, exc)
+ continue
+
+ conf.append((key.strip().split("."), value))
+
+ else:
+ # url
+ if gconf or lconf:
+ yield util.ExtendedUrl(line, gconf, lconf)
+ gconf = []
+ lconf = []
+ else:
+ yield line
+
+
+def main():
+ try:
+ if sys.stdout.encoding.lower() != "utf-8":
+ output.replace_std_streams()
+
+ parser = option.build_parser()
+ args = parser.parse_args()
+ log = output.initialize_logging(args.loglevel)
+
+ # configuration
+ if args.load_config:
+ config.load()
+ if args.cfgfiles:
+ config.load(args.cfgfiles, strict=True)
+ if args.yamlfiles:
+ config.load(args.yamlfiles, strict=True, fmt="yaml")
+ if args.postprocessors:
+ config.set(("postprocessors", ), args.postprocessors)
+ for key, value in args.options:
+ config.set(key, value)
+
+ # stream logging handler
+ output.configure_logging_handler(
+ "log", logging.getLogger().handlers[0])
+
+ # file logging handler
+ handler = output.setup_logging_handler(
+ "logfile", lvl=args.loglevel)
+ if handler:
+ logging.getLogger().addHandler(handler)
+
+ # loglevels
+ if args.loglevel >= logging.ERROR:
+ config.set(("output", "mode"), "null")
+ elif args.loglevel <= logging.DEBUG:
+ import platform
+ import subprocess
+ import os.path
+ import requests
+
+ head = ""
+ try:
+ out, err = subprocess.Popen(
+ ("git", "rev-parse", "--short", "HEAD"),
+ stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE,
+ cwd=os.path.dirname(os.path.abspath(__file__)),
+ ).communicate()
+ if out and not err:
+ head = " - Git HEAD: " + out.decode().rstrip()
+ except (OSError, subprocess.SubprocessError):
+ pass
+
+ log.debug("Version %s%s", __version__, head)
+ log.debug("Python %s - %s",
+ platform.python_version(), platform.platform())
+ try:
+ log.debug("requests %s - urllib3 %s",
+ requests.__version__,
+ requests.packages.urllib3.__version__)
+ except AttributeError:
+ pass
+
+ if args.list_modules:
+ for module_name in extractor.modules:
+ print(module_name)
+ elif args.list_extractors:
+ for extr in extractor.extractors():
+ if not extr.__doc__:
+ continue
+ print(extr.__name__)
+ print(extr.__doc__)
+ print("Category:", extr.category,
+ "- Subcategory:", extr.subcategory)
+ test = next(extr._get_tests(), None)
+ if test:
+ print("Example :", test[0])
+ print()
+ elif args.clear_cache:
+ from . import cache
+ log = logging.getLogger("cache")
+ cnt = cache.clear()
+
+ if cnt is None:
+ log.error("Database file not available")
+ else:
+ log.info(
+ "Deleted %d %s from '%s'",
+ cnt, "entry" if cnt == 1 else "entries", cache._path(),
+ )
+ else:
+ if not args.urls and not args.inputfile:
+ parser.error(
+ "The following arguments are required: URL\n"
+ "Use 'gallery-dl --help' to get a list of all options.")
+
+ if args.list_urls:
+ jobtype = job.UrlJob
+ jobtype.maxdepth = args.list_urls
+ else:
+ jobtype = args.jobtype or job.DownloadJob
+
+ urls = args.urls
+ if args.inputfile:
+ try:
+ if args.inputfile == "-":
+ file = sys.stdin
+ else:
+ file = open(args.inputfile, encoding="utf-8")
+ urls += parse_inputfile(file, log)
+ file.close()
+ except OSError as exc:
+ log.warning("input file: %s", exc)
+
+ # unsupported file logging handler
+ handler = output.setup_logging_handler(
+ "unsupportedfile", fmt="{message}")
+ if handler:
+ ulog = logging.getLogger("unsupported")
+ ulog.addHandler(handler)
+ ulog.propagate = False
+ job.Job.ulog = ulog
+
+ pformat = config.get(("output", "progress"), True)
+ if pformat and len(urls) > 1 and args.loglevel < logging.ERROR:
+ urls = progress(urls, pformat)
+
+ for url in urls:
+ try:
+ log.debug("Starting %s for '%s'", jobtype.__name__, url)
+ if isinstance(url, util.ExtendedUrl):
+ for key, value in url.gconfig:
+ config.set(key, value)
+ with config.apply(url.lconfig):
+ jobtype(url.value).run()
+ else:
+ jobtype(url).run()
+ except exception.NoExtractorError:
+ log.error("No suitable extractor found for '%s'", url)
+
+ except KeyboardInterrupt:
+ print("\nKeyboardInterrupt", file=sys.stderr)
+ except BrokenPipeError:
+ pass
+ except IOError as exc:
+ import errno
+ if exc.errno != errno.EPIPE:
+ raise