2015-04-05 16:23:20 +02:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
2017-02-20 22:02:49 +01:00
|
|
|
# Copyright 2015-2017 Mike Fährmann
|
2015-04-05 16:23:20 +02:00
|
|
|
#
|
|
|
|
# This program is free software; you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU General Public License version 2 as
|
|
|
|
# published by the Free Software Foundation.
|
|
|
|
|
2017-04-12 18:43:41 +02:00
|
|
|
import sys
|
2017-12-04 17:06:17 +01:00
|
|
|
import time
|
2015-12-12 01:16:02 +01:00
|
|
|
import json
|
|
|
|
import hashlib
|
2017-03-28 13:12:44 +02:00
|
|
|
from . import extractor, downloader, config, util, output, exception
|
2015-11-24 19:47:51 +01:00
|
|
|
from .extractor.message import Message
|
2015-04-05 16:23:20 +02:00
|
|
|
|
2017-01-30 19:40:15 +01:00
|
|
|
|
2015-12-12 00:11:05 +01:00
|
|
|
class Job():
|
|
|
|
"""Base class for Job-types"""
|
2017-05-27 16:16:57 +02:00
|
|
|
ufile = None
|
2015-04-08 01:51:48 +02:00
|
|
|
|
2017-09-30 18:52:23 +02:00
|
|
|
def __init__(self, url, parent=None):
|
2017-02-25 23:53:31 +01:00
|
|
|
self.url = url
|
2015-11-21 00:30:31 +01:00
|
|
|
self.extractor = extractor.find(url)
|
2015-06-28 12:45:52 +02:00
|
|
|
if self.extractor is None:
|
2016-07-14 14:25:56 +02:00
|
|
|
raise exception.NoExtractorError(url)
|
2017-06-02 09:10:58 +02:00
|
|
|
self.extractor.log.debug(
|
|
|
|
"Using %s for %s", self.extractor.__class__.__name__, url)
|
2015-12-12 00:11:05 +01:00
|
|
|
|
2017-09-06 17:08:50 +02:00
|
|
|
# url predicates
|
|
|
|
predicates = [util.UniquePredicate()]
|
2017-09-08 17:52:00 +02:00
|
|
|
image = config.get(("_", "image"), {})
|
2017-09-09 14:51:31 +02:00
|
|
|
if "filter" in image:
|
|
|
|
predicates.append(util.FilterPredicate(image["filter"]))
|
2017-09-08 17:52:00 +02:00
|
|
|
if "range" in image:
|
|
|
|
pred = util.RangePredicate(image["range"])
|
2017-03-03 17:26:50 +01:00
|
|
|
if pred.lower > 1:
|
|
|
|
pred.index += self.extractor.skip(pred.lower - 1)
|
2017-09-06 17:08:50 +02:00
|
|
|
predicates.append(pred)
|
|
|
|
self.pred_url = util.build_predicate(predicates)
|
|
|
|
|
|
|
|
# queue predicates
|
|
|
|
predicates = []
|
2017-09-08 17:52:00 +02:00
|
|
|
chapter = config.get(("_", "chapter"), {})
|
|
|
|
if "filter" in chapter:
|
|
|
|
predicates.append(util.FilterPredicate(chapter["filter"]))
|
2017-09-09 14:51:31 +02:00
|
|
|
if "range" in chapter:
|
|
|
|
predicates.append(util.RangePredicate(chapter["range"]))
|
2017-09-06 17:08:50 +02:00
|
|
|
self.pred_queue = util.build_predicate(predicates)
|
2016-09-24 10:45:11 +02:00
|
|
|
|
2017-09-30 18:52:23 +02:00
|
|
|
# category transfer
|
|
|
|
if parent and parent.extractor.categorytransfer:
|
|
|
|
self.extractor.category = parent.extractor.category
|
|
|
|
self.extractor.subcategory = parent.extractor.subcategory
|
|
|
|
|
2017-02-23 21:51:29 +01:00
|
|
|
def run(self):
|
|
|
|
"""Execute or run the job"""
|
|
|
|
try:
|
2017-03-11 01:47:57 +01:00
|
|
|
log = self.extractor.log
|
2017-02-23 21:51:29 +01:00
|
|
|
for msg in self.extractor:
|
2017-02-26 02:06:56 +01:00
|
|
|
self.dispatch(msg)
|
2017-02-25 23:53:31 +01:00
|
|
|
except exception.AuthenticationError:
|
2017-03-11 01:47:57 +01:00
|
|
|
log.error("Authentication failed. Please provide a valid "
|
|
|
|
"username/password pair.")
|
2017-02-25 23:53:31 +01:00
|
|
|
except exception.AuthorizationError:
|
2017-03-11 01:47:57 +01:00
|
|
|
log.error("You do not have permission to access the resource "
|
|
|
|
"at '%s'", self.url)
|
2017-08-05 16:11:46 +02:00
|
|
|
except exception.NotFoundError as exc:
|
|
|
|
res = str(exc) or "resource (gallery/image/user)"
|
2017-03-11 01:47:57 +01:00
|
|
|
log.error("The %s at '%s' does not exist", res, self.url)
|
2017-08-05 16:11:46 +02:00
|
|
|
except exception.HttpError as exc:
|
2017-09-25 13:01:10 +02:00
|
|
|
log.error("HTTP request failed: %s", exc)
|
2017-08-11 21:48:37 +02:00
|
|
|
except exception.FormatError as exc:
|
|
|
|
err, obj = exc.args
|
2017-09-08 17:52:00 +02:00
|
|
|
log.error("Applying %s format string failed: %s: %s",
|
2017-08-11 21:48:37 +02:00
|
|
|
obj, err.__class__.__name__, err)
|
2017-09-08 17:52:00 +02:00
|
|
|
except exception.FilterError as exc:
|
|
|
|
err = exc.args[0]
|
|
|
|
log.error("Evaluating filter expression failed: %s: %s",
|
|
|
|
err.__class__.__name__, err)
|
2017-02-23 21:51:29 +01:00
|
|
|
except exception.StopExtraction:
|
|
|
|
pass
|
2017-08-10 16:29:05 +02:00
|
|
|
except OSError as exc:
|
|
|
|
log.error("Unable to download data: %s", exc)
|
2017-04-18 11:38:48 +02:00
|
|
|
except Exception as exc:
|
2017-08-10 16:29:05 +02:00
|
|
|
log.error(("An unexpected error occurred: %s - %s. "
|
|
|
|
"Please run gallery-dl again with the --verbose flag, "
|
|
|
|
"copy its output and report this issue on "
|
|
|
|
"https://github.com/mikf/gallery-dl/issues ."),
|
|
|
|
exc.__class__.__name__, exc)
|
|
|
|
log.debug("Traceback", exc_info=True)
|
2017-04-18 11:38:48 +02:00
|
|
|
|
2017-02-26 02:06:56 +01:00
|
|
|
def dispatch(self, msg):
|
|
|
|
"""Call the appropriate message handler"""
|
2017-03-17 09:39:46 +01:00
|
|
|
if msg[0] == Message.Url:
|
2017-09-06 17:08:50 +02:00
|
|
|
_, url, kwds = msg
|
|
|
|
if self.pred_url(url, kwds):
|
|
|
|
self.update_kwdict(kwds)
|
|
|
|
self.handle_url(url, kwds)
|
2017-02-26 02:06:56 +01:00
|
|
|
|
|
|
|
elif msg[0] == Message.Directory:
|
|
|
|
self.update_kwdict(msg[1])
|
|
|
|
self.handle_directory(msg[1])
|
|
|
|
|
2017-03-17 09:39:46 +01:00
|
|
|
elif msg[0] == Message.Queue:
|
2017-09-12 16:19:00 +02:00
|
|
|
_, url, kwds = msg
|
|
|
|
if self.pred_queue(url, kwds):
|
|
|
|
self.handle_queue(url, kwds)
|
2017-02-26 02:06:56 +01:00
|
|
|
|
|
|
|
elif msg[0] == Message.Version:
|
|
|
|
if msg[1] != 1:
|
|
|
|
raise "unsupported message-version ({}, {})".format(
|
|
|
|
self.extractor.category, msg[1]
|
|
|
|
)
|
|
|
|
# TODO: support for multiple message versions
|
|
|
|
|
2017-04-12 18:43:41 +02:00
|
|
|
def handle_url(self, url, keywords):
|
2016-09-24 10:45:11 +02:00
|
|
|
"""Handle Message.Url"""
|
|
|
|
|
|
|
|
def handle_directory(self, keywords):
|
|
|
|
"""Handle Message.Directory"""
|
|
|
|
|
2017-09-12 16:19:00 +02:00
|
|
|
def handle_queue(self, url, keywords):
|
2016-09-24 10:45:11 +02:00
|
|
|
"""Handle Message.Queue"""
|
|
|
|
|
|
|
|
def update_kwdict(self, kwdict):
|
|
|
|
"""Add 'category' and 'subcategory' keywords"""
|
|
|
|
kwdict["category"] = self.extractor.category
|
|
|
|
kwdict["subcategory"] = self.extractor.subcategory
|
2015-12-12 00:11:05 +01:00
|
|
|
|
2017-05-27 16:16:57 +02:00
|
|
|
def _write_unsupported(self, url):
|
|
|
|
if self.ufile:
|
|
|
|
print(url, file=self.ufile, flush=True)
|
|
|
|
|
2017-01-30 19:40:15 +01:00
|
|
|
|
2015-12-12 00:11:05 +01:00
|
|
|
class DownloadJob(Job):
|
|
|
|
"""Download images into appropriate directory/filename locations"""
|
|
|
|
|
2017-09-30 18:52:23 +02:00
|
|
|
def __init__(self, url, parent=None):
|
2017-10-06 15:38:35 +02:00
|
|
|
Job.__init__(self, url, parent)
|
2017-03-28 13:12:44 +02:00
|
|
|
self.pathfmt = util.PathFormat(self.extractor)
|
2017-12-04 17:06:17 +01:00
|
|
|
self.sleep = self.extractor.config("sleep")
|
2015-04-08 01:51:48 +02:00
|
|
|
self.downloaders = {}
|
2016-09-30 12:32:48 +02:00
|
|
|
self.out = output.select()
|
2015-04-08 01:51:48 +02:00
|
|
|
|
2016-09-24 10:45:11 +02:00
|
|
|
def handle_url(self, url, keywords):
|
|
|
|
"""Download the resource specified in 'url'"""
|
2016-09-30 12:32:48 +02:00
|
|
|
self.pathfmt.set_keywords(keywords)
|
|
|
|
if self.pathfmt.exists():
|
|
|
|
self.out.skip(self.pathfmt.path)
|
2015-04-08 01:51:48 +02:00
|
|
|
return
|
2017-12-04 17:06:17 +01:00
|
|
|
if self.sleep:
|
|
|
|
time.sleep(self.sleep)
|
2015-11-12 02:35:30 +01:00
|
|
|
dlinstance = self.get_downloader(url)
|
2016-09-30 12:32:48 +02:00
|
|
|
dlinstance.download(url, self.pathfmt)
|
2015-04-08 01:51:48 +02:00
|
|
|
|
2016-09-24 10:45:11 +02:00
|
|
|
def handle_directory(self, keywords):
|
2015-04-08 01:51:48 +02:00
|
|
|
"""Set and create the target directory for downloads"""
|
2016-09-30 12:32:48 +02:00
|
|
|
self.pathfmt.set_directory(keywords)
|
2015-04-08 01:51:48 +02:00
|
|
|
|
2017-09-12 16:19:00 +02:00
|
|
|
def handle_queue(self, url, keywords):
|
2016-09-24 10:45:11 +02:00
|
|
|
try:
|
2017-09-30 18:52:23 +02:00
|
|
|
DownloadJob(url, self).run()
|
2017-05-24 15:15:06 +02:00
|
|
|
except exception.NoExtractorError:
|
2017-05-27 16:16:57 +02:00
|
|
|
self._write_unsupported(url)
|
2016-09-24 10:45:11 +02:00
|
|
|
|
2015-04-08 01:51:48 +02:00
|
|
|
def get_downloader(self, url):
|
|
|
|
"""Return, and possibly construct, a downloader suitable for 'url'"""
|
|
|
|
pos = url.find(":")
|
|
|
|
scheme = url[:pos] if pos != -1 else "http"
|
|
|
|
if scheme == "https":
|
|
|
|
scheme = "http"
|
2015-11-12 02:35:30 +01:00
|
|
|
instance = self.downloaders.get(scheme)
|
|
|
|
if instance is None:
|
|
|
|
klass = downloader.find(scheme)
|
2017-06-30 19:38:14 +02:00
|
|
|
instance = klass(self.extractor.session, self.out)
|
2015-11-12 02:35:30 +01:00
|
|
|
self.downloaders[scheme] = instance
|
|
|
|
return instance
|
|
|
|
|
2015-11-13 01:02:49 +01:00
|
|
|
|
2015-12-12 00:11:05 +01:00
|
|
|
class KeywordJob(Job):
|
|
|
|
"""Print available keywords"""
|
2015-11-13 01:02:49 +01:00
|
|
|
|
2017-05-17 14:31:14 +02:00
|
|
|
def handle_url(self, url, keywords):
|
2017-09-30 18:52:23 +02:00
|
|
|
print("\nKeywords for filenames and --filter:")
|
|
|
|
print("------------------------------------")
|
2017-05-17 14:31:14 +02:00
|
|
|
self.print_keywords(keywords)
|
|
|
|
raise exception.StopExtraction()
|
|
|
|
|
|
|
|
def handle_directory(self, keywords):
|
|
|
|
print("Keywords for directory names:")
|
|
|
|
print("-----------------------------")
|
|
|
|
self.print_keywords(keywords)
|
2015-11-13 01:02:49 +01:00
|
|
|
|
2017-09-12 16:19:00 +02:00
|
|
|
def handle_queue(self, url, keywords):
|
2017-09-26 20:50:49 +02:00
|
|
|
if not keywords:
|
|
|
|
self.extractor.log.info(
|
2017-10-09 23:20:17 +02:00
|
|
|
"This extractor delegates work to other extractors "
|
2017-09-26 20:50:49 +02:00
|
|
|
"and does not provide any keywords on its own. Try "
|
|
|
|
"'gallery-dl -K \"%s\"' instead.", url)
|
|
|
|
else:
|
|
|
|
print("Keywords for --chapter-filter:")
|
|
|
|
print("------------------------------")
|
|
|
|
self.print_keywords(keywords)
|
|
|
|
if self.extractor.categorytransfer:
|
|
|
|
print()
|
2017-09-30 18:52:23 +02:00
|
|
|
KeywordJob(url, self).run()
|
2017-08-10 17:36:21 +02:00
|
|
|
raise exception.StopExtraction()
|
|
|
|
|
2015-11-13 01:02:49 +01:00
|
|
|
@staticmethod
|
2017-05-17 14:31:14 +02:00
|
|
|
def print_keywords(keywords, prefix=""):
|
2015-12-12 01:16:02 +01:00
|
|
|
"""Print key-value pairs with formatting"""
|
2017-05-17 14:31:14 +02:00
|
|
|
suffix = "]" if prefix else ""
|
2015-11-13 01:02:49 +01:00
|
|
|
for key, value in sorted(keywords.items()):
|
2017-05-17 14:31:14 +02:00
|
|
|
key = prefix + key + suffix
|
2017-05-15 18:30:47 +02:00
|
|
|
|
|
|
|
if isinstance(value, dict):
|
2017-05-17 14:31:14 +02:00
|
|
|
KeywordJob.print_keywords(value, key + "[")
|
2017-05-15 18:30:47 +02:00
|
|
|
|
|
|
|
elif isinstance(value, list):
|
|
|
|
if value and isinstance(value[0], dict):
|
2017-05-17 14:31:14 +02:00
|
|
|
KeywordJob.print_keywords(value[0], key + "[][")
|
2017-05-15 18:30:47 +02:00
|
|
|
else:
|
2017-05-17 14:31:14 +02:00
|
|
|
print(key, "[]", sep="")
|
2017-05-15 18:30:47 +02:00
|
|
|
for val in value:
|
2017-05-17 14:31:14 +02:00
|
|
|
print(" -", val)
|
2017-05-15 18:30:47 +02:00
|
|
|
|
|
|
|
else:
|
|
|
|
# string or number
|
2017-05-17 14:31:14 +02:00
|
|
|
print(key, "\n ", value, sep="")
|
2015-12-10 02:14:28 +01:00
|
|
|
|
|
|
|
|
2015-12-12 00:11:05 +01:00
|
|
|
class UrlJob(Job):
|
|
|
|
"""Print download urls"""
|
2017-02-17 22:18:16 +01:00
|
|
|
maxdepth = -1
|
|
|
|
|
2017-09-30 18:52:23 +02:00
|
|
|
def __init__(self, url, parent=None, depth=1):
|
|
|
|
Job.__init__(self, url, parent)
|
2017-02-17 22:18:16 +01:00
|
|
|
self.depth = depth
|
|
|
|
if depth == self.maxdepth:
|
2017-09-12 16:19:00 +02:00
|
|
|
self.handle_queue = self.handle_url
|
2015-12-10 02:14:28 +01:00
|
|
|
|
2017-05-23 11:48:00 +02:00
|
|
|
@staticmethod
|
|
|
|
def handle_url(url, _):
|
2016-09-24 10:45:11 +02:00
|
|
|
print(url)
|
2016-08-11 13:20:21 +02:00
|
|
|
|
2017-09-12 16:19:00 +02:00
|
|
|
def handle_queue(self, url, _):
|
2016-09-24 10:45:11 +02:00
|
|
|
try:
|
2017-09-30 18:52:23 +02:00
|
|
|
UrlJob(url, self, self.depth + 1).run()
|
2016-09-24 10:45:11 +02:00
|
|
|
except exception.NoExtractorError:
|
2017-05-27 16:16:57 +02:00
|
|
|
self._write_unsupported(url)
|
2017-05-23 11:48:00 +02:00
|
|
|
|
2015-12-12 01:16:02 +01:00
|
|
|
|
2017-02-26 02:06:56 +01:00
|
|
|
class TestJob(DownloadJob):
|
|
|
|
"""Generate test-results for extractor runs"""
|
2015-12-12 01:16:02 +01:00
|
|
|
|
2015-12-21 22:49:04 +01:00
|
|
|
class HashIO():
|
2016-09-24 10:45:11 +02:00
|
|
|
"""Minimal file-like interface"""
|
2015-12-21 22:49:04 +01:00
|
|
|
|
|
|
|
def __init__(self, hashobj):
|
|
|
|
self.hashobj = hashobj
|
2016-09-30 12:32:48 +02:00
|
|
|
self.path = ""
|
2017-10-20 18:56:18 +02:00
|
|
|
self.size = 0
|
2016-09-30 12:32:48 +02:00
|
|
|
self.has_extension = True
|
|
|
|
|
|
|
|
def __enter__(self):
|
|
|
|
return self
|
|
|
|
|
|
|
|
def __exit__(self, *args):
|
|
|
|
pass
|
|
|
|
|
2017-10-25 12:55:36 +02:00
|
|
|
def open(self, mode):
|
2017-10-20 18:56:18 +02:00
|
|
|
self.size = 0
|
2016-09-30 12:32:48 +02:00
|
|
|
return self
|
2015-12-21 22:49:04 +01:00
|
|
|
|
|
|
|
def write(self, content):
|
2016-09-24 10:45:11 +02:00
|
|
|
"""Update SHA1 hash"""
|
2017-10-20 18:56:18 +02:00
|
|
|
self.size += len(content)
|
2015-12-21 22:49:04 +01:00
|
|
|
self.hashobj.update(content)
|
|
|
|
|
2017-10-20 18:56:18 +02:00
|
|
|
def tell(self):
|
|
|
|
return self.size
|
|
|
|
|
2017-10-25 12:55:36 +02:00
|
|
|
def part_size(self):
|
|
|
|
return 0
|
|
|
|
|
2017-10-06 15:38:35 +02:00
|
|
|
def __init__(self, url, parent=None, content=False):
|
|
|
|
DownloadJob.__init__(self, url, parent)
|
2017-01-30 19:40:15 +01:00
|
|
|
self.content = content
|
2017-08-25 22:01:14 +02:00
|
|
|
self.urllist = []
|
2017-01-30 19:40:15 +01:00
|
|
|
self.hash_url = hashlib.sha1()
|
2015-12-12 01:16:02 +01:00
|
|
|
self.hash_keyword = hashlib.sha1()
|
2015-12-21 22:49:04 +01:00
|
|
|
self.hash_content = hashlib.sha1()
|
|
|
|
if content:
|
|
|
|
self.fileobj = self.HashIO(self.hash_content)
|
2015-12-12 01:16:02 +01:00
|
|
|
|
2017-02-26 02:06:56 +01:00
|
|
|
def run(self):
|
2017-02-27 23:05:08 +01:00
|
|
|
for msg in self.extractor:
|
|
|
|
self.dispatch(msg)
|
2017-02-26 02:06:56 +01:00
|
|
|
|
2016-09-24 10:45:11 +02:00
|
|
|
def handle_url(self, url, keywords):
|
|
|
|
self.update_url(url)
|
|
|
|
self.update_keyword(keywords)
|
|
|
|
self.update_content(url)
|
2015-12-12 01:16:02 +01:00
|
|
|
|
2016-09-24 10:45:11 +02:00
|
|
|
def handle_directory(self, keywords):
|
|
|
|
self.update_keyword(keywords)
|
2015-12-12 01:16:02 +01:00
|
|
|
|
2017-09-12 16:19:00 +02:00
|
|
|
def handle_queue(self, url, keywords):
|
2015-12-12 01:16:02 +01:00
|
|
|
self.update_url(url)
|
2017-09-12 16:19:00 +02:00
|
|
|
self.update_keyword(keywords)
|
2015-12-12 01:16:02 +01:00
|
|
|
|
|
|
|
def update_url(self, url):
|
2016-09-24 10:45:11 +02:00
|
|
|
"""Update the URL hash"""
|
2017-08-25 22:01:14 +02:00
|
|
|
self.urllist.append(url)
|
2015-12-12 01:16:02 +01:00
|
|
|
self.hash_url.update(url.encode())
|
|
|
|
|
|
|
|
def update_keyword(self, kwdict):
|
2016-09-24 10:45:11 +02:00
|
|
|
"""Update the keyword hash"""
|
2015-12-12 01:16:02 +01:00
|
|
|
self.hash_keyword.update(
|
|
|
|
json.dumps(kwdict, sort_keys=True).encode()
|
|
|
|
)
|
2015-12-21 22:49:04 +01:00
|
|
|
|
|
|
|
def update_content(self, url):
|
2016-09-24 10:45:11 +02:00
|
|
|
"""Update the content hash"""
|
2015-12-21 22:49:04 +01:00
|
|
|
if self.content:
|
|
|
|
self.get_downloader(url).download(url, self.fileobj)
|
2017-04-12 18:43:41 +02:00
|
|
|
|
|
|
|
|
|
|
|
class DataJob(Job):
|
|
|
|
"""Collect extractor results and dump them"""
|
|
|
|
|
2017-10-06 15:38:35 +02:00
|
|
|
def __init__(self, url, parent=None, file=sys.stdout):
|
|
|
|
Job.__init__(self, url, parent)
|
2017-04-12 18:43:41 +02:00
|
|
|
self.file = file
|
|
|
|
self.data = []
|
|
|
|
self.ensure_ascii = config.get(("output", "ascii"), True)
|
|
|
|
|
|
|
|
def run(self):
|
|
|
|
# collect data
|
|
|
|
try:
|
|
|
|
for msg in self.extractor:
|
2017-11-18 17:35:57 +01:00
|
|
|
self.dispatch(msg)
|
2017-04-12 18:43:41 +02:00
|
|
|
except Exception as exc:
|
|
|
|
self.data.append((exc.__class__.__name__, str(exc)))
|
2017-11-18 17:35:57 +01:00
|
|
|
except BaseException:
|
|
|
|
pass
|
2017-04-12 18:43:41 +02:00
|
|
|
|
|
|
|
# dump to 'file'
|
|
|
|
json.dump(
|
|
|
|
self.data, self.file,
|
|
|
|
sort_keys=True, indent=2, ensure_ascii=self.ensure_ascii
|
|
|
|
)
|
|
|
|
self.file.write("\n")
|
2017-11-18 17:35:57 +01:00
|
|
|
|
|
|
|
def handle_url(self, url, keywords):
|
|
|
|
self.data.append((Message.Url, url, keywords.copy()))
|
|
|
|
|
|
|
|
def handle_directory(self, keywords):
|
|
|
|
self.data.append((Message.Directory, keywords.copy()))
|
|
|
|
|
|
|
|
def handle_queue(self, url, keywords):
|
|
|
|
self.data.append((Message.Queue, url, keywords.copy()))
|