mirror of
https://github.com/mikf/gallery-dl.git
synced 2024-11-22 10:42:34 +01:00
[sankaku] rewrite/improve (fixes #44)
- add wait-time between HTTP requests similar to exhentai - add 'wait-min' and 'wait-max' options - increase retry-count for HTTP requests to 10 - implement user authentication (non-authenticated users can only view images up to page 25) - implement 'skip()' functionality (only works up to page 50) - implement image-retrieval for pages >= 51 - fix issue with multiple tags
This commit is contained in:
parent
9aecc67841
commit
6af921a952
@ -238,7 +238,8 @@ Description The username and password to use when attempting to log in to
|
|||||||
|
|
||||||
Specifying username and password is
|
Specifying username and password is
|
||||||
required for the ``pixiv``, ``nijie`` and ``seiga`` modules and
|
required for the ``pixiv``, ``nijie`` and ``seiga`` modules and
|
||||||
optional (but strongly recommended) for ``batoto`` and ``exhentai``.
|
optional (but strongly recommended) for ``batoto``, ``exhentai``
|
||||||
|
and ``sankaku``.
|
||||||
|
|
||||||
These values can also be set via the ``-u/--username`` and
|
These values can also be set via the ``-u/--username`` and
|
||||||
``-p/--password`` command-line options or by using a |.netrc|_ file.
|
``-p/--password`` command-line options or by using a |.netrc|_ file.
|
||||||
@ -522,6 +523,20 @@ Description The ``refresh_token`` value you get from linking your Reddit account
|
|||||||
=========== =====
|
=========== =====
|
||||||
|
|
||||||
|
|
||||||
|
extractor.sankaku.wait-min & .wait-max
|
||||||
|
--------------------------------------
|
||||||
|
=========== =====
|
||||||
|
Type ``float``
|
||||||
|
Default ``2.0`` and ``4.0``
|
||||||
|
Description Minimum and maximum wait time in seconds between each image
|
||||||
|
|
||||||
|
Sankaku Channel responds with ``429 Too Many Requests`` if it
|
||||||
|
receives too many HTTP requests in a certain amount of time.
|
||||||
|
Waiting a few seconds between each request tries to prevent that.
|
||||||
|
=========== =====
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
API Tokens & IDs
|
API Tokens & IDs
|
||||||
================
|
================
|
||||||
|
|
||||||
|
@ -49,6 +49,11 @@
|
|||||||
"username": null,
|
"username": null,
|
||||||
"password": null
|
"password": null
|
||||||
},
|
},
|
||||||
|
"sankaku":
|
||||||
|
{
|
||||||
|
"username": null,
|
||||||
|
"password": null
|
||||||
|
},
|
||||||
"seiga":
|
"seiga":
|
||||||
{
|
{
|
||||||
"username": null,
|
"username": null,
|
||||||
|
@ -63,7 +63,7 @@ RebeccaBlackTech https://rbt.asia/ Threads
|
|||||||
Reddit https://reddit.com/ Submissions, Subreddits Optional (OAuth)
|
Reddit https://reddit.com/ Submissions, Subreddits Optional (OAuth)
|
||||||
Rule 34 https://rule34.xxx/ Posts, Tag-Searches
|
Rule 34 https://rule34.xxx/ Posts, Tag-Searches
|
||||||
Safebooru https://safebooru.org/ Posts, Tag-Searches
|
Safebooru https://safebooru.org/ Posts, Tag-Searches
|
||||||
Sankaku Channel https://chan.sankakucomplex.com/ Tag-Searches
|
Sankaku Channel https://chan.sankakucomplex.com/ Tag-Searches Optional
|
||||||
Sea Otter Scans https://reader.seaotterscans.com/ Chapters, Manga
|
Sea Otter Scans https://reader.seaotterscans.com/ Chapters, Manga
|
||||||
Sen Manga http://raw.senmanga.com/ Chapters
|
Sen Manga http://raw.senmanga.com/ Chapters
|
||||||
Sense-Scans http://sensescans.com/ Chapters, Manga
|
Sense-Scans http://sensescans.com/ Chapters, Manga
|
||||||
|
@ -8,11 +8,14 @@
|
|||||||
|
|
||||||
"""Extract images from https://chan.sankakucomplex.com/"""
|
"""Extract images from https://chan.sankakucomplex.com/"""
|
||||||
|
|
||||||
from .common import AsynchronousExtractor, Message
|
from .common import Extractor, Message
|
||||||
from .. import text
|
from .. import text, util, exception
|
||||||
|
from ..cache import cache
|
||||||
|
import time
|
||||||
|
import random
|
||||||
|
|
||||||
|
|
||||||
class SankakuTagExtractor(AsynchronousExtractor):
|
class SankakuTagExtractor(Extractor):
|
||||||
"""Extractor for images from chan.sankakucomplex.com by search-tags"""
|
"""Extractor for images from chan.sankakucomplex.com by search-tags"""
|
||||||
category = "sankaku"
|
category = "sankaku"
|
||||||
subcategory = "tag"
|
subcategory = "tag"
|
||||||
@ -24,16 +27,30 @@ class SankakuTagExtractor(AsynchronousExtractor):
|
|||||||
"pattern": (r"https://cs\.sankakucomplex\.com/data/[^/]{2}/[^/]{2}"
|
"pattern": (r"https://cs\.sankakucomplex\.com/data/[^/]{2}/[^/]{2}"
|
||||||
r"/[^/]{32}\.\w+\?e=\d+&m=[^&#]+"),
|
r"/[^/]{32}\.\w+\?e=\d+&m=[^&#]+"),
|
||||||
})]
|
})]
|
||||||
url = "https://chan.sankakucomplex.com/"
|
root = "https://chan.sankakucomplex.com"
|
||||||
|
cookienames = ("login", "pass_hash")
|
||||||
|
cookiedomain = "chan.sankakucomplex.com"
|
||||||
|
|
||||||
def __init__(self, match):
|
def __init__(self, match):
|
||||||
AsynchronousExtractor.__init__(self)
|
Extractor.__init__(self)
|
||||||
self.tags = text.unquote(match.group(1))
|
self.logged_in = True
|
||||||
|
self.pagestart = 1
|
||||||
|
self.tags = text.unquote(match.group(1).replace("+", " "))
|
||||||
|
self.wait_min = self.config("wait-min", 2)
|
||||||
|
self.wait_max = self.config("wait-max", 4)
|
||||||
|
if self.wait_max < self.wait_min:
|
||||||
|
self.wait_max = self.wait_min
|
||||||
self.session.headers["User-Agent"] = (
|
self.session.headers["User-Agent"] = (
|
||||||
"Mozilla/5.0 Gecko/20100101 Firefox/40.0"
|
"Mozilla/5.0 Gecko/20100101 Firefox/40.0"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
def skip(self, num):
|
||||||
|
pages = min(num // 20, 49)
|
||||||
|
self.pagestart += pages
|
||||||
|
return pages * 20
|
||||||
|
|
||||||
def items(self):
|
def items(self):
|
||||||
|
self.login()
|
||||||
data = self.get_job_metadata()
|
data = self.get_job_metadata()
|
||||||
yield Message.Version, 1
|
yield Message.Version, 1
|
||||||
yield Message.Directory, data
|
yield Message.Directory, data
|
||||||
@ -48,36 +65,67 @@ class SankakuTagExtractor(AsynchronousExtractor):
|
|||||||
def get_images(self):
|
def get_images(self):
|
||||||
params = {
|
params = {
|
||||||
"tags": self.tags,
|
"tags": self.tags,
|
||||||
"page": 1,
|
"page": self.pagestart,
|
||||||
}
|
}
|
||||||
while True:
|
while self.logged_in or params["page"] <= 25:
|
||||||
count = 0
|
image = None
|
||||||
page = self.request(self.url, params=params).text
|
page = self.request(self.root, params=params, retries=10).text
|
||||||
pos = text.extract(page, '<div id=more-popular-posts-link>', '')[1]
|
pos = text.extract(page, '<div id=more-popular-posts-link>', '')[1]
|
||||||
while True:
|
for image_id in text.extract_iter(
|
||||||
image_id, pos = text.extract(
|
page, '<span class="thumb blacklisted" id=p', '>', pos):
|
||||||
page, '<span class="thumb blacklisted" id=p', '>', pos
|
self.wait()
|
||||||
)
|
|
||||||
if not image_id:
|
|
||||||
break
|
|
||||||
image = self.get_image_metadata(image_id)
|
image = self.get_image_metadata(image_id)
|
||||||
count += 1
|
|
||||||
yield image
|
yield image
|
||||||
if count < 20:
|
if not image:
|
||||||
return
|
return
|
||||||
params["page"] += 1
|
params["page"] += 1
|
||||||
|
params["next"] = image["id"] - 1
|
||||||
|
|
||||||
def get_image_metadata(self, image_id):
|
def get_image_metadata(self, image_id):
|
||||||
url = "https://chan.sankakucomplex.com/post/show/" + image_id
|
url = "https://chan.sankakucomplex.com/post/show/" + image_id
|
||||||
page = self.request(url).text
|
page = self.request(url, retries=10).text
|
||||||
image_url, pos = text.extract(page, '<li>Original: <a href="', '"')
|
image_url, pos = text.extract(page, '<li>Original: <a href="', '"')
|
||||||
width , pos = text.extract(page, '>', 'x', pos)
|
width , pos = text.extract(page, '>', 'x', pos)
|
||||||
height , pos = text.extract(page, '', ' ', pos)
|
height , pos = text.extract(page, '', ' ', pos)
|
||||||
data = text.nameext_from_url(image_url, {
|
data = text.nameext_from_url(image_url, {
|
||||||
"id": image_id,
|
"id": util.safe_int(image_id),
|
||||||
"file_url": "https:" + text.unescape(image_url),
|
"file_url": "https:" + text.unescape(image_url),
|
||||||
"width": width,
|
"width": util.safe_int(width),
|
||||||
"height": height,
|
"height": util.safe_int(height),
|
||||||
})
|
})
|
||||||
data["md5"] = data["name"]
|
data["md5"] = data["name"]
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
def wait(self):
|
||||||
|
"""Wait for a randomly chosen amount of seconds"""
|
||||||
|
time.sleep(random.uniform(self.wait_min, self.wait_max))
|
||||||
|
|
||||||
|
def login(self):
|
||||||
|
"""Login and set necessary cookies"""
|
||||||
|
if self._check_cookies(self.cookienames):
|
||||||
|
return
|
||||||
|
username, password = self._get_auth_info()
|
||||||
|
if username:
|
||||||
|
cookies = self._login_impl(username, password)
|
||||||
|
for key, value in cookies.items():
|
||||||
|
self.session.cookies.set(
|
||||||
|
key, value, domain=self.cookiedomain)
|
||||||
|
else:
|
||||||
|
self.logged_in = False
|
||||||
|
|
||||||
|
@cache(maxage=90*24*60*60, keyarg=1)
|
||||||
|
def _login_impl(self, username, password):
|
||||||
|
"""Actual login implementation"""
|
||||||
|
self.log.info("Logging in as %s", username)
|
||||||
|
params = {
|
||||||
|
"url": "",
|
||||||
|
"user[name]": username,
|
||||||
|
"user[password]": password,
|
||||||
|
"commit": "Login",
|
||||||
|
}
|
||||||
|
response = self.request(self.root + "/user/authenticate",
|
||||||
|
method="POST", params=params)
|
||||||
|
if not response.history or response.url != self.root + "/user/home":
|
||||||
|
raise exception.AuthenticationError()
|
||||||
|
response = response.history[0]
|
||||||
|
return {c: response.cookies[c] for c in self.cookienames}
|
||||||
|
@ -82,6 +82,7 @@ AUTH_MAP = {
|
|||||||
"nijie" : "Required",
|
"nijie" : "Required",
|
||||||
"pixiv" : "Required",
|
"pixiv" : "Required",
|
||||||
"reddit" : "Optional (OAuth)",
|
"reddit" : "Optional (OAuth)",
|
||||||
|
"sankaku" : "Optional",
|
||||||
"seiga" : "Required",
|
"seiga" : "Required",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -90,7 +90,7 @@ class TestCookiedict(unittest.TestCase):
|
|||||||
self.assertEqual(sorted(cookies.values()), sorted(self.cdict.values()))
|
self.assertEqual(sorted(cookies.values()), sorted(self.cdict.values()))
|
||||||
|
|
||||||
def test_domain(self):
|
def test_domain(self):
|
||||||
for category in ["batoto", "exhentai", "nijie", "seiga"]:
|
for category in ["batoto", "exhentai", "nijie", "sankaku", "seiga"]:
|
||||||
extr = _get_extractor(category)
|
extr = _get_extractor(category)
|
||||||
cookies = extr.session.cookies
|
cookies = extr.session.cookies
|
||||||
for key in self.cdict.keys():
|
for key in self.cdict.keys():
|
||||||
@ -109,6 +109,7 @@ class TestCookieLogin(unittest.TestCase):
|
|||||||
"batoto": ("member_id", "pass_hash"),
|
"batoto": ("member_id", "pass_hash"),
|
||||||
"exhentai": ("ipb_member_id", "ipb_pass_hash"),
|
"exhentai": ("ipb_member_id", "ipb_pass_hash"),
|
||||||
"nijie": ("nemail", "nlogin"),
|
"nijie": ("nemail", "nlogin"),
|
||||||
|
"sankaku": ("login", "pass_hash"),
|
||||||
"seiga": ("user_session",),
|
"seiga": ("user_session",),
|
||||||
}
|
}
|
||||||
for category, cookienames in extr_cookies.items():
|
for category, cookienames in extr_cookies.items():
|
||||||
|
Loading…
Reference in New Issue
Block a user