1
0
mirror of https://github.com/mikf/gallery-dl.git synced 2024-11-24 03:32:33 +01:00
gallery-dl/gallery_dl/extractor/simplyhentai.py
Mike Fährmann bd08ee2859
remove most 'yield Message.Version' statements
only leave them in oauth.py as noop results
2021-08-16 03:10:48 +02:00

196 lines
7.5 KiB
Python

# -*- coding: utf-8 -*-
# Copyright 2018-2021 Mike Fährmann
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License version 2 as
# published by the Free Software Foundation.
"""Extract hentai-manga from https://www.simply-hentai.com/"""
from .common import GalleryExtractor, Extractor, Message
from .. import text, util, exception
class SimplyhentaiGalleryExtractor(GalleryExtractor):
"""Extractor for image galleries from simply-hentai.com"""
category = "simplyhentai"
archive_fmt = "{image_id}"
pattern = (r"(?:https?://)?(?!videos\.)([\w-]+\.)?simply-hentai\.com"
r"(?!/(?:album|gifs?|images?|series)(?:/|$))"
r"((?:/(?!(?:page|all-pages)(?:/|\.|$))[^/?#]+)+)")
test = (
(("https://original-work.simply-hentai.com"
"/amazon-no-hiyaku-amazon-elixir"), {
"url": "21613585ae5ec2f69ea579e9713f536fceab5bd5",
"keyword": "9e87a0973553b2922ddee37958b8f5d87910af72",
}),
("https://www.simply-hentai.com/notfound", {
"exception": exception.GalleryDLException,
}),
# custom subdomain
("https://pokemon.simply-hentai.com/mao-friends-9bc39"),
# www subdomain, two path segments
("https://www.simply-hentai.com/vocaloid/black-magnet"),
)
def __init__(self, match):
subdomain, path = match.groups()
if subdomain and subdomain not in ("www.", "old."):
path = "/" + subdomain.rstrip(".") + path
url = "https://old.simply-hentai.com" + path
GalleryExtractor.__init__(self, match, url)
self.session.headers["Referer"] = url
def metadata(self, page):
extr = text.extract_from(page)
split = text.split_html
title = extr('<meta property="og:title" content="', '"')
image = extr('<meta property="og:image" content="', '"')
if not title:
raise exception.NotFoundError("gallery")
data = {
"title" : text.unescape(title),
"gallery_id": text.parse_int(image.split("/")[-2]),
"parody" : split(extr('box-title">Series</div>', '</div>')),
"language" : text.remove_html(extr(
'box-title">Language</div>', '</div>')) or None,
"characters": split(extr('box-title">Characters</div>', '</div>')),
"tags" : split(extr('box-title">Tags</div>', '</div>')),
"artist" : split(extr('box-title">Artists</div>', '</div>')),
"date" : text.parse_datetime(text.remove_html(
extr('Uploaded', '</div>')), "%d.%m.%Y"),
}
data["lang"] = util.language_to_code(data["language"])
return data
def images(self, _):
url = self.gallery_url + "/all-pages"
headers = {"Accept": "application/json"}
images = self.request(url, headers=headers).json()
return [
(
urls["full"].replace("/giant_thumb_", "/"),
{"image_id": text.parse_int(image_id)},
)
for image_id, urls in sorted(images.items())
]
class SimplyhentaiImageExtractor(Extractor):
"""Extractor for individual images from simply-hentai.com"""
category = "simplyhentai"
subcategory = "image"
directory_fmt = ("{category}", "{type}s")
filename_fmt = "{category}_{token}{title:?_//}.{extension}"
archive_fmt = "{token}"
pattern = (r"(?:https?://)?(?:www\.)?(simply-hentai\.com"
r"/(image|gif)/[^/?#]+)")
test = (
(("https://www.simply-hentai.com/image"
"/pheromomania-vol-1-kanzenban-isao-3949d8b3-400c-4b6"), {
"url": "3d8eb55240a960134891bd77fe1df7988fcdc455",
"keyword": "e10e5588481cab68329ef6ec1e5325206b2079a2",
}),
("https://www.simply-hentai.com/gif/8915dfcf-0b6a-47c", {
"url": "f73916527211b4a40f26568ee26cd8999f5f4f30",
"keyword": "f94d775177fed918759c8a78a50976f867425b48",
}),
)
def __init__(self, match):
Extractor.__init__(self, match)
self.page_url = "https://old." + match.group(1)
self.type = match.group(2)
def items(self):
extr = text.extract_from(self.request(self.page_url).text)
title = extr('"og:title" content="' , '"')
descr = extr('"og:description" content="', '"')
url = extr('&quot;image&quot;:&quot;' , '&')
url = extr("&quot;content&quot;:&quot;", "&") or url
tags = text.extract(descr, " tagged with ", " online for free ")[0]
if tags:
tags = tags.split(", ")
tags[-1] = tags[-1].partition(" ")[2]
else:
tags = []
if url.startswith("//"):
url = "https:" + url
data = text.nameext_from_url(url, {
"title": text.unescape(title) if title else "",
"tags": tags,
"type": self.type,
})
data["token"] = data["filename"].rpartition("_")[2]
yield Message.Directory, data
yield Message.Url, url, data
class SimplyhentaiVideoExtractor(Extractor):
"""Extractor for hentai videos from simply-hentai.com"""
category = "simplyhentai"
subcategory = "video"
directory_fmt = ("{category}", "{type}s")
filename_fmt = "{title}{episode:?_//>02}.{extension}"
archive_fmt = "{title}_{episode}"
pattern = r"(?:https?://)?(videos\.simply-hentai\.com/[^/?#]+)"
test = (
("https://videos.simply-hentai.com/creamy-pie-episode-02", {
"pattern": r"https://www\.googleapis\.com/drive/v3/files"
r"/0B1ecQ8ZVLm3JcHZzQzBnVy1ZUmc\?alt=media&key=[\w-]+",
"keyword": "706790708b14773efc1e075ddd3b738a375348a5",
"options": (("verify", False),),
"count": 1,
}),
(("https://videos.simply-hentai.com"
"/1715-tifa-in-hentai-gang-bang-3d-movie"), {
"url": "ad9a36ae06c601b6490e3c401834b4949d947eb0",
"keyword": "f9dad94fbde9c95859e631ff4f07297a9567b874",
}),
)
def __init__(self, match):
Extractor.__init__(self, match)
self.page_url = "https://" + match.group(1)
def items(self):
page = self.request(self.page_url).text
title, pos = text.extract(page, "<title>", "</title>")
tags , pos = text.extract(page, ">Tags</div>", "</div>", pos)
date , pos = text.extract(page, ">Upload Date</div>", "</div>", pos)
title = title.rpartition(" - ")[0]
if "<video" in page:
video_url = text.extract(page, '<source src="', '"', pos)[0]
episode = 0
else:
# video url from myhentai.tv embed
pos = page.index('<div class="video-frame-container">', pos)
embed_url = text.extract(page, 'src="', '"', pos)[0].replace(
"embedplayer.php?link=", "embed.php?name=")
embed_page = self.request(embed_url).text
video_url = text.extract(embed_page, '"file":"', '"')[0]
title, _, episode = title.rpartition(" Episode ")
if video_url.startswith("//"):
video_url = "https:" + video_url
data = text.nameext_from_url(video_url, {
"title": text.unescape(title),
"episode": text.parse_int(episode),
"tags": text.split_html(tags)[::2],
"type": "video",
"date": text.parse_datetime(text.remove_html(
date), "%B %d, %Y %H:%M"),
})
yield Message.Directory, data
yield Message.Url, video_url, data