1
0
mirror of https://github.com/mikf/gallery-dl.git synced 2024-11-26 12:42:29 +01:00

[seisoparty] remove module

This commit is contained in:
Mike Fährmann 2021-11-09 22:41:04 +01:00
parent efa178cc91
commit 37c9dedee1
No known key found for this signature in database
GPG Key ID: 5680CA389D365A88
3 changed files with 0 additions and 208 deletions

View File

@ -673,12 +673,6 @@ Consider all sites to be NSFW unless otherwise known.
<td>Articles, Tag Searches</td> <td>Articles, Tag Searches</td>
<td></td> <td></td>
</tr> </tr>
<tr>
<td>Seiso</td>
<td>https://seiso.party/</td>
<td>Favorites, Posts, User Profiles</td>
<td>Supported</td>
</tr>
<tr> <tr>
<td>Sen Manga</td> <td>Sen Manga</td>
<td>https://raw.senmanga.com/</td> <td>https://raw.senmanga.com/</td>

View File

@ -111,7 +111,6 @@ modules = [
"sankaku", "sankaku",
"sankakucomplex", "sankakucomplex",
"seiga", "seiga",
"seisoparty",
"senmanga", "senmanga",
"sexcom", "sexcom",
"simplyhentai", "simplyhentai",

View File

@ -1,201 +0,0 @@
# -*- coding: utf-8 -*-
# Copyright 2021 Mike Fährmann
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License version 2 as
# published by the Free Software Foundation.
"""Extractors for https://seiso.party/"""
from .common import Extractor, Message
from .. import text, exception
from ..cache import cache
import re
class SeisopartyExtractor(Extractor):
"""Base class for seisoparty extractors"""
category = "seisoparty"
root = "https://seiso.party"
directory_fmt = ("{category}", "{service}", "{username}")
filename_fmt = "{id}_{title}_{num:>02}_{filename}.{extension}"
archive_fmt = "{service}_{user}_{id}_{num}"
cookiedomain = ".seiso.party"
def __init__(self, match):
Extractor.__init__(self, match)
self.user_name = None
self._find_files = re.compile(
r'href="(https://cdn(?:-\d)?\.seiso\.party/files/[^"]+)').findall
def items(self):
self._prepare_ddosguard_cookies()
for post in self.posts():
files = post.pop("files")
yield Message.Directory, post
for post["num"], url in enumerate(files, 1):
yield Message.Url, url, text.nameext_from_url(url, post)
def _parse_post(self, page, post_id):
extr = text.extract_from(page)
return {
"service" : self.service,
"user" : self.user_id,
"username": self.user_name,
"id" : post_id,
"date" : text.parse_datetime(extr(
'<div class="margin-bottom-15 minor-text">', '<'),
"%Y-%m-%d %H:%M:%S %Z"),
"title" : text.unescape(extr('class="post-title">', '<')),
"content" : text.unescape(extr("\n<p>\n", "\n</p>\n").strip()),
"files" : self._find_files(page),
}
def login(self):
username, password = self._get_auth_info()
if username:
self._update_cookies(self._login_impl(username, password))
@cache(maxage=28*24*3600, keyarg=1)
def _login_impl(self, username, password):
self.log.info("Logging in as %s", username)
url = self.root + "/account/login"
data = {"username": username, "password": password}
response = self.request(url, method="POST", data=data)
if response.url.endswith("/account/login") and \
"Username or password is incorrect" in response.text:
raise exception.AuthenticationError()
return {c.name: c.value for c in response.history[0].cookies}
class SeisopartyUserExtractor(SeisopartyExtractor):
"""Extractor for all posts from a seiso.party user listing"""
subcategory = "user"
pattern = r"(?:https?://)?seiso\.party/artists/([^/?#]+)/([^/?#]+)"
test = (
("https://seiso.party/artists/fanbox/21", {
"pattern": r"https://cdn\.seiso\.party/files/fanbox/\d+/",
"count": ">=15",
"keyword": {
"content": str,
"date": "type:datetime",
"id": r"re:\d+",
"num": int,
"service": "fanbox",
"title": str,
"user": "21",
"username": "",
},
}),
)
def __init__(self, match):
SeisopartyExtractor.__init__(self, match)
self.service, self.user_id = match.groups()
def posts(self):
url = "{}/artists/{}/{}".format(self.root, self.service, self.user_id)
page = self.request(url).text
self.user_name, pos = text.extract(page, '<span class="title">', '<')
url = self.root + text.extract(
page, 'href="', '"', page.index('id="content"', pos))[0]
response = self.request(url)
headers = {"Referer": url}
while True:
yield self._parse_post(response.text, url.rpartition("/")[2])
response = self.request(url + "/next", headers=headers)
if url == response.url:
return
url = headers["Referer"] = response.url
class SeisopartyPostExtractor(SeisopartyExtractor):
"""Extractor for a single seiso.party post"""
subcategory = "post"
pattern = r"(?:https?://)?seiso\.party/post/([^/?#]+)/([^/?#]+)/([^/?#]+)"
test = (
("https://seiso.party/post/fanbox/21/371", {
"url": "75f13b92de0ce399b6163c3de18f1f36011c2366",
"count": 2,
"keyword": {
"content": "この前描いためぐるちゃんのPSDファイルです。<br/>"
"どうぞよろしくお願いします。",
"date": "dt:2021-05-06 12:38:31",
"extension": "re:psd|jpg",
"filename": "re:backcourt|ffb2ccb7a3586d05f9a4620329dd131e",
"id": "371",
"num": int,
"service": "fanbox",
"title": "MEGURU.PSD",
"user": "21",
"username": "",
},
}),
("https://seiso.party/post/patreon/429/95949", {
"pattern": r"https://cdn-2\.seiso\.party/files/patreon/95949/",
"count": 2,
}),
)
def __init__(self, match):
SeisopartyExtractor.__init__(self, match)
self.service, self.user_id, self.post_id = match.groups()
def posts(self):
url = "{}/artists/{}/{}".format(self.root, self.service, self.user_id)
page = self.request(url).text
self.user_name, pos = text.extract(page, '<span class="title">', '<')
url = "{}/post/{}/{}/{}".format(
self.root, self.service, self.user_id, self.post_id)
return (self._parse_post(self.request(url).text, self.post_id),)
class SeisopartyFavoriteExtractor(SeisopartyExtractor):
"""Extractor for seiso.party favorites"""
subcategory = "favorite"
pattern = r"(?:https?://)?seiso\.party/favorites/artists/?(?:\?([^#]+))?"
test = (
("https://seiso.party/favorites/artists", {
"pattern": SeisopartyUserExtractor.pattern,
"url": "0c862434bc3bbbe84cbf41c3a6152473a8cde683",
"count": 3,
}),
("https://seiso.party/favorites/artists?sort=id&sort_direction=asc", {
"url": "629a8b9c6d3a8a64f521908bdb3d7426ac03f8d3",
}),
)
def __init__(self, match):
SeisopartyExtractor.__init__(self, match)
self.query = match.group(1)
def items(self):
self._prepare_ddosguard_cookies()
self.login()
url = self.root + "/favorites/artists"
data = {"_extractor": SeisopartyUserExtractor}
params = text.parse_query(self.query)
params["page"] = text.parse_int(params.get("page"), 1)
while True:
page = self.request(url, params=params).text
cnt = 0
for card in text.extract_iter(
page, '<div class="artist-card', '</a>'):
path = text.extract(card, '<a href="', '"')[0]
yield Message.Queue, self.root + path, data
cnt += 1
if cnt < 25:
return
params["page"] += 1