mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-02 17:22:31 +01:00
InfoExtractors: use _download_webpage in more IEs
IEs without tests are intact.
This commit is contained in:
parent
feecf22511
commit
46bfb42258
@ -112,7 +112,7 @@
|
|||||||
{
|
{
|
||||||
"name": "Escapist",
|
"name": "Escapist",
|
||||||
"url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
|
"url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
|
||||||
"file": "6618-Breaking-Down-Baldurs-Gate.flv",
|
"file": "6618-Breaking-Down-Baldurs-Gate.mp4",
|
||||||
"md5": "c6793dbda81388f4264c1ba18684a74d",
|
"md5": "c6793dbda81388f4264c1ba18684a74d",
|
||||||
"skip": "Fails with timeout on Travis"
|
"skip": "Fails with timeout on Travis"
|
||||||
},
|
},
|
||||||
|
@ -1069,13 +1069,7 @@ def _real_extract(self, url, new_video=True):
|
|||||||
|
|
||||||
# Retrieve video webpage to extract further information
|
# Retrieve video webpage to extract further information
|
||||||
request = compat_urllib_request.Request(url, None, std_headers)
|
request = compat_urllib_request.Request(url, None, std_headers)
|
||||||
try:
|
webpage = self._download_webpage(request, video_id)
|
||||||
self.report_download_webpage(video_id)
|
|
||||||
webpage_bytes = compat_urllib_request.urlopen(request).read()
|
|
||||||
webpage = webpage_bytes.decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'Unable to retrieve video webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
# Now we begin extracting as much information as we can from what we
|
# Now we begin extracting as much information as we can from what we
|
||||||
# retrieved. First we extract the information common to all extractors,
|
# retrieved. First we extract the information common to all extractors,
|
||||||
@ -1682,10 +1676,6 @@ def suitable(cls, url):
|
|||||||
"""Receives a URL and returns True if suitable for this IE."""
|
"""Receives a URL and returns True if suitable for this IE."""
|
||||||
return re.match(cls._VALID_URL, url, re.VERBOSE) is not None
|
return re.match(cls._VALID_URL, url, re.VERBOSE) is not None
|
||||||
|
|
||||||
def report_download_page(self, playlist_id, pagenum):
|
|
||||||
"""Report attempt to download playlist page with given number."""
|
|
||||||
self._downloader.to_screen(u'[youtube] PL %s: Downloading page #%s' % (playlist_id, pagenum))
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
# Extract playlist id
|
# Extract playlist id
|
||||||
mobj = re.match(self._VALID_URL, url, re.VERBOSE)
|
mobj = re.match(self._VALID_URL, url, re.VERBOSE)
|
||||||
@ -1699,14 +1689,8 @@ def _real_extract(self, url):
|
|||||||
videos = []
|
videos = []
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
self.report_download_page(playlist_id, page_num)
|
|
||||||
|
|
||||||
url = self._TEMPLATE_URL % (playlist_id, self._MAX_RESULTS, self._MAX_RESULTS * (page_num - 1) + 1)
|
url = self._TEMPLATE_URL % (playlist_id, self._MAX_RESULTS, self._MAX_RESULTS * (page_num - 1) + 1)
|
||||||
try:
|
page = self._download_webpage(url, playlist_id, u'Downloading page #%s' % page_num)
|
||||||
page = compat_urllib_request.urlopen(url).read().decode('utf8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
response = json.loads(page)
|
response = json.loads(page)
|
||||||
@ -1745,10 +1729,6 @@ class YoutubeChannelIE(InfoExtractor):
|
|||||||
_MORE_PAGES_URL = 'http://www.youtube.com/channel_ajax?action_load_more_videos=1&flow=list&paging=%s&view=0&sort=da&channel_id=%s'
|
_MORE_PAGES_URL = 'http://www.youtube.com/channel_ajax?action_load_more_videos=1&flow=list&paging=%s&view=0&sort=da&channel_id=%s'
|
||||||
IE_NAME = u'youtube:channel'
|
IE_NAME = u'youtube:channel'
|
||||||
|
|
||||||
def report_download_page(self, channel_id, pagenum):
|
|
||||||
"""Report attempt to download channel page with given number."""
|
|
||||||
self._downloader.to_screen(u'[youtube] Channel %s: Downloading page #%s' % (channel_id, pagenum))
|
|
||||||
|
|
||||||
def extract_videos_from_page(self, page):
|
def extract_videos_from_page(self, page):
|
||||||
ids_in_page = []
|
ids_in_page = []
|
||||||
for mobj in re.finditer(r'href="/watch\?v=([0-9A-Za-z_-]+)&?', page):
|
for mobj in re.finditer(r'href="/watch\?v=([0-9A-Za-z_-]+)&?', page):
|
||||||
@ -1768,14 +1748,9 @@ def _real_extract(self, url):
|
|||||||
video_ids = []
|
video_ids = []
|
||||||
pagenum = 1
|
pagenum = 1
|
||||||
|
|
||||||
self.report_download_page(channel_id, pagenum)
|
|
||||||
url = self._TEMPLATE_URL % (channel_id, pagenum)
|
url = self._TEMPLATE_URL % (channel_id, pagenum)
|
||||||
request = compat_urllib_request.Request(url)
|
page = self._download_webpage(url, channel_id,
|
||||||
try:
|
u'Downloading page #%s' % pagenum)
|
||||||
page = compat_urllib_request.urlopen(request).read().decode('utf8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
# Extract video identifiers
|
# Extract video identifiers
|
||||||
ids_in_page = self.extract_videos_from_page(page)
|
ids_in_page = self.extract_videos_from_page(page)
|
||||||
@ -1786,14 +1761,9 @@ def _real_extract(self, url):
|
|||||||
while True:
|
while True:
|
||||||
pagenum = pagenum + 1
|
pagenum = pagenum + 1
|
||||||
|
|
||||||
self.report_download_page(channel_id, pagenum)
|
|
||||||
url = self._MORE_PAGES_URL % (pagenum, channel_id)
|
url = self._MORE_PAGES_URL % (pagenum, channel_id)
|
||||||
request = compat_urllib_request.Request(url)
|
page = self._download_webpage(url, channel_id,
|
||||||
try:
|
u'Downloading page #%s' % pagenum)
|
||||||
page = compat_urllib_request.urlopen(request).read().decode('utf8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
page = json.loads(page)
|
page = json.loads(page)
|
||||||
|
|
||||||
@ -1820,11 +1790,6 @@ class YoutubeUserIE(InfoExtractor):
|
|||||||
_VIDEO_INDICATOR = r'/watch\?v=(.+?)[\<&]'
|
_VIDEO_INDICATOR = r'/watch\?v=(.+?)[\<&]'
|
||||||
IE_NAME = u'youtube:user'
|
IE_NAME = u'youtube:user'
|
||||||
|
|
||||||
def report_download_page(self, username, start_index):
|
|
||||||
"""Report attempt to download user page."""
|
|
||||||
self._downloader.to_screen(u'[youtube] user %s: Downloading video ids from %d to %d' %
|
|
||||||
(username, start_index, start_index + self._GDATA_PAGE_SIZE))
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
# Extract username
|
# Extract username
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
@ -1844,15 +1809,10 @@ def _real_extract(self, url):
|
|||||||
|
|
||||||
while True:
|
while True:
|
||||||
start_index = pagenum * self._GDATA_PAGE_SIZE + 1
|
start_index = pagenum * self._GDATA_PAGE_SIZE + 1
|
||||||
self.report_download_page(username, start_index)
|
|
||||||
|
|
||||||
request = compat_urllib_request.Request(self._GDATA_URL % (username, self._GDATA_PAGE_SIZE, start_index))
|
gdata_url = self._GDATA_URL % (username, self._GDATA_PAGE_SIZE, start_index)
|
||||||
|
page = self._download_webpage(gdata_url, username,
|
||||||
try:
|
u'Downloading video ids from %d to %d' % (start_index, start_index + self._GDATA_PAGE_SIZE))
|
||||||
page = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
# Extract video identifiers
|
# Extract video identifiers
|
||||||
ids_in_page = []
|
ids_in_page = []
|
||||||
@ -1886,11 +1846,6 @@ class BlipTVUserIE(InfoExtractor):
|
|||||||
_PAGE_SIZE = 12
|
_PAGE_SIZE = 12
|
||||||
IE_NAME = u'blip.tv:user'
|
IE_NAME = u'blip.tv:user'
|
||||||
|
|
||||||
def report_download_page(self, username, pagenum):
|
|
||||||
"""Report attempt to download user page."""
|
|
||||||
self.to_screen(u'user %s: Downloading video ids from page %d' %
|
|
||||||
(username, pagenum))
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
# Extract username
|
# Extract username
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
@ -1902,15 +1857,9 @@ def _real_extract(self, url):
|
|||||||
|
|
||||||
page_base = 'http://m.blip.tv/pr/show_get_full_episode_list?users_id=%s&lite=0&esi=1'
|
page_base = 'http://m.blip.tv/pr/show_get_full_episode_list?users_id=%s&lite=0&esi=1'
|
||||||
|
|
||||||
request = compat_urllib_request.Request(url)
|
page = self._download_webpage(url, username, u'Downloading user page')
|
||||||
|
mobj = re.search(r'data-users-id="([^"]+)"', page)
|
||||||
try:
|
page_base = page_base % mobj.group(1)
|
||||||
page = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
|
||||||
mobj = re.search(r'data-users-id="([^"]+)"', page)
|
|
||||||
page_base = page_base % mobj.group(1)
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
|
|
||||||
# Download video ids using BlipTV Ajax calls. Result size per
|
# Download video ids using BlipTV Ajax calls. Result size per
|
||||||
@ -1922,14 +1871,9 @@ def _real_extract(self, url):
|
|||||||
pagenum = 1
|
pagenum = 1
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
self.report_download_page(username, pagenum)
|
|
||||||
url = page_base + "&page=" + str(pagenum)
|
url = page_base + "&page=" + str(pagenum)
|
||||||
request = compat_urllib_request.Request( url )
|
page = self._download_webpage(url, username,
|
||||||
try:
|
u'Downloading video ids from page %d' % pagenum)
|
||||||
page = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
# Extract video identifiers
|
# Extract video identifiers
|
||||||
ids_in_page = []
|
ids_in_page = []
|
||||||
@ -2288,12 +2232,6 @@ def suitable(cls, url):
|
|||||||
"""Receives a URL and returns True if suitable for this IE."""
|
"""Receives a URL and returns True if suitable for this IE."""
|
||||||
return re.match(cls._VALID_URL, url, re.VERBOSE) is not None
|
return re.match(cls._VALID_URL, url, re.VERBOSE) is not None
|
||||||
|
|
||||||
def report_config_download(self, episode_id, media_id):
|
|
||||||
self.to_screen(u'%s: Downloading configuration for %s' % (episode_id, media_id))
|
|
||||||
|
|
||||||
def report_index_download(self, episode_id):
|
|
||||||
self.to_screen(u'%s: Downloading show index' % episode_id)
|
|
||||||
|
|
||||||
def _print_formats(self, formats):
|
def _print_formats(self, formats):
|
||||||
print('Available formats:')
|
print('Available formats:')
|
||||||
for x in formats:
|
for x in formats:
|
||||||
@ -2327,15 +2265,8 @@ def _real_extract(self, url):
|
|||||||
else:
|
else:
|
||||||
epTitle = mobj.group('episode')
|
epTitle = mobj.group('episode')
|
||||||
|
|
||||||
req = compat_urllib_request.Request(url)
|
|
||||||
self.report_extraction(epTitle)
|
self.report_extraction(epTitle)
|
||||||
try:
|
webpage = self._download_webpage(url, epTitle)
|
||||||
htmlHandle = compat_urllib_request.urlopen(req)
|
|
||||||
html = htmlHandle.read()
|
|
||||||
webpage = html.decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
if dlNewest:
|
if dlNewest:
|
||||||
url = htmlHandle.geturl()
|
url = htmlHandle.geturl()
|
||||||
mobj = re.match(self._VALID_URL, url, re.VERBOSE)
|
mobj = re.match(self._VALID_URL, url, re.VERBOSE)
|
||||||
@ -2363,12 +2294,9 @@ def _real_extract(self, url):
|
|||||||
|
|
||||||
uri = mMovieParams[0][1]
|
uri = mMovieParams[0][1]
|
||||||
indexUrl = 'http://shadow.comedycentral.com/feeds/video_player/mrss/?' + compat_urllib_parse.urlencode({'uri': uri})
|
indexUrl = 'http://shadow.comedycentral.com/feeds/video_player/mrss/?' + compat_urllib_parse.urlencode({'uri': uri})
|
||||||
self.report_index_download(epTitle)
|
indexXml = self._download_webpage(indexUrl, epTitle,
|
||||||
try:
|
u'Downloading show index',
|
||||||
indexXml = compat_urllib_request.urlopen(indexUrl).read()
|
u'unable to download episode index')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download episode index: ' + compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
results = []
|
results = []
|
||||||
|
|
||||||
@ -2383,13 +2311,8 @@ def _real_extract(self, url):
|
|||||||
|
|
||||||
configUrl = ('http://www.comedycentral.com/global/feeds/entertainment/media/mediaGenEntertainment.jhtml?' +
|
configUrl = ('http://www.comedycentral.com/global/feeds/entertainment/media/mediaGenEntertainment.jhtml?' +
|
||||||
compat_urllib_parse.urlencode({'uri': mediaId}))
|
compat_urllib_parse.urlencode({'uri': mediaId}))
|
||||||
configReq = compat_urllib_request.Request(configUrl)
|
configXml = self._download_webpage(configUrl, epTitle,
|
||||||
self.report_config_download(epTitle, shortMediaId)
|
u'Downloading configuration for %s' % shortMediaId)
|
||||||
try:
|
|
||||||
configXml = compat_urllib_request.urlopen(configReq).read()
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
cdoc = xml.etree.ElementTree.fromstring(configXml)
|
cdoc = xml.etree.ElementTree.fromstring(configXml)
|
||||||
turls = []
|
turls = []
|
||||||
@ -2446,9 +2369,6 @@ class EscapistIE(InfoExtractor):
|
|||||||
_VALID_URL = r'^(https?://)?(www\.)?escapistmagazine\.com/videos/view/(?P<showname>[^/]+)/(?P<episode>[^/?]+)[/?]?.*$'
|
_VALID_URL = r'^(https?://)?(www\.)?escapistmagazine\.com/videos/view/(?P<showname>[^/]+)/(?P<episode>[^/?]+)[/?]?.*$'
|
||||||
IE_NAME = u'escapist'
|
IE_NAME = u'escapist'
|
||||||
|
|
||||||
def report_config_download(self, showName):
|
|
||||||
self.to_screen(u'%s: Downloading configuration' % showName)
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
if mobj is None:
|
if mobj is None:
|
||||||
@ -2458,14 +2378,7 @@ def _real_extract(self, url):
|
|||||||
videoId = mobj.group('episode')
|
videoId = mobj.group('episode')
|
||||||
|
|
||||||
self.report_extraction(showName)
|
self.report_extraction(showName)
|
||||||
try:
|
webPage = self._download_webpage(url, showName)
|
||||||
webPage = compat_urllib_request.urlopen(url)
|
|
||||||
webPageBytes = webPage.read()
|
|
||||||
m = re.match(r'text/html; charset="?([^"]+)"?', webPage.headers['Content-Type'])
|
|
||||||
webPage = webPageBytes.decode(m.group(1) if m else 'utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download webpage: ' + compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
descMatch = re.search('<meta name="description" content="([^"]*)"', webPage)
|
descMatch = re.search('<meta name="description" content="([^"]*)"', webPage)
|
||||||
description = unescapeHTML(descMatch.group(1))
|
description = unescapeHTML(descMatch.group(1))
|
||||||
@ -2476,14 +2389,9 @@ def _real_extract(self, url):
|
|||||||
configUrlMatch = re.search('config=(.*)$', playerUrl)
|
configUrlMatch = re.search('config=(.*)$', playerUrl)
|
||||||
configUrl = compat_urllib_parse.unquote(configUrlMatch.group(1))
|
configUrl = compat_urllib_parse.unquote(configUrlMatch.group(1))
|
||||||
|
|
||||||
self.report_config_download(showName)
|
configJSON = self._download_webpage(configUrl, showName,
|
||||||
try:
|
u'Downloading configuration',
|
||||||
configJSON = compat_urllib_request.urlopen(configUrl)
|
u'unable to download configuration')
|
||||||
m = re.match(r'text/html; charset="?([^"]+)"?', configJSON.headers['Content-Type'])
|
|
||||||
configJSON = configJSON.read().decode(m.group(1) if m else 'utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download configuration: ' + compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
# Technically, it's JavaScript, not JSON
|
# Technically, it's JavaScript, not JSON
|
||||||
configJSON = configJSON.replace("'", '"')
|
configJSON = configJSON.replace("'", '"')
|
||||||
@ -2661,31 +2569,22 @@ def _real_extract(self, url):
|
|||||||
# extract simple title (uploader + slug of song title)
|
# extract simple title (uploader + slug of song title)
|
||||||
slug_title = mobj.group(2)
|
slug_title = mobj.group(2)
|
||||||
simple_title = uploader + u'-' + slug_title
|
simple_title = uploader + u'-' + slug_title
|
||||||
|
full_title = '%s/%s' % (uploader, slug_title)
|
||||||
|
|
||||||
self.report_resolve('%s/%s' % (uploader, slug_title))
|
self.report_resolve(full_title)
|
||||||
|
|
||||||
url = 'http://soundcloud.com/%s/%s' % (uploader, slug_title)
|
url = 'http://soundcloud.com/%s/%s' % (uploader, slug_title)
|
||||||
resolv_url = 'http://api.soundcloud.com/resolve.json?url=' + url + '&client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
resolv_url = 'http://api.soundcloud.com/resolve.json?url=' + url + '&client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
||||||
request = compat_urllib_request.Request(resolv_url)
|
info_json = self._download_webpage(resolv_url, full_title, u'Downloading info JSON')
|
||||||
try:
|
|
||||||
info_json_bytes = compat_urllib_request.urlopen(request).read()
|
|
||||||
info_json = info_json_bytes.decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
info = json.loads(info_json)
|
info = json.loads(info_json)
|
||||||
video_id = info['id']
|
video_id = info['id']
|
||||||
self.report_extraction('%s/%s' % (uploader, slug_title))
|
self.report_extraction(full_title)
|
||||||
|
|
||||||
streams_url = 'https://api.sndcdn.com/i1/tracks/' + str(video_id) + '/streams?client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
streams_url = 'https://api.sndcdn.com/i1/tracks/' + str(video_id) + '/streams?client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
||||||
request = compat_urllib_request.Request(streams_url)
|
stream_json = self._download_webpage(streams_url, full_title,
|
||||||
try:
|
u'Downloading stream definitions',
|
||||||
stream_json_bytes = compat_urllib_request.urlopen(request).read()
|
u'unable to download stream definitions')
|
||||||
stream_json = stream_json_bytes.decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download stream definitions: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
streams = json.loads(stream_json)
|
streams = json.loads(stream_json)
|
||||||
mediaURL = streams['http_mp3_128_url']
|
mediaURL = streams['http_mp3_128_url']
|
||||||
@ -2728,18 +2627,13 @@ def _real_extract(self, url):
|
|||||||
# extract simple title (uploader + slug of song title)
|
# extract simple title (uploader + slug of song title)
|
||||||
slug_title = mobj.group(2)
|
slug_title = mobj.group(2)
|
||||||
simple_title = uploader + u'-' + slug_title
|
simple_title = uploader + u'-' + slug_title
|
||||||
|
full_title = '%s/sets/%s' % (uploader, slug_title)
|
||||||
|
|
||||||
self.report_resolve('%s/sets/%s' % (uploader, slug_title))
|
self.report_resolve(full_title)
|
||||||
|
|
||||||
url = 'http://soundcloud.com/%s/sets/%s' % (uploader, slug_title)
|
url = 'http://soundcloud.com/%s/sets/%s' % (uploader, slug_title)
|
||||||
resolv_url = 'http://api.soundcloud.com/resolve.json?url=' + url + '&client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
resolv_url = 'http://api.soundcloud.com/resolve.json?url=' + url + '&client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
||||||
request = compat_urllib_request.Request(resolv_url)
|
info_json = self._download_webpage(resolv_url, full_title)
|
||||||
try:
|
|
||||||
info_json_bytes = compat_urllib_request.urlopen(request).read()
|
|
||||||
info_json = info_json_bytes.decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
videos = []
|
videos = []
|
||||||
info = json.loads(info_json)
|
info = json.loads(info_json)
|
||||||
@ -2748,19 +2642,14 @@ def _real_extract(self, url):
|
|||||||
self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err['error_message']))
|
self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err['error_message']))
|
||||||
return
|
return
|
||||||
|
|
||||||
|
self.report_extraction(full_title)
|
||||||
for track in info['tracks']:
|
for track in info['tracks']:
|
||||||
video_id = track['id']
|
video_id = track['id']
|
||||||
self.report_extraction('%s/sets/%s' % (uploader, slug_title))
|
|
||||||
|
|
||||||
streams_url = 'https://api.sndcdn.com/i1/tracks/' + str(video_id) + '/streams?client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
streams_url = 'https://api.sndcdn.com/i1/tracks/' + str(video_id) + '/streams?client_id=b45b1aa10f1ac2941910a7f0d10f8e28'
|
||||||
request = compat_urllib_request.Request(streams_url)
|
stream_json = self._download_webpage(streams_url, video_id, u'Downloading track info JSON')
|
||||||
try:
|
|
||||||
stream_json_bytes = compat_urllib_request.urlopen(request).read()
|
|
||||||
stream_json = stream_json_bytes.decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download stream definitions: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
|
self.report_extraction(video_id)
|
||||||
streams = json.loads(stream_json)
|
streams = json.loads(stream_json)
|
||||||
mediaURL = streams['http_mp3_128_url']
|
mediaURL = streams['http_mp3_128_url']
|
||||||
|
|
||||||
@ -3155,18 +3044,11 @@ def _real_extract(self, url):
|
|||||||
|
|
||||||
info_url = 'http://v.youku.com/player/getPlayList/VideoIDS/' + video_id
|
info_url = 'http://v.youku.com/player/getPlayList/VideoIDS/' + video_id
|
||||||
|
|
||||||
request = compat_urllib_request.Request(info_url, None, std_headers)
|
jsondata = self._download_webpage(info_url, video_id)
|
||||||
try:
|
|
||||||
self.report_download_webpage(video_id)
|
|
||||||
jsondata = compat_urllib_request.urlopen(request).read()
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'Unable to retrieve video webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
self.report_extraction(video_id)
|
self.report_extraction(video_id)
|
||||||
try:
|
try:
|
||||||
jsonstr = jsondata.decode('utf-8')
|
config = json.loads(jsondata)
|
||||||
config = json.loads(jsonstr)
|
|
||||||
|
|
||||||
video_title = config['data'][0]['title']
|
video_title = config['data'][0]['title']
|
||||||
seed = config['data'][0]['seed']
|
seed = config['data'][0]['seed']
|
||||||
@ -3234,15 +3116,8 @@ def _real_extract(self, url):
|
|||||||
return
|
return
|
||||||
video_id = mobj.group(1)
|
video_id = mobj.group(1)
|
||||||
|
|
||||||
self.report_download_webpage(video_id)
|
|
||||||
|
|
||||||
# Get webpage content
|
# Get webpage content
|
||||||
try:
|
webpage = self._download_webpage(url, video_id)
|
||||||
webpage_bytes = compat_urllib_request.urlopen(url).read()
|
|
||||||
webpage = webpage_bytes.decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download video webpage: %s' % err)
|
|
||||||
return
|
|
||||||
|
|
||||||
result = re.search(self.VIDEO_URL_RE, webpage)
|
result = re.search(self.VIDEO_URL_RE, webpage)
|
||||||
if result is None:
|
if result is None:
|
||||||
@ -3314,12 +3189,7 @@ def _real_extract(self, url):
|
|||||||
|
|
||||||
# Step 1, Retrieve post webpage to extract further information
|
# Step 1, Retrieve post webpage to extract further information
|
||||||
self.report_extract_entry(post_url)
|
self.report_extract_entry(post_url)
|
||||||
request = compat_urllib_request.Request(post_url)
|
webpage = self._download_webpage(post_url, video_id, u'Downloading entry webpage')
|
||||||
try:
|
|
||||||
webpage = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'Unable to retrieve entry webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
# Extract update date
|
# Extract update date
|
||||||
upload_date = None
|
upload_date = None
|
||||||
@ -3356,12 +3226,7 @@ def _real_extract(self, url):
|
|||||||
self._downloader.report_error(u'unable to extract video page URL')
|
self._downloader.report_error(u'unable to extract video page URL')
|
||||||
|
|
||||||
video_page = mobj.group(1)
|
video_page = mobj.group(1)
|
||||||
request = compat_urllib_request.Request(video_page)
|
webpage = self._download_webpage(video_page, video_id, u'Downloading video page')
|
||||||
try:
|
|
||||||
webpage = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'Unable to retrieve video webpage: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
self.report_extract_vid_page(video_page)
|
self.report_extract_vid_page(video_page)
|
||||||
|
|
||||||
|
|
||||||
@ -3448,14 +3313,10 @@ def report_download_page(self, channel, offset):
|
|||||||
(channel, offset, offset + self._JUSTIN_PAGE_LIMIT))
|
(channel, offset, offset + self._JUSTIN_PAGE_LIMIT))
|
||||||
|
|
||||||
# Return count of items, list of *valid* items
|
# Return count of items, list of *valid* items
|
||||||
def _parse_page(self, url):
|
def _parse_page(self, url, video_id):
|
||||||
try:
|
webpage = self._download_webpage(url, video_id,
|
||||||
urlh = compat_urllib_request.urlopen(url)
|
u'Downloading video info JSON',
|
||||||
webpage_bytes = urlh.read()
|
u'unable to download video info JSON')
|
||||||
webpage = webpage_bytes.decode('utf-8', 'ignore')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
self._downloader.report_error(u'unable to download video info JSON: %s' % compat_str(err))
|
|
||||||
return
|
|
||||||
|
|
||||||
response = json.loads(webpage)
|
response = json.loads(webpage)
|
||||||
if type(response) != list:
|
if type(response) != list:
|
||||||
@ -3507,7 +3368,7 @@ def _real_extract(self, url):
|
|||||||
if paged:
|
if paged:
|
||||||
self.report_download_page(video_id, offset)
|
self.report_download_page(video_id, offset)
|
||||||
page_url = api + ('?offset=%d&limit=%d' % (offset, limit))
|
page_url = api + ('?offset=%d&limit=%d' % (offset, limit))
|
||||||
page_count, page_info = self._parse_page(page_url)
|
page_count, page_info = self._parse_page(page_url, video_id)
|
||||||
info.extend(page_info)
|
info.extend(page_info)
|
||||||
if not paged or page_count != limit:
|
if not paged or page_count != limit:
|
||||||
break
|
break
|
||||||
@ -3627,14 +3488,13 @@ class WorldStarHipHopIE(InfoExtractor):
|
|||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
_src_url = r"""(http://(hw-videos|hw-post1).*(?:mp4|flv))"""
|
_src_url = r"""(http://(hw-videos|hw-post1).*(?:mp4|flv))"""
|
||||||
|
|
||||||
webpage_src = compat_urllib_request.urlopen(url).read()
|
|
||||||
webpage_src = webpage_src.decode('utf-8')
|
|
||||||
|
|
||||||
mobj = re.search(_src_url, webpage_src)
|
|
||||||
|
|
||||||
m = re.match(self._VALID_URL, url)
|
m = re.match(self._VALID_URL, url)
|
||||||
video_id = m.group('id')
|
video_id = m.group('id')
|
||||||
|
|
||||||
|
webpage_src = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
mobj = re.search(_src_url, webpage_src)
|
||||||
|
|
||||||
if mobj is not None:
|
if mobj is not None:
|
||||||
video_url = mobj.group()
|
video_url = mobj.group()
|
||||||
if 'mp4' in video_url:
|
if 'mp4' in video_url:
|
||||||
|
Loading…
Reference in New Issue
Block a user