diff --git a/README.md b/README.md
index 2ca021125..9dffd99ab 100644
--- a/README.md
+++ b/README.md
@@ -46,15 +46,16 @@ ## Filesystem Options:
-A, --auto-number number downloaded files starting from 00000
-o, --output TEMPLATE output filename template. Use %(title)s to get the
title, %(uploader)s for the uploader name,
- %(autonumber)s to get an automatically incremented
- number, %(ext)s for the filename extension,
- %(upload_date)s for the upload date (YYYYMMDD),
- %(extractor)s for the provider (youtube, metacafe,
- etc), %(id)s for the video id and %% for a literal
- percent. Use - to output to stdout. Can also be
- used to download to a different directory, for
- example with -o '/my/downloads/%(uploader)s/%(title
- )s-%(id)s.%(ext)s' .
+ %(uploader_id)s for the uploader nickname if
+ different, %(autonumber)s to get an automatically
+ incremented number, %(ext)s for the filename
+ extension, %(upload_date)s for the upload date
+ (YYYYMMDD), %(extractor)s for the provider
+ (youtube, metacafe, etc), %(id)s for the video id
+ and %% for a literal percent. Use - to output to
+ stdout. Can also be used to download to a different
+ directory, for example with -o '/my/downloads/%(upl
+ oader)s/%(title)s-%(id)s.%(ext)s' .
--restrict-filenames Restrict filenames to only ASCII characters, and
avoid "&" and spaces in filenames
-a, --batch-file FILE file containing URLs to download ('-' for stdin)
diff --git a/test/test_download.py b/test/test_download.py
index bce0e4fcd..1ee1b334d 100644
--- a/test/test_download.py
+++ b/test/test_download.py
@@ -6,6 +6,7 @@
import json
import unittest
import sys
+import hashlib
import socket
# Allow direct execution
@@ -24,12 +25,15 @@
proxy_handler = compat_urllib_request.ProxyHandler()
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
compat_urllib_request.install_opener(opener)
-socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
class FileDownloader(youtube_dl.FileDownloader):
def __init__(self, *args, **kwargs):
- youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
self.to_stderr = self.to_screen
+ self.processed_info_dicts = []
+ return youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
+ def process_info(self, info_dict):
+ self.processed_info_dicts.append(info_dict)
+ return youtube_dl.FileDownloader.process_info(self, info_dict)
def _file_md5(fn):
with open(fn, 'rb') as f:
@@ -40,6 +44,7 @@ def _file_md5(fn):
with io.open(PARAMETERS_FILE, encoding='utf-8') as pf:
parameters = json.load(pf)
+
class TestDownload(unittest.TestCase):
def setUp(self):
self.parameters = parameters
@@ -68,18 +73,28 @@ def test_template(self):
if 'skip' in test_case:
print('Skipping: {0}'.format(test_case['skip']))
return
+
params = dict(self.parameters) # Duplicate it locally
for p in test_case.get('params', {}):
params[p] = test_case['params'][p]
+
fd = FileDownloader(params)
fd.add_info_extractor(ie())
for ien in test_case.get('add_ie', []):
fd.add_info_extractor(getattr(youtube_dl.InfoExtractors, ien + 'IE')())
fd.download([test_case['url']])
+
self.assertTrue(os.path.exists(test_case['file']))
if 'md5' in test_case:
md5_for_file = _file_md5(test_case['file'])
self.assertEqual(md5_for_file, test_case['md5'])
+ info_dict = fd.processed_info_dicts[0]
+ for (info_field, value) in test_case.get('info_dict', {}).items():
+ if value.startswith('md5:'):
+ md5_info_value = hashlib.md5(info_dict.get(info_field, '')).hexdigest()
+ self.assertEqual(value[3:], md5_info_value)
+ else:
+ self.assertEqual(value, info_dict.get(info_field))
return test_template
diff --git a/test/test_write_info_json.py b/test/test_write_info_json.py
index ebf543980..8134dda37 100644
--- a/test/test_write_info_json.py
+++ b/test/test_write_info_json.py
@@ -3,7 +3,6 @@
import json
import os
-import socket
import sys
import unittest
@@ -22,7 +21,6 @@
proxy_handler = compat_urllib_request.ProxyHandler()
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
compat_urllib_request.install_opener(opener)
-socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
class FileDownloader(youtube_dl.FileDownloader):
def __init__(self, *args, **kwargs):
diff --git a/test/test_youtube_lists.py b/test/test_youtube_lists.py
index e352e5ab9..3044e0852 100644
--- a/test/test_youtube_lists.py
+++ b/test/test_youtube_lists.py
@@ -2,7 +2,6 @@
import sys
import unittest
-import socket
import json
# Allow direct execution
@@ -22,7 +21,6 @@
proxy_handler = compat_urllib_request.ProxyHandler()
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
compat_urllib_request.install_opener(opener)
-socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
class FakeDownloader(object):
def __init__(self):
diff --git a/test/test_youtube_subtitles.py b/test/test_youtube_subtitles.py
index 64a391d14..5d3566a35 100644
--- a/test/test_youtube_subtitles.py
+++ b/test/test_youtube_subtitles.py
@@ -2,7 +2,6 @@
import sys
import unittest
-import socket
import json
import io
import hashlib
@@ -24,7 +23,6 @@
proxy_handler = compat_urllib_request.ProxyHandler()
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
compat_urllib_request.install_opener(opener)
-socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
class FakeDownloader(object):
def __init__(self):
diff --git a/test/tests.json b/test/tests.json
index 9bf56082e..dbff62676 100644
--- a/test/tests.json
+++ b/test/tests.json
@@ -2,7 +2,14 @@
{
"name": "Youtube",
"url": "http://www.youtube.com/watch?v=BaW_jenozKc",
- "file": "BaW_jenozKc.mp4"
+ "file": "BaW_jenozKc.mp4",
+ "info_dict": {
+ "title": "youtube-dl test video \"'/\\ä↭𝕐",
+ "uploader": "Philipp Hagemeister",
+ "uploader_id": "phihag",
+ "upload_date": "20121002",
+ "description": "test chars: \"'/\\ä↭𝕐\n\nThis is a test video for youtube-dl.\n\nFor more information, contact phihag@phihag.de ."
+ }
},
{
"name": "Dailymotion",
@@ -30,9 +37,16 @@
},
{
"name": "Vimeo",
- "md5": "60540a4ec7cc378ec84b919c0aed5023",
- "url": "http://vimeo.com/14160053",
- "file": "14160053.mp4"
+ "md5": "8879b6cc097e987f02484baf890129e5",
+ "url": "http://vimeo.com/56015672",
+ "file": "56015672.mp4",
+ "info_dict": {
+ "title": "youtube-dl test video - ★ \" ' 幸 / \\ ä ↭ 𝕐",
+ "uploader": "Filippo Valsorda",
+ "uploader_id": "user7108434",
+ "upload_date": "20121220",
+ "description": "This is a test case for youtube-dl.\nFor more information, see github.com/rg3/youtube-dl\nTest chars: ★ \" ' 幸 / \\ ä ↭ 𝕐"
+ }
},
{
"name": "Soundcloud",
@@ -81,7 +95,8 @@
"name": "Escapist",
"url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
"file": "6618-Breaking-Down-Baldurs-Gate.flv",
- "md5": "c6793dbda81388f4264c1ba18684a74d"
+ "md5": "c6793dbda81388f4264c1ba18684a74d",
+ "skip": "Fails with timeout on Travis"
},
{
"name": "GooglePlus",
diff --git a/youtube_dl/FileDownloader.py b/youtube_dl/FileDownloader.py
index a861086c3..d9a4ecd3a 100644
--- a/youtube_dl/FileDownloader.py
+++ b/youtube_dl/FileDownloader.py
@@ -433,11 +433,8 @@ def process_info(self, info_dict):
try:
srtfn = filename.rsplit('.', 1)[0] + u'.srt'
self.report_writesubtitles(srtfn)
- srtfile = open(encodeFilename(srtfn), 'wb')
- try:
- srtfile.write(info_dict['subtitles'].encode('utf-8'))
- finally:
- srtfile.close()
+ with io.open(encodeFilename(srtfn), 'w', encoding='utf-8') as srtfile:
+ srtfile.write(info_dict['subtitles'])
except (OSError, IOError):
self.trouble(u'ERROR: Cannot write subtitles file ' + descfn)
return
diff --git a/youtube_dl/InfoExtractors.py b/youtube_dl/InfoExtractors.py
index d94ebde34..697c031c5 100755
--- a/youtube_dl/InfoExtractors.py
+++ b/youtube_dl/InfoExtractors.py
@@ -32,7 +32,7 @@ class InfoExtractor(object):
id: Video identifier.
url: Final video URL.
- uploader: Nickname of the video uploader, unescaped.
+ uploader: Full name of the video uploader, unescaped.
upload_date: Video upload date (YYYYMMDD).
title: Video title, unescaped.
ext: Video filename extension.
@@ -42,6 +42,7 @@ class InfoExtractor(object):
format: The video format, defaults to ext (used for --get-format)
thumbnail: Full URL to a video thumbnail image.
description: One-line video description.
+ uploader_id: Nickname or id of the video uploader.
player_url: SWF Player URL (used for rtmpdump).
subtitles: The .srt file contents.
urlhandle: [internal] The urlHandle to be used to download the file,
@@ -219,6 +220,34 @@ def _closed_captions_xml_to_srt(self, xml_string):
srt += caption + '\n\n'
return srt
+ def _extract_subtitles(self, video_id):
+ self.report_video_subtitles_download(video_id)
+ request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
+ try:
+ srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
+ except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
+ return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
+ srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
+ srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
+ if not srt_lang_list:
+ return (u'WARNING: video has no closed captions', None)
+ if self._downloader.params.get('subtitleslang', False):
+ srt_lang = self._downloader.params.get('subtitleslang')
+ elif 'en' in srt_lang_list:
+ srt_lang = 'en'
+ else:
+ srt_lang = list(srt_lang_list.keys())[0]
+ if not srt_lang in srt_lang_list:
+ return (u'WARNING: no closed captions found in the specified language', None)
+ request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
+ try:
+ srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
+ except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
+ return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
+ if not srt_xml:
+ return (u'WARNING: unable to download video subtitles', None)
+ return (None, self._closed_captions_xml_to_srt(srt_xml))
+
def _print_formats(self, formats):
print('Available formats:')
for x in formats:
@@ -356,10 +385,18 @@ def _real_extract(self, url):
# uploader
if 'author' not in video_info:
- self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
+ self._downloader.trouble(u'ERROR: unable to extract uploader name')
return
video_uploader = compat_urllib_parse.unquote_plus(video_info['author'][0])
+ # uploader_id
+ video_uploader_id = None
+ mobj = re.search(r'', video_webpage)
+ if mobj is not None:
+ video_uploader_id = mobj.group(1)
+ else:
+ self._downloader.trouble(u'WARNING: unable to extract uploader nickname')
+
# title
if 'title' not in video_info:
self._downloader.trouble(u'ERROR: unable to extract video title')
@@ -395,35 +432,9 @@ def _real_extract(self, url):
# closed captions
video_subtitles = None
if self._downloader.params.get('writesubtitles', False):
- try:
- self.report_video_subtitles_download(video_id)
- request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
- try:
- srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
- except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
- raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
- srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
- srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
- if not srt_lang_list:
- raise Trouble(u'WARNING: video has no closed captions')
- if self._downloader.params.get('subtitleslang', False):
- srt_lang = self._downloader.params.get('subtitleslang')
- elif 'en' in srt_lang_list:
- srt_lang = 'en'
- else:
- srt_lang = srt_lang_list.keys()[0]
- if not srt_lang in srt_lang_list:
- raise Trouble(u'WARNING: no closed captions found in the specified language')
- request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
- try:
- srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
- except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
- raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
- if not srt_xml:
- raise Trouble(u'WARNING: unable to download video subtitles')
- video_subtitles = self._closed_captions_xml_to_srt(srt_xml)
- except Trouble as trouble:
- self._downloader.trouble(str(trouble))
+ (srt_error, video_subtitles) = self._extract_subtitles(video_id)
+ if srt_error:
+ self._downloader.trouble(srt_error)
if 'length_seconds' not in video_info:
self._downloader.trouble(u'WARNING: unable to extract video duration')
@@ -443,7 +454,7 @@ def _real_extract(self, url):
elif 'url_encoded_fmt_stream_map' in video_info and len(video_info['url_encoded_fmt_stream_map']) >= 1:
url_data_strs = video_info['url_encoded_fmt_stream_map'][0].split(',')
url_data = [compat_parse_qs(uds) for uds in url_data_strs]
- url_data = filter(lambda ud: 'itag' in ud and 'url' in ud, url_data)
+ url_data = [ud for ud in url_data if 'itag' in ud and 'url' in ud]
url_map = dict((ud['itag'][0], ud['url'][0] + '&signature=' + ud['sig'][0]) for ud in url_data)
format_limit = self._downloader.params.get('format_limit', None)
@@ -493,6 +504,7 @@ def _real_extract(self, url):
'id': video_id,
'url': video_real_url,
'uploader': video_uploader,
+ 'uploader_id': video_uploader_id,
'upload_date': upload_date,
'title': video_title,
'ext': video_extension,
@@ -992,8 +1004,9 @@ def _real_extract(self, url, new_video=True):
# Extract title
video_title = config["video"]["title"]
- # Extract uploader
+ # Extract uploader and uploader_id
video_uploader = config["video"]["owner"]["name"]
+ video_uploader_id = config["video"]["owner"]["url"].split('/')[-1]
# Extract video thumbnail
video_thumbnail = config["video"]["thumbnail"]
@@ -1005,9 +1018,9 @@ def _real_extract(self, url, new_video=True):
# Extract upload date
video_upload_date = None
- mobj = re.search(r'[^:]*: (.*?)( \([^\(]*\))?', webpage)
+ mobj = re.search(r'
html = html.replace('\n', ' ')
- html = re.sub('\s*<\s*br\s*/?\s*>\s*', '\n', html)
+ html = re.sub(r'\s*<\s*br\s*/?\s*>\s*', '\n', html)
+ html = re.sub(r'<\s*/\s*p\s*>\s*<\s*p[^>]*>', '\n', html)
# Strip html tags
html = re.sub('<.*?>', '', html)
# Replace html entities
@@ -465,14 +466,6 @@ def __init__(self, downloaded, expected):
self.downloaded = downloaded
self.expected = expected
-
-class Trouble(Exception):
- """Trouble helper exception
-
- This is an exception to be handled with
- FileDownloader.trouble
- """
-
class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
"""Handler for HTTP requests and responses.