mirror of
				https://gitlab.com/ytdl-org/youtube-dl.git
				synced 2025-11-04 08:57:08 -05:00 
			
		
		
		
	Merge branch 'master' of github.com:rg3/youtube-dl
This commit is contained in:
		
							
								
								
									
										19
									
								
								README.md
									
									
									
									
									
								
							
							
						
						
									
										19
									
								
								README.md
									
									
									
									
									
								
							@@ -46,15 +46,16 @@ which means you can modify it, redistribute it or use it however you like.
 | 
				
			|||||||
    -A, --auto-number        number downloaded files starting from 00000
 | 
					    -A, --auto-number        number downloaded files starting from 00000
 | 
				
			||||||
    -o, --output TEMPLATE    output filename template. Use %(title)s to get the
 | 
					    -o, --output TEMPLATE    output filename template. Use %(title)s to get the
 | 
				
			||||||
                             title, %(uploader)s for the uploader name,
 | 
					                             title, %(uploader)s for the uploader name,
 | 
				
			||||||
                             %(autonumber)s to get an automatically incremented
 | 
					                             %(uploader_id)s for the uploader nickname if
 | 
				
			||||||
                             number, %(ext)s for the filename extension,
 | 
					                             different, %(autonumber)s to get an automatically
 | 
				
			||||||
                             %(upload_date)s for the upload date (YYYYMMDD),
 | 
					                             incremented number, %(ext)s for the filename
 | 
				
			||||||
                             %(extractor)s for the provider (youtube, metacafe,
 | 
					                             extension, %(upload_date)s for the upload date
 | 
				
			||||||
                             etc), %(id)s for the video id and %% for a literal
 | 
					                             (YYYYMMDD), %(extractor)s for the provider
 | 
				
			||||||
                             percent. Use - to output to stdout. Can also be
 | 
					                             (youtube, metacafe, etc), %(id)s for the video id
 | 
				
			||||||
                             used to download to a different directory, for
 | 
					                             and %% for a literal percent. Use - to output to
 | 
				
			||||||
                             example with -o '/my/downloads/%(uploader)s/%(title
 | 
					                             stdout. Can also be used to download to a different
 | 
				
			||||||
                             )s-%(id)s.%(ext)s' .
 | 
					                             directory, for example with -o '/my/downloads/%(upl
 | 
				
			||||||
 | 
					                             oader)s/%(title)s-%(id)s.%(ext)s' .
 | 
				
			||||||
    --restrict-filenames     Restrict filenames to only ASCII characters, and
 | 
					    --restrict-filenames     Restrict filenames to only ASCII characters, and
 | 
				
			||||||
                             avoid "&" and spaces in filenames
 | 
					                             avoid "&" and spaces in filenames
 | 
				
			||||||
    -a, --batch-file FILE    file containing URLs to download ('-' for stdin)
 | 
					    -a, --batch-file FILE    file containing URLs to download ('-' for stdin)
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -6,6 +6,7 @@ import os
 | 
				
			|||||||
import json
 | 
					import json
 | 
				
			||||||
import unittest
 | 
					import unittest
 | 
				
			||||||
import sys
 | 
					import sys
 | 
				
			||||||
 | 
					import hashlib
 | 
				
			||||||
import socket
 | 
					import socket
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# Allow direct execution
 | 
					# Allow direct execution
 | 
				
			||||||
@@ -24,12 +25,15 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
 | 
				
			|||||||
proxy_handler = compat_urllib_request.ProxyHandler()
 | 
					proxy_handler = compat_urllib_request.ProxyHandler()
 | 
				
			||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
					opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
				
			||||||
compat_urllib_request.install_opener(opener)
 | 
					compat_urllib_request.install_opener(opener)
 | 
				
			||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
class FileDownloader(youtube_dl.FileDownloader):
 | 
					class FileDownloader(youtube_dl.FileDownloader):
 | 
				
			||||||
    def __init__(self, *args, **kwargs):
 | 
					    def __init__(self, *args, **kwargs):
 | 
				
			||||||
        youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
 | 
					 | 
				
			||||||
        self.to_stderr = self.to_screen
 | 
					        self.to_stderr = self.to_screen
 | 
				
			||||||
 | 
					        self.processed_info_dicts = []
 | 
				
			||||||
 | 
					        return youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
 | 
				
			||||||
 | 
					    def process_info(self, info_dict):
 | 
				
			||||||
 | 
					        self.processed_info_dicts.append(info_dict)
 | 
				
			||||||
 | 
					        return youtube_dl.FileDownloader.process_info(self, info_dict)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def _file_md5(fn):
 | 
					def _file_md5(fn):
 | 
				
			||||||
    with open(fn, 'rb') as f:
 | 
					    with open(fn, 'rb') as f:
 | 
				
			||||||
@@ -40,6 +44,7 @@ with io.open(DEF_FILE, encoding='utf-8') as deff:
 | 
				
			|||||||
with io.open(PARAMETERS_FILE, encoding='utf-8') as pf:
 | 
					with io.open(PARAMETERS_FILE, encoding='utf-8') as pf:
 | 
				
			||||||
    parameters = json.load(pf)
 | 
					    parameters = json.load(pf)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
class TestDownload(unittest.TestCase):
 | 
					class TestDownload(unittest.TestCase):
 | 
				
			||||||
    def setUp(self):
 | 
					    def setUp(self):
 | 
				
			||||||
        self.parameters = parameters
 | 
					        self.parameters = parameters
 | 
				
			||||||
@@ -68,18 +73,28 @@ def generator(test_case):
 | 
				
			|||||||
        if 'skip' in test_case:
 | 
					        if 'skip' in test_case:
 | 
				
			||||||
            print('Skipping: {0}'.format(test_case['skip']))
 | 
					            print('Skipping: {0}'.format(test_case['skip']))
 | 
				
			||||||
            return
 | 
					            return
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        params = dict(self.parameters) # Duplicate it locally
 | 
					        params = dict(self.parameters) # Duplicate it locally
 | 
				
			||||||
        for p in test_case.get('params', {}):
 | 
					        for p in test_case.get('params', {}):
 | 
				
			||||||
            params[p] = test_case['params'][p]
 | 
					            params[p] = test_case['params'][p]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        fd = FileDownloader(params)
 | 
					        fd = FileDownloader(params)
 | 
				
			||||||
        fd.add_info_extractor(ie())
 | 
					        fd.add_info_extractor(ie())
 | 
				
			||||||
        for ien in test_case.get('add_ie', []):
 | 
					        for ien in test_case.get('add_ie', []):
 | 
				
			||||||
            fd.add_info_extractor(getattr(youtube_dl.InfoExtractors, ien + 'IE')())
 | 
					            fd.add_info_extractor(getattr(youtube_dl.InfoExtractors, ien + 'IE')())
 | 
				
			||||||
        fd.download([test_case['url']])
 | 
					        fd.download([test_case['url']])
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        self.assertTrue(os.path.exists(test_case['file']))
 | 
					        self.assertTrue(os.path.exists(test_case['file']))
 | 
				
			||||||
        if 'md5' in test_case:
 | 
					        if 'md5' in test_case:
 | 
				
			||||||
            md5_for_file = _file_md5(test_case['file'])
 | 
					            md5_for_file = _file_md5(test_case['file'])
 | 
				
			||||||
            self.assertEqual(md5_for_file, test_case['md5'])
 | 
					            self.assertEqual(md5_for_file, test_case['md5'])
 | 
				
			||||||
 | 
					        info_dict = fd.processed_info_dicts[0]
 | 
				
			||||||
 | 
					        for (info_field, value) in test_case.get('info_dict', {}).items():
 | 
				
			||||||
 | 
					            if value.startswith('md5:'):
 | 
				
			||||||
 | 
					                md5_info_value = hashlib.md5(info_dict.get(info_field, '')).hexdigest()
 | 
				
			||||||
 | 
					                self.assertEqual(value[3:], md5_info_value)
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                self.assertEqual(value, info_dict.get(info_field))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    return test_template
 | 
					    return test_template
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -3,7 +3,6 @@
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
import json
 | 
					import json
 | 
				
			||||||
import os
 | 
					import os
 | 
				
			||||||
import socket
 | 
					 | 
				
			||||||
import sys
 | 
					import sys
 | 
				
			||||||
import unittest
 | 
					import unittest
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -22,7 +21,6 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
 | 
				
			|||||||
proxy_handler = compat_urllib_request.ProxyHandler()
 | 
					proxy_handler = compat_urllib_request.ProxyHandler()
 | 
				
			||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
					opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
				
			||||||
compat_urllib_request.install_opener(opener)
 | 
					compat_urllib_request.install_opener(opener)
 | 
				
			||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
class FileDownloader(youtube_dl.FileDownloader):
 | 
					class FileDownloader(youtube_dl.FileDownloader):
 | 
				
			||||||
    def __init__(self, *args, **kwargs):
 | 
					    def __init__(self, *args, **kwargs):
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -2,7 +2,6 @@
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
import sys
 | 
					import sys
 | 
				
			||||||
import unittest
 | 
					import unittest
 | 
				
			||||||
import socket
 | 
					 | 
				
			||||||
import json
 | 
					import json
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# Allow direct execution
 | 
					# Allow direct execution
 | 
				
			||||||
@@ -22,7 +21,6 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
 | 
				
			|||||||
proxy_handler = compat_urllib_request.ProxyHandler()
 | 
					proxy_handler = compat_urllib_request.ProxyHandler()
 | 
				
			||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
					opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
				
			||||||
compat_urllib_request.install_opener(opener)
 | 
					compat_urllib_request.install_opener(opener)
 | 
				
			||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
class FakeDownloader(object):
 | 
					class FakeDownloader(object):
 | 
				
			||||||
    def __init__(self):
 | 
					    def __init__(self):
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -2,7 +2,6 @@
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
import sys
 | 
					import sys
 | 
				
			||||||
import unittest
 | 
					import unittest
 | 
				
			||||||
import socket
 | 
					 | 
				
			||||||
import json
 | 
					import json
 | 
				
			||||||
import io
 | 
					import io
 | 
				
			||||||
import hashlib
 | 
					import hashlib
 | 
				
			||||||
@@ -24,7 +23,6 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
 | 
				
			|||||||
proxy_handler = compat_urllib_request.ProxyHandler()
 | 
					proxy_handler = compat_urllib_request.ProxyHandler()
 | 
				
			||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
					opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
 | 
				
			||||||
compat_urllib_request.install_opener(opener)
 | 
					compat_urllib_request.install_opener(opener)
 | 
				
			||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
class FakeDownloader(object):
 | 
					class FakeDownloader(object):
 | 
				
			||||||
    def __init__(self):
 | 
					    def __init__(self):
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -2,7 +2,14 @@
 | 
				
			|||||||
  {
 | 
					  {
 | 
				
			||||||
    "name": "Youtube",
 | 
					    "name": "Youtube",
 | 
				
			||||||
    "url":  "http://www.youtube.com/watch?v=BaW_jenozKc",
 | 
					    "url":  "http://www.youtube.com/watch?v=BaW_jenozKc",
 | 
				
			||||||
    "file":  "BaW_jenozKc.mp4"
 | 
					    "file":  "BaW_jenozKc.mp4",
 | 
				
			||||||
 | 
					    "info_dict": {
 | 
				
			||||||
 | 
					      "title": "youtube-dl test video \"'/\\ä↭𝕐",
 | 
				
			||||||
 | 
					      "uploader": "Philipp Hagemeister",
 | 
				
			||||||
 | 
					      "uploader_id": "phihag",
 | 
				
			||||||
 | 
					      "upload_date": "20121002",
 | 
				
			||||||
 | 
					      "description": "test chars:  \"'/\\ä↭𝕐\n\nThis is a test video for youtube-dl.\n\nFor more information, contact phihag@phihag.de ."
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
  },
 | 
					  },
 | 
				
			||||||
  {
 | 
					  {
 | 
				
			||||||
    "name": "Dailymotion",
 | 
					    "name": "Dailymotion",
 | 
				
			||||||
@@ -30,9 +37,16 @@
 | 
				
			|||||||
  },
 | 
					  },
 | 
				
			||||||
  {
 | 
					  {
 | 
				
			||||||
    "name": "Vimeo",
 | 
					    "name": "Vimeo",
 | 
				
			||||||
    "md5":  "60540a4ec7cc378ec84b919c0aed5023",
 | 
					    "md5":  "8879b6cc097e987f02484baf890129e5",
 | 
				
			||||||
    "url":  "http://vimeo.com/14160053",
 | 
					    "url":  "http://vimeo.com/56015672",
 | 
				
			||||||
    "file": "14160053.mp4"
 | 
					    "file": "56015672.mp4",
 | 
				
			||||||
 | 
					    "info_dict": {
 | 
				
			||||||
 | 
					      "title": "youtube-dl test video - ★ \" ' 幸 / \\ ä ↭ 𝕐",
 | 
				
			||||||
 | 
					      "uploader": "Filippo Valsorda",
 | 
				
			||||||
 | 
					      "uploader_id": "user7108434",
 | 
				
			||||||
 | 
					      "upload_date": "20121220",
 | 
				
			||||||
 | 
					      "description": "This is a test case for youtube-dl.\nFor more information, see github.com/rg3/youtube-dl\nTest chars: ★ \" ' 幸 / \\ ä ↭ 𝕐"
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
  },
 | 
					  },
 | 
				
			||||||
  {
 | 
					  {
 | 
				
			||||||
    "name": "Soundcloud",
 | 
					    "name": "Soundcloud",
 | 
				
			||||||
@@ -81,7 +95,8 @@
 | 
				
			|||||||
    "name": "Escapist",
 | 
					    "name": "Escapist",
 | 
				
			||||||
    "url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
 | 
					    "url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
 | 
				
			||||||
    "file": "6618-Breaking-Down-Baldurs-Gate.flv",
 | 
					    "file": "6618-Breaking-Down-Baldurs-Gate.flv",
 | 
				
			||||||
    "md5": "c6793dbda81388f4264c1ba18684a74d"
 | 
					    "md5": "c6793dbda81388f4264c1ba18684a74d",
 | 
				
			||||||
 | 
					    "skip": "Fails with timeout on Travis"
 | 
				
			||||||
  },
 | 
					  },
 | 
				
			||||||
  {
 | 
					  {
 | 
				
			||||||
    "name": "GooglePlus",
 | 
					    "name": "GooglePlus",
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -433,11 +433,8 @@ class FileDownloader(object):
 | 
				
			|||||||
            try:
 | 
					            try:
 | 
				
			||||||
                srtfn = filename.rsplit('.', 1)[0] + u'.srt'
 | 
					                srtfn = filename.rsplit('.', 1)[0] + u'.srt'
 | 
				
			||||||
                self.report_writesubtitles(srtfn)
 | 
					                self.report_writesubtitles(srtfn)
 | 
				
			||||||
                srtfile = open(encodeFilename(srtfn), 'wb')
 | 
					                with io.open(encodeFilename(srtfn), 'w', encoding='utf-8') as srtfile:
 | 
				
			||||||
                try:
 | 
					                    srtfile.write(info_dict['subtitles'])
 | 
				
			||||||
                    srtfile.write(info_dict['subtitles'].encode('utf-8'))
 | 
					 | 
				
			||||||
                finally:
 | 
					 | 
				
			||||||
                    srtfile.close()
 | 
					 | 
				
			||||||
            except (OSError, IOError):
 | 
					            except (OSError, IOError):
 | 
				
			||||||
                self.trouble(u'ERROR: Cannot write subtitles file ' + descfn)
 | 
					                self.trouble(u'ERROR: Cannot write subtitles file ' + descfn)
 | 
				
			||||||
                return
 | 
					                return
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -32,7 +32,7 @@ class InfoExtractor(object):
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
    id:             Video identifier.
 | 
					    id:             Video identifier.
 | 
				
			||||||
    url:            Final video URL.
 | 
					    url:            Final video URL.
 | 
				
			||||||
    uploader:       Nickname of the video uploader, unescaped.
 | 
					    uploader:       Full name of the video uploader, unescaped.
 | 
				
			||||||
    upload_date:    Video upload date (YYYYMMDD).
 | 
					    upload_date:    Video upload date (YYYYMMDD).
 | 
				
			||||||
    title:          Video title, unescaped.
 | 
					    title:          Video title, unescaped.
 | 
				
			||||||
    ext:            Video filename extension.
 | 
					    ext:            Video filename extension.
 | 
				
			||||||
@@ -42,6 +42,7 @@ class InfoExtractor(object):
 | 
				
			|||||||
    format:         The video format, defaults to ext (used for --get-format)
 | 
					    format:         The video format, defaults to ext (used for --get-format)
 | 
				
			||||||
    thumbnail:      Full URL to a video thumbnail image.
 | 
					    thumbnail:      Full URL to a video thumbnail image.
 | 
				
			||||||
    description:    One-line video description.
 | 
					    description:    One-line video description.
 | 
				
			||||||
 | 
					    uploader_id:    Nickname or id of the video uploader.
 | 
				
			||||||
    player_url:     SWF Player URL (used for rtmpdump).
 | 
					    player_url:     SWF Player URL (used for rtmpdump).
 | 
				
			||||||
    subtitles:      The .srt file contents.
 | 
					    subtitles:      The .srt file contents.
 | 
				
			||||||
    urlhandle:      [internal] The urlHandle to be used to download the file,
 | 
					    urlhandle:      [internal] The urlHandle to be used to download the file,
 | 
				
			||||||
@@ -219,6 +220,34 @@ class YoutubeIE(InfoExtractor):
 | 
				
			|||||||
            srt += caption + '\n\n'
 | 
					            srt += caption + '\n\n'
 | 
				
			||||||
        return srt
 | 
					        return srt
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def _extract_subtitles(self, video_id):
 | 
				
			||||||
 | 
					        self.report_video_subtitles_download(video_id)
 | 
				
			||||||
 | 
					        request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
 | 
				
			||||||
 | 
					        try:
 | 
				
			||||||
 | 
					            srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
 | 
				
			||||||
 | 
					        except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
 | 
				
			||||||
 | 
					            return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
 | 
				
			||||||
 | 
					        srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
 | 
				
			||||||
 | 
					        srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
 | 
				
			||||||
 | 
					        if not srt_lang_list:
 | 
				
			||||||
 | 
					            return (u'WARNING: video has no closed captions', None)
 | 
				
			||||||
 | 
					        if self._downloader.params.get('subtitleslang', False):
 | 
				
			||||||
 | 
					            srt_lang = self._downloader.params.get('subtitleslang')
 | 
				
			||||||
 | 
					        elif 'en' in srt_lang_list:
 | 
				
			||||||
 | 
					            srt_lang = 'en'
 | 
				
			||||||
 | 
					        else:
 | 
				
			||||||
 | 
					            srt_lang = list(srt_lang_list.keys())[0]
 | 
				
			||||||
 | 
					        if not srt_lang in srt_lang_list:
 | 
				
			||||||
 | 
					            return (u'WARNING: no closed captions found in the specified language', None)
 | 
				
			||||||
 | 
					        request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
 | 
				
			||||||
 | 
					        try:
 | 
				
			||||||
 | 
					            srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
 | 
				
			||||||
 | 
					        except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
 | 
				
			||||||
 | 
					            return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
 | 
				
			||||||
 | 
					        if not srt_xml:
 | 
				
			||||||
 | 
					            return (u'WARNING: unable to download video subtitles', None)
 | 
				
			||||||
 | 
					        return (None, self._closed_captions_xml_to_srt(srt_xml))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    def _print_formats(self, formats):
 | 
					    def _print_formats(self, formats):
 | 
				
			||||||
        print('Available formats:')
 | 
					        print('Available formats:')
 | 
				
			||||||
        for x in formats:
 | 
					        for x in formats:
 | 
				
			||||||
@@ -356,10 +385,18 @@ class YoutubeIE(InfoExtractor):
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
        # uploader
 | 
					        # uploader
 | 
				
			||||||
        if 'author' not in video_info:
 | 
					        if 'author' not in video_info:
 | 
				
			||||||
            self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
 | 
					            self._downloader.trouble(u'ERROR: unable to extract uploader name')
 | 
				
			||||||
            return
 | 
					            return
 | 
				
			||||||
        video_uploader = compat_urllib_parse.unquote_plus(video_info['author'][0])
 | 
					        video_uploader = compat_urllib_parse.unquote_plus(video_info['author'][0])
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # uploader_id
 | 
				
			||||||
 | 
					        video_uploader_id = None
 | 
				
			||||||
 | 
					        mobj = re.search(r'<link itemprop="url" href="http://www.youtube.com/user/([^"]+)">', video_webpage)
 | 
				
			||||||
 | 
					        if mobj is not None:
 | 
				
			||||||
 | 
					            video_uploader_id = mobj.group(1)
 | 
				
			||||||
 | 
					        else:
 | 
				
			||||||
 | 
					            self._downloader.trouble(u'WARNING: unable to extract uploader nickname')
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        # title
 | 
					        # title
 | 
				
			||||||
        if 'title' not in video_info:
 | 
					        if 'title' not in video_info:
 | 
				
			||||||
            self._downloader.trouble(u'ERROR: unable to extract video title')
 | 
					            self._downloader.trouble(u'ERROR: unable to extract video title')
 | 
				
			||||||
@@ -395,35 +432,9 @@ class YoutubeIE(InfoExtractor):
 | 
				
			|||||||
        # closed captions
 | 
					        # closed captions
 | 
				
			||||||
        video_subtitles = None
 | 
					        video_subtitles = None
 | 
				
			||||||
        if self._downloader.params.get('writesubtitles', False):
 | 
					        if self._downloader.params.get('writesubtitles', False):
 | 
				
			||||||
            try:
 | 
					            (srt_error, video_subtitles) = self._extract_subtitles(video_id)
 | 
				
			||||||
                self.report_video_subtitles_download(video_id)
 | 
					            if srt_error:
 | 
				
			||||||
                request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
 | 
					                self._downloader.trouble(srt_error)
 | 
				
			||||||
                try:
 | 
					 | 
				
			||||||
                    srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
 | 
					 | 
				
			||||||
                except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
 | 
					 | 
				
			||||||
                    raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
 | 
					 | 
				
			||||||
                srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
 | 
					 | 
				
			||||||
                srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
 | 
					 | 
				
			||||||
                if not srt_lang_list:
 | 
					 | 
				
			||||||
                    raise Trouble(u'WARNING: video has no closed captions')
 | 
					 | 
				
			||||||
                if self._downloader.params.get('subtitleslang', False):
 | 
					 | 
				
			||||||
                    srt_lang = self._downloader.params.get('subtitleslang')
 | 
					 | 
				
			||||||
                elif 'en' in srt_lang_list:
 | 
					 | 
				
			||||||
                    srt_lang = 'en'
 | 
					 | 
				
			||||||
                else:
 | 
					 | 
				
			||||||
                    srt_lang = srt_lang_list.keys()[0]
 | 
					 | 
				
			||||||
                if not srt_lang in srt_lang_list:
 | 
					 | 
				
			||||||
                    raise Trouble(u'WARNING: no closed captions found in the specified language')
 | 
					 | 
				
			||||||
                request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
 | 
					 | 
				
			||||||
                try:
 | 
					 | 
				
			||||||
                    srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
 | 
					 | 
				
			||||||
                except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
 | 
					 | 
				
			||||||
                    raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
 | 
					 | 
				
			||||||
                if not srt_xml:
 | 
					 | 
				
			||||||
                    raise Trouble(u'WARNING: unable to download video subtitles')
 | 
					 | 
				
			||||||
                video_subtitles = self._closed_captions_xml_to_srt(srt_xml)
 | 
					 | 
				
			||||||
            except Trouble as trouble:
 | 
					 | 
				
			||||||
                self._downloader.trouble(str(trouble))
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
        if 'length_seconds' not in video_info:
 | 
					        if 'length_seconds' not in video_info:
 | 
				
			||||||
            self._downloader.trouble(u'WARNING: unable to extract video duration')
 | 
					            self._downloader.trouble(u'WARNING: unable to extract video duration')
 | 
				
			||||||
@@ -443,7 +454,7 @@ class YoutubeIE(InfoExtractor):
 | 
				
			|||||||
        elif 'url_encoded_fmt_stream_map' in video_info and len(video_info['url_encoded_fmt_stream_map']) >= 1:
 | 
					        elif 'url_encoded_fmt_stream_map' in video_info and len(video_info['url_encoded_fmt_stream_map']) >= 1:
 | 
				
			||||||
            url_data_strs = video_info['url_encoded_fmt_stream_map'][0].split(',')
 | 
					            url_data_strs = video_info['url_encoded_fmt_stream_map'][0].split(',')
 | 
				
			||||||
            url_data = [compat_parse_qs(uds) for uds in url_data_strs]
 | 
					            url_data = [compat_parse_qs(uds) for uds in url_data_strs]
 | 
				
			||||||
            url_data = filter(lambda ud: 'itag' in ud and 'url' in ud, url_data)
 | 
					            url_data = [ud for ud in url_data if 'itag' in ud and 'url' in ud]
 | 
				
			||||||
            url_map = dict((ud['itag'][0], ud['url'][0] + '&signature=' + ud['sig'][0]) for ud in url_data)
 | 
					            url_map = dict((ud['itag'][0], ud['url'][0] + '&signature=' + ud['sig'][0]) for ud in url_data)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            format_limit = self._downloader.params.get('format_limit', None)
 | 
					            format_limit = self._downloader.params.get('format_limit', None)
 | 
				
			||||||
@@ -493,6 +504,7 @@ class YoutubeIE(InfoExtractor):
 | 
				
			|||||||
                'id':       video_id,
 | 
					                'id':       video_id,
 | 
				
			||||||
                'url':      video_real_url,
 | 
					                'url':      video_real_url,
 | 
				
			||||||
                'uploader': video_uploader,
 | 
					                'uploader': video_uploader,
 | 
				
			||||||
 | 
					                'uploader_id': video_uploader_id,
 | 
				
			||||||
                'upload_date':  upload_date,
 | 
					                'upload_date':  upload_date,
 | 
				
			||||||
                'title':    video_title,
 | 
					                'title':    video_title,
 | 
				
			||||||
                'ext':      video_extension,
 | 
					                'ext':      video_extension,
 | 
				
			||||||
@@ -992,8 +1004,9 @@ class VimeoIE(InfoExtractor):
 | 
				
			|||||||
        # Extract title
 | 
					        # Extract title
 | 
				
			||||||
        video_title = config["video"]["title"]
 | 
					        video_title = config["video"]["title"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        # Extract uploader
 | 
					        # Extract uploader and uploader_id
 | 
				
			||||||
        video_uploader = config["video"]["owner"]["name"]
 | 
					        video_uploader = config["video"]["owner"]["name"]
 | 
				
			||||||
 | 
					        video_uploader_id = config["video"]["owner"]["url"].split('/')[-1]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        # Extract video thumbnail
 | 
					        # Extract video thumbnail
 | 
				
			||||||
        video_thumbnail = config["video"]["thumbnail"]
 | 
					        video_thumbnail = config["video"]["thumbnail"]
 | 
				
			||||||
@@ -1005,9 +1018,9 @@ class VimeoIE(InfoExtractor):
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
        # Extract upload date
 | 
					        # Extract upload date
 | 
				
			||||||
        video_upload_date = None
 | 
					        video_upload_date = None
 | 
				
			||||||
        mobj = re.search(r'<span id="clip-date" style="display:none">[^:]*: (.*?)( \([^\(]*\))?</span>', webpage)
 | 
					        mobj = re.search(r'<meta itemprop="dateCreated" content="(\d{4})-(\d{2})-(\d{2})T', webpage)
 | 
				
			||||||
        if mobj is not None:
 | 
					        if mobj is not None:
 | 
				
			||||||
            video_upload_date = mobj.group(1)
 | 
					            video_upload_date = mobj.group(1) + mobj.group(2) + mobj.group(3)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        # Vimeo specific: extract request signature and timestamp
 | 
					        # Vimeo specific: extract request signature and timestamp
 | 
				
			||||||
        sig = config['request']['signature']
 | 
					        sig = config['request']['signature']
 | 
				
			||||||
@@ -1045,6 +1058,7 @@ class VimeoIE(InfoExtractor):
 | 
				
			|||||||
            'id':       video_id,
 | 
					            'id':       video_id,
 | 
				
			||||||
            'url':      video_url,
 | 
					            'url':      video_url,
 | 
				
			||||||
            'uploader': video_uploader,
 | 
					            'uploader': video_uploader,
 | 
				
			||||||
 | 
					            'uploader_id': video_uploader_id,
 | 
				
			||||||
            'upload_date':  video_upload_date,
 | 
					            'upload_date':  video_upload_date,
 | 
				
			||||||
            'title':    video_title,
 | 
					            'title':    video_title,
 | 
				
			||||||
            'ext':      video_extension,
 | 
					            'ext':      video_extension,
 | 
				
			||||||
@@ -2113,7 +2127,7 @@ class FacebookIE(InfoExtractor):
 | 
				
			|||||||
        video_description = video_info.get('description', 'No description available.')
 | 
					        video_description = video_info.get('description', 'No description available.')
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        url_map = video_info['video_urls']
 | 
					        url_map = video_info['video_urls']
 | 
				
			||||||
        if len(url_map.keys()) > 0:
 | 
					        if len(list(url_map.keys())) > 0:
 | 
				
			||||||
            # Decide which formats to download
 | 
					            # Decide which formats to download
 | 
				
			||||||
            req_format = self._downloader.params.get('format', None)
 | 
					            req_format = self._downloader.params.get('format', None)
 | 
				
			||||||
            format_limit = self._downloader.params.get('format_limit', None)
 | 
					            format_limit = self._downloader.params.get('format_limit', None)
 | 
				
			||||||
@@ -2973,7 +2987,7 @@ class MixcloudIE(InfoExtractor):
 | 
				
			|||||||
                if file_url is not None:
 | 
					                if file_url is not None:
 | 
				
			||||||
                    break # got it!
 | 
					                    break # got it!
 | 
				
			||||||
        else:
 | 
					        else:
 | 
				
			||||||
            if req_format not in formats.keys():
 | 
					            if req_format not in list(formats.keys()):
 | 
				
			||||||
                self._downloader.trouble(u'ERROR: format is not available')
 | 
					                self._downloader.trouble(u'ERROR: format is not available')
 | 
				
			||||||
                return
 | 
					                return
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -3272,7 +3286,7 @@ class YoukuIE(InfoExtractor):
 | 
				
			|||||||
            seed = config['data'][0]['seed']
 | 
					            seed = config['data'][0]['seed']
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            format = self._downloader.params.get('format', None)
 | 
					            format = self._downloader.params.get('format', None)
 | 
				
			||||||
            supported_format = config['data'][0]['streamfileids'].keys()
 | 
					            supported_format = list(config['data'][0]['streamfileids'].keys())
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            if format is None or format == 'best':
 | 
					            if format is None or format == 'best':
 | 
				
			||||||
                if 'hd2' in supported_format:
 | 
					                if 'hd2' in supported_format:
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -307,7 +307,7 @@ def parseOpts():
 | 
				
			|||||||
            action='store_true', dest='autonumber',
 | 
					            action='store_true', dest='autonumber',
 | 
				
			||||||
            help='number downloaded files starting from 00000', default=False)
 | 
					            help='number downloaded files starting from 00000', default=False)
 | 
				
			||||||
    filesystem.add_option('-o', '--output',
 | 
					    filesystem.add_option('-o', '--output',
 | 
				
			||||||
            dest='outtmpl', metavar='TEMPLATE', help='output filename template. Use %(title)s to get the title, %(uploader)s for the uploader name, %(autonumber)s to get an automatically incremented number, %(ext)s for the filename extension, %(upload_date)s for the upload date (YYYYMMDD), %(extractor)s for the provider (youtube, metacafe, etc), %(id)s for the video id and %% for a literal percent. Use - to output to stdout. Can also be used to download to a different directory, for example with -o \'/my/downloads/%(uploader)s/%(title)s-%(id)s.%(ext)s\' .')
 | 
					            dest='outtmpl', metavar='TEMPLATE', help='output filename template. Use %(title)s to get the title, %(uploader)s for the uploader name, %(uploader_id)s for the uploader nickname if different, %(autonumber)s to get an automatically incremented number, %(ext)s for the filename extension, %(upload_date)s for the upload date (YYYYMMDD), %(extractor)s for the provider (youtube, metacafe, etc), %(id)s for the video id and %% for a literal percent. Use - to output to stdout. Can also be used to download to a different directory, for example with -o \'/my/downloads/%(uploader)s/%(title)s-%(id)s.%(ext)s\' .')
 | 
				
			||||||
    filesystem.add_option('--restrict-filenames',
 | 
					    filesystem.add_option('--restrict-filenames',
 | 
				
			||||||
            action='store_true', dest='restrictfilenames',
 | 
					            action='store_true', dest='restrictfilenames',
 | 
				
			||||||
            help='Restrict filenames to only ASCII characters, and avoid "&" and spaces in filenames', default=False)
 | 
					            help='Restrict filenames to only ASCII characters, and avoid "&" and spaces in filenames', default=False)
 | 
				
			||||||
@@ -454,8 +454,8 @@ def _real_main():
 | 
				
			|||||||
    if opts.list_extractors:
 | 
					    if opts.list_extractors:
 | 
				
			||||||
        for ie in extractors:
 | 
					        for ie in extractors:
 | 
				
			||||||
            print(ie.IE_NAME + (' (CURRENTLY BROKEN)' if not ie._WORKING else ''))
 | 
					            print(ie.IE_NAME + (' (CURRENTLY BROKEN)' if not ie._WORKING else ''))
 | 
				
			||||||
            matchedUrls = filter(lambda url: ie.suitable(url), all_urls)
 | 
					            matchedUrls = [url for url in all_urls if ie.suitable(url)]
 | 
				
			||||||
            all_urls = filter(lambda url: url not in matchedUrls, all_urls)
 | 
					            all_urls = [url for url in all_urls if url not in matchedUrls]
 | 
				
			||||||
            for mu in matchedUrls:
 | 
					            for mu in matchedUrls:
 | 
				
			||||||
                print(u'  ' + mu)
 | 
					                print(u'  ' + mu)
 | 
				
			||||||
        sys.exit(0)
 | 
					        sys.exit(0)
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -298,7 +298,8 @@ def clean_html(html):
 | 
				
			|||||||
    """Clean an HTML snippet into a readable string"""
 | 
					    """Clean an HTML snippet into a readable string"""
 | 
				
			||||||
    # Newline vs <br />
 | 
					    # Newline vs <br />
 | 
				
			||||||
    html = html.replace('\n', ' ')
 | 
					    html = html.replace('\n', ' ')
 | 
				
			||||||
    html = re.sub('\s*<\s*br\s*/?\s*>\s*', '\n', html)
 | 
					    html = re.sub(r'\s*<\s*br\s*/?\s*>\s*', '\n', html)
 | 
				
			||||||
 | 
					    html = re.sub(r'<\s*/\s*p\s*>\s*<\s*p[^>]*>', '\n', html)
 | 
				
			||||||
    # Strip html tags
 | 
					    # Strip html tags
 | 
				
			||||||
    html = re.sub('<.*?>', '', html)
 | 
					    html = re.sub('<.*?>', '', html)
 | 
				
			||||||
    # Replace html entities
 | 
					    # Replace html entities
 | 
				
			||||||
@@ -465,14 +466,6 @@ class ContentTooShortError(Exception):
 | 
				
			|||||||
        self.downloaded = downloaded
 | 
					        self.downloaded = downloaded
 | 
				
			||||||
        self.expected = expected
 | 
					        self.expected = expected
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					 | 
				
			||||||
class Trouble(Exception):
 | 
					 | 
				
			||||||
    """Trouble helper exception
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    This is an exception to be handled with
 | 
					 | 
				
			||||||
    FileDownloader.trouble
 | 
					 | 
				
			||||||
    """
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
 | 
					class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
 | 
				
			||||||
    """Handler for HTTP requests and responses.
 | 
					    """Handler for HTTP requests and responses.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user