Merge branch 'master' of github.com:rg3/youtube-dl
This commit is contained in:
commit
a7c0f8602e
10 changed files with 106 additions and 77 deletions
19
README.md
19
README.md
|
@ -46,15 +46,16 @@ which means you can modify it, redistribute it or use it however you like.
|
|||
-A, --auto-number number downloaded files starting from 00000
|
||||
-o, --output TEMPLATE output filename template. Use %(title)s to get the
|
||||
title, %(uploader)s for the uploader name,
|
||||
%(autonumber)s to get an automatically incremented
|
||||
number, %(ext)s for the filename extension,
|
||||
%(upload_date)s for the upload date (YYYYMMDD),
|
||||
%(extractor)s for the provider (youtube, metacafe,
|
||||
etc), %(id)s for the video id and %% for a literal
|
||||
percent. Use - to output to stdout. Can also be
|
||||
used to download to a different directory, for
|
||||
example with -o '/my/downloads/%(uploader)s/%(title
|
||||
)s-%(id)s.%(ext)s' .
|
||||
%(uploader_id)s for the uploader nickname if
|
||||
different, %(autonumber)s to get an automatically
|
||||
incremented number, %(ext)s for the filename
|
||||
extension, %(upload_date)s for the upload date
|
||||
(YYYYMMDD), %(extractor)s for the provider
|
||||
(youtube, metacafe, etc), %(id)s for the video id
|
||||
and %% for a literal percent. Use - to output to
|
||||
stdout. Can also be used to download to a different
|
||||
directory, for example with -o '/my/downloads/%(upl
|
||||
oader)s/%(title)s-%(id)s.%(ext)s' .
|
||||
--restrict-filenames Restrict filenames to only ASCII characters, and
|
||||
avoid "&" and spaces in filenames
|
||||
-a, --batch-file FILE file containing URLs to download ('-' for stdin)
|
||||
|
|
|
@ -6,6 +6,7 @@ import os
|
|||
import json
|
||||
import unittest
|
||||
import sys
|
||||
import hashlib
|
||||
import socket
|
||||
|
||||
# Allow direct execution
|
||||
|
@ -24,12 +25,15 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
|
|||
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||
compat_urllib_request.install_opener(opener)
|
||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
||||
|
||||
class FileDownloader(youtube_dl.FileDownloader):
|
||||
def __init__(self, *args, **kwargs):
|
||||
youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
|
||||
self.to_stderr = self.to_screen
|
||||
self.processed_info_dicts = []
|
||||
return youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
|
||||
def process_info(self, info_dict):
|
||||
self.processed_info_dicts.append(info_dict)
|
||||
return youtube_dl.FileDownloader.process_info(self, info_dict)
|
||||
|
||||
def _file_md5(fn):
|
||||
with open(fn, 'rb') as f:
|
||||
|
@ -40,6 +44,7 @@ with io.open(DEF_FILE, encoding='utf-8') as deff:
|
|||
with io.open(PARAMETERS_FILE, encoding='utf-8') as pf:
|
||||
parameters = json.load(pf)
|
||||
|
||||
|
||||
class TestDownload(unittest.TestCase):
|
||||
def setUp(self):
|
||||
self.parameters = parameters
|
||||
|
@ -68,18 +73,28 @@ def generator(test_case):
|
|||
if 'skip' in test_case:
|
||||
print('Skipping: {0}'.format(test_case['skip']))
|
||||
return
|
||||
|
||||
params = dict(self.parameters) # Duplicate it locally
|
||||
for p in test_case.get('params', {}):
|
||||
params[p] = test_case['params'][p]
|
||||
|
||||
fd = FileDownloader(params)
|
||||
fd.add_info_extractor(ie())
|
||||
for ien in test_case.get('add_ie', []):
|
||||
fd.add_info_extractor(getattr(youtube_dl.InfoExtractors, ien + 'IE')())
|
||||
fd.download([test_case['url']])
|
||||
|
||||
self.assertTrue(os.path.exists(test_case['file']))
|
||||
if 'md5' in test_case:
|
||||
md5_for_file = _file_md5(test_case['file'])
|
||||
self.assertEqual(md5_for_file, test_case['md5'])
|
||||
info_dict = fd.processed_info_dicts[0]
|
||||
for (info_field, value) in test_case.get('info_dict', {}).items():
|
||||
if value.startswith('md5:'):
|
||||
md5_info_value = hashlib.md5(info_dict.get(info_field, '')).hexdigest()
|
||||
self.assertEqual(value[3:], md5_info_value)
|
||||
else:
|
||||
self.assertEqual(value, info_dict.get(info_field))
|
||||
|
||||
return test_template
|
||||
|
||||
|
|
|
@ -3,7 +3,6 @@
|
|||
|
||||
import json
|
||||
import os
|
||||
import socket
|
||||
import sys
|
||||
import unittest
|
||||
|
||||
|
@ -22,7 +21,6 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
|
|||
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||
compat_urllib_request.install_opener(opener)
|
||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
||||
|
||||
class FileDownloader(youtube_dl.FileDownloader):
|
||||
def __init__(self, *args, **kwargs):
|
||||
|
|
|
@ -2,7 +2,6 @@
|
|||
|
||||
import sys
|
||||
import unittest
|
||||
import socket
|
||||
import json
|
||||
|
||||
# Allow direct execution
|
||||
|
@ -22,7 +21,6 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
|
|||
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||
compat_urllib_request.install_opener(opener)
|
||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
||||
|
||||
class FakeDownloader(object):
|
||||
def __init__(self):
|
||||
|
|
|
@ -2,7 +2,6 @@
|
|||
|
||||
import sys
|
||||
import unittest
|
||||
import socket
|
||||
import json
|
||||
import io
|
||||
import hashlib
|
||||
|
@ -24,7 +23,6 @@ cookie_processor = compat_urllib_request.HTTPCookieProcessor(jar)
|
|||
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||
compat_urllib_request.install_opener(opener)
|
||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
||||
|
||||
class FakeDownloader(object):
|
||||
def __init__(self):
|
||||
|
|
|
@ -2,7 +2,14 @@
|
|||
{
|
||||
"name": "Youtube",
|
||||
"url": "http://www.youtube.com/watch?v=BaW_jenozKc",
|
||||
"file": "BaW_jenozKc.mp4"
|
||||
"file": "BaW_jenozKc.mp4",
|
||||
"info_dict": {
|
||||
"title": "youtube-dl test video \"'/\\ä↭𝕐",
|
||||
"uploader": "Philipp Hagemeister",
|
||||
"uploader_id": "phihag",
|
||||
"upload_date": "20121002",
|
||||
"description": "test chars: \"'/\\ä↭𝕐\n\nThis is a test video for youtube-dl.\n\nFor more information, contact phihag@phihag.de ."
|
||||
}
|
||||
},
|
||||
{
|
||||
"name": "Dailymotion",
|
||||
|
@ -30,9 +37,16 @@
|
|||
},
|
||||
{
|
||||
"name": "Vimeo",
|
||||
"md5": "60540a4ec7cc378ec84b919c0aed5023",
|
||||
"url": "http://vimeo.com/14160053",
|
||||
"file": "14160053.mp4"
|
||||
"md5": "8879b6cc097e987f02484baf890129e5",
|
||||
"url": "http://vimeo.com/56015672",
|
||||
"file": "56015672.mp4",
|
||||
"info_dict": {
|
||||
"title": "youtube-dl test video - ★ \" ' 幸 / \\ ä ↭ 𝕐",
|
||||
"uploader": "Filippo Valsorda",
|
||||
"uploader_id": "user7108434",
|
||||
"upload_date": "20121220",
|
||||
"description": "This is a test case for youtube-dl.\nFor more information, see github.com/rg3/youtube-dl\nTest chars: ★ \" ' 幸 / \\ ä ↭ 𝕐"
|
||||
}
|
||||
},
|
||||
{
|
||||
"name": "Soundcloud",
|
||||
|
@ -81,7 +95,8 @@
|
|||
"name": "Escapist",
|
||||
"url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
|
||||
"file": "6618-Breaking-Down-Baldurs-Gate.flv",
|
||||
"md5": "c6793dbda81388f4264c1ba18684a74d"
|
||||
"md5": "c6793dbda81388f4264c1ba18684a74d",
|
||||
"skip": "Fails with timeout on Travis"
|
||||
},
|
||||
{
|
||||
"name": "GooglePlus",
|
||||
|
|
|
@ -433,11 +433,8 @@ class FileDownloader(object):
|
|||
try:
|
||||
srtfn = filename.rsplit('.', 1)[0] + u'.srt'
|
||||
self.report_writesubtitles(srtfn)
|
||||
srtfile = open(encodeFilename(srtfn), 'wb')
|
||||
try:
|
||||
srtfile.write(info_dict['subtitles'].encode('utf-8'))
|
||||
finally:
|
||||
srtfile.close()
|
||||
with io.open(encodeFilename(srtfn), 'w', encoding='utf-8') as srtfile:
|
||||
srtfile.write(info_dict['subtitles'])
|
||||
except (OSError, IOError):
|
||||
self.trouble(u'ERROR: Cannot write subtitles file ' + descfn)
|
||||
return
|
||||
|
|
|
@ -32,7 +32,7 @@ class InfoExtractor(object):
|
|||
|
||||
id: Video identifier.
|
||||
url: Final video URL.
|
||||
uploader: Nickname of the video uploader, unescaped.
|
||||
uploader: Full name of the video uploader, unescaped.
|
||||
upload_date: Video upload date (YYYYMMDD).
|
||||
title: Video title, unescaped.
|
||||
ext: Video filename extension.
|
||||
|
@ -42,6 +42,7 @@ class InfoExtractor(object):
|
|||
format: The video format, defaults to ext (used for --get-format)
|
||||
thumbnail: Full URL to a video thumbnail image.
|
||||
description: One-line video description.
|
||||
uploader_id: Nickname or id of the video uploader.
|
||||
player_url: SWF Player URL (used for rtmpdump).
|
||||
subtitles: The .srt file contents.
|
||||
urlhandle: [internal] The urlHandle to be used to download the file,
|
||||
|
@ -219,6 +220,34 @@ class YoutubeIE(InfoExtractor):
|
|||
srt += caption + '\n\n'
|
||||
return srt
|
||||
|
||||
def _extract_subtitles(self, video_id):
|
||||
self.report_video_subtitles_download(video_id)
|
||||
request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
|
||||
try:
|
||||
srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||
return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
|
||||
srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
|
||||
srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
|
||||
if not srt_lang_list:
|
||||
return (u'WARNING: video has no closed captions', None)
|
||||
if self._downloader.params.get('subtitleslang', False):
|
||||
srt_lang = self._downloader.params.get('subtitleslang')
|
||||
elif 'en' in srt_lang_list:
|
||||
srt_lang = 'en'
|
||||
else:
|
||||
srt_lang = list(srt_lang_list.keys())[0]
|
||||
if not srt_lang in srt_lang_list:
|
||||
return (u'WARNING: no closed captions found in the specified language', None)
|
||||
request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
|
||||
try:
|
||||
srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||
return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
|
||||
if not srt_xml:
|
||||
return (u'WARNING: unable to download video subtitles', None)
|
||||
return (None, self._closed_captions_xml_to_srt(srt_xml))
|
||||
|
||||
def _print_formats(self, formats):
|
||||
print('Available formats:')
|
||||
for x in formats:
|
||||
|
@ -356,10 +385,18 @@ class YoutubeIE(InfoExtractor):
|
|||
|
||||
# uploader
|
||||
if 'author' not in video_info:
|
||||
self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
|
||||
self._downloader.trouble(u'ERROR: unable to extract uploader name')
|
||||
return
|
||||
video_uploader = compat_urllib_parse.unquote_plus(video_info['author'][0])
|
||||
|
||||
# uploader_id
|
||||
video_uploader_id = None
|
||||
mobj = re.search(r'<link itemprop="url" href="http://www.youtube.com/user/([^"]+)">', video_webpage)
|
||||
if mobj is not None:
|
||||
video_uploader_id = mobj.group(1)
|
||||
else:
|
||||
self._downloader.trouble(u'WARNING: unable to extract uploader nickname')
|
||||
|
||||
# title
|
||||
if 'title' not in video_info:
|
||||
self._downloader.trouble(u'ERROR: unable to extract video title')
|
||||
|
@ -395,35 +432,9 @@ class YoutubeIE(InfoExtractor):
|
|||
# closed captions
|
||||
video_subtitles = None
|
||||
if self._downloader.params.get('writesubtitles', False):
|
||||
try:
|
||||
self.report_video_subtitles_download(video_id)
|
||||
request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
|
||||
try:
|
||||
srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
||||
srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
|
||||
srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
|
||||
if not srt_lang_list:
|
||||
raise Trouble(u'WARNING: video has no closed captions')
|
||||
if self._downloader.params.get('subtitleslang', False):
|
||||
srt_lang = self._downloader.params.get('subtitleslang')
|
||||
elif 'en' in srt_lang_list:
|
||||
srt_lang = 'en'
|
||||
else:
|
||||
srt_lang = srt_lang_list.keys()[0]
|
||||
if not srt_lang in srt_lang_list:
|
||||
raise Trouble(u'WARNING: no closed captions found in the specified language')
|
||||
request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
|
||||
try:
|
||||
srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
||||
if not srt_xml:
|
||||
raise Trouble(u'WARNING: unable to download video subtitles')
|
||||
video_subtitles = self._closed_captions_xml_to_srt(srt_xml)
|
||||
except Trouble as trouble:
|
||||
self._downloader.trouble(str(trouble))
|
||||
(srt_error, video_subtitles) = self._extract_subtitles(video_id)
|
||||
if srt_error:
|
||||
self._downloader.trouble(srt_error)
|
||||
|
||||
if 'length_seconds' not in video_info:
|
||||
self._downloader.trouble(u'WARNING: unable to extract video duration')
|
||||
|
@ -443,7 +454,7 @@ class YoutubeIE(InfoExtractor):
|
|||
elif 'url_encoded_fmt_stream_map' in video_info and len(video_info['url_encoded_fmt_stream_map']) >= 1:
|
||||
url_data_strs = video_info['url_encoded_fmt_stream_map'][0].split(',')
|
||||
url_data = [compat_parse_qs(uds) for uds in url_data_strs]
|
||||
url_data = filter(lambda ud: 'itag' in ud and 'url' in ud, url_data)
|
||||
url_data = [ud for ud in url_data if 'itag' in ud and 'url' in ud]
|
||||
url_map = dict((ud['itag'][0], ud['url'][0] + '&signature=' + ud['sig'][0]) for ud in url_data)
|
||||
|
||||
format_limit = self._downloader.params.get('format_limit', None)
|
||||
|
@ -493,6 +504,7 @@ class YoutubeIE(InfoExtractor):
|
|||
'id': video_id,
|
||||
'url': video_real_url,
|
||||
'uploader': video_uploader,
|
||||
'uploader_id': video_uploader_id,
|
||||
'upload_date': upload_date,
|
||||
'title': video_title,
|
||||
'ext': video_extension,
|
||||
|
@ -992,8 +1004,9 @@ class VimeoIE(InfoExtractor):
|
|||
# Extract title
|
||||
video_title = config["video"]["title"]
|
||||
|
||||
# Extract uploader
|
||||
# Extract uploader and uploader_id
|
||||
video_uploader = config["video"]["owner"]["name"]
|
||||
video_uploader_id = config["video"]["owner"]["url"].split('/')[-1]
|
||||
|
||||
# Extract video thumbnail
|
||||
video_thumbnail = config["video"]["thumbnail"]
|
||||
|
@ -1005,9 +1018,9 @@ class VimeoIE(InfoExtractor):
|
|||
|
||||
# Extract upload date
|
||||
video_upload_date = None
|
||||
mobj = re.search(r'<span id="clip-date" style="display:none">[^:]*: (.*?)( \([^\(]*\))?</span>', webpage)
|
||||
mobj = re.search(r'<meta itemprop="dateCreated" content="(\d{4})-(\d{2})-(\d{2})T', webpage)
|
||||
if mobj is not None:
|
||||
video_upload_date = mobj.group(1)
|
||||
video_upload_date = mobj.group(1) + mobj.group(2) + mobj.group(3)
|
||||
|
||||
# Vimeo specific: extract request signature and timestamp
|
||||
sig = config['request']['signature']
|
||||
|
@ -1045,6 +1058,7 @@ class VimeoIE(InfoExtractor):
|
|||
'id': video_id,
|
||||
'url': video_url,
|
||||
'uploader': video_uploader,
|
||||
'uploader_id': video_uploader_id,
|
||||
'upload_date': video_upload_date,
|
||||
'title': video_title,
|
||||
'ext': video_extension,
|
||||
|
@ -2113,7 +2127,7 @@ class FacebookIE(InfoExtractor):
|
|||
video_description = video_info.get('description', 'No description available.')
|
||||
|
||||
url_map = video_info['video_urls']
|
||||
if len(url_map.keys()) > 0:
|
||||
if len(list(url_map.keys())) > 0:
|
||||
# Decide which formats to download
|
||||
req_format = self._downloader.params.get('format', None)
|
||||
format_limit = self._downloader.params.get('format_limit', None)
|
||||
|
@ -2973,7 +2987,7 @@ class MixcloudIE(InfoExtractor):
|
|||
if file_url is not None:
|
||||
break # got it!
|
||||
else:
|
||||
if req_format not in formats.keys():
|
||||
if req_format not in list(formats.keys()):
|
||||
self._downloader.trouble(u'ERROR: format is not available')
|
||||
return
|
||||
|
||||
|
@ -3272,7 +3286,7 @@ class YoukuIE(InfoExtractor):
|
|||
seed = config['data'][0]['seed']
|
||||
|
||||
format = self._downloader.params.get('format', None)
|
||||
supported_format = config['data'][0]['streamfileids'].keys()
|
||||
supported_format = list(config['data'][0]['streamfileids'].keys())
|
||||
|
||||
if format is None or format == 'best':
|
||||
if 'hd2' in supported_format:
|
||||
|
|
|
@ -307,7 +307,7 @@ def parseOpts():
|
|||
action='store_true', dest='autonumber',
|
||||
help='number downloaded files starting from 00000', default=False)
|
||||
filesystem.add_option('-o', '--output',
|
||||
dest='outtmpl', metavar='TEMPLATE', help='output filename template. Use %(title)s to get the title, %(uploader)s for the uploader name, %(autonumber)s to get an automatically incremented number, %(ext)s for the filename extension, %(upload_date)s for the upload date (YYYYMMDD), %(extractor)s for the provider (youtube, metacafe, etc), %(id)s for the video id and %% for a literal percent. Use - to output to stdout. Can also be used to download to a different directory, for example with -o \'/my/downloads/%(uploader)s/%(title)s-%(id)s.%(ext)s\' .')
|
||||
dest='outtmpl', metavar='TEMPLATE', help='output filename template. Use %(title)s to get the title, %(uploader)s for the uploader name, %(uploader_id)s for the uploader nickname if different, %(autonumber)s to get an automatically incremented number, %(ext)s for the filename extension, %(upload_date)s for the upload date (YYYYMMDD), %(extractor)s for the provider (youtube, metacafe, etc), %(id)s for the video id and %% for a literal percent. Use - to output to stdout. Can also be used to download to a different directory, for example with -o \'/my/downloads/%(uploader)s/%(title)s-%(id)s.%(ext)s\' .')
|
||||
filesystem.add_option('--restrict-filenames',
|
||||
action='store_true', dest='restrictfilenames',
|
||||
help='Restrict filenames to only ASCII characters, and avoid "&" and spaces in filenames', default=False)
|
||||
|
@ -454,8 +454,8 @@ def _real_main():
|
|||
if opts.list_extractors:
|
||||
for ie in extractors:
|
||||
print(ie.IE_NAME + (' (CURRENTLY BROKEN)' if not ie._WORKING else ''))
|
||||
matchedUrls = filter(lambda url: ie.suitable(url), all_urls)
|
||||
all_urls = filter(lambda url: url not in matchedUrls, all_urls)
|
||||
matchedUrls = [url for url in all_urls if ie.suitable(url)]
|
||||
all_urls = [url for url in all_urls if url not in matchedUrls]
|
||||
for mu in matchedUrls:
|
||||
print(u' ' + mu)
|
||||
sys.exit(0)
|
||||
|
|
|
@ -298,7 +298,8 @@ def clean_html(html):
|
|||
"""Clean an HTML snippet into a readable string"""
|
||||
# Newline vs <br />
|
||||
html = html.replace('\n', ' ')
|
||||
html = re.sub('\s*<\s*br\s*/?\s*>\s*', '\n', html)
|
||||
html = re.sub(r'\s*<\s*br\s*/?\s*>\s*', '\n', html)
|
||||
html = re.sub(r'<\s*/\s*p\s*>\s*<\s*p[^>]*>', '\n', html)
|
||||
# Strip html tags
|
||||
html = re.sub('<.*?>', '', html)
|
||||
# Replace html entities
|
||||
|
@ -465,14 +466,6 @@ class ContentTooShortError(Exception):
|
|||
self.downloaded = downloaded
|
||||
self.expected = expected
|
||||
|
||||
|
||||
class Trouble(Exception):
|
||||
"""Trouble helper exception
|
||||
|
||||
This is an exception to be handled with
|
||||
FileDownloader.trouble
|
||||
"""
|
||||
|
||||
class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
|
||||
"""Handler for HTTP requests and responses.
|
||||
|
||||
|
|
Loading…
Reference in a new issue