mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-09 11:47:56 -05:00
265 lines
9.7 KiB
Python
265 lines
9.7 KiB
Python
|
import base64
|
||
|
import os
|
||
|
import re
|
||
|
import socket
|
||
|
import sys
|
||
|
|
||
|
from ..utils import (
|
||
|
compat_http_client,
|
||
|
compat_urllib_error,
|
||
|
compat_urllib_request,
|
||
|
compat_str,
|
||
|
|
||
|
clean_html,
|
||
|
compiled_regex_type,
|
||
|
ExtractorError,
|
||
|
)
|
||
|
|
||
|
class InfoExtractor(object):
|
||
|
"""Information Extractor class.
|
||
|
|
||
|
Information extractors are the classes that, given a URL, extract
|
||
|
information about the video (or videos) the URL refers to. This
|
||
|
information includes the real video URL, the video title, author and
|
||
|
others. The information is stored in a dictionary which is then
|
||
|
passed to the FileDownloader. The FileDownloader processes this
|
||
|
information possibly downloading the video to the file system, among
|
||
|
other possible outcomes.
|
||
|
|
||
|
The dictionaries must include the following fields:
|
||
|
|
||
|
id: Video identifier.
|
||
|
url: Final video URL.
|
||
|
title: Video title, unescaped.
|
||
|
ext: Video filename extension.
|
||
|
|
||
|
The following fields are optional:
|
||
|
|
||
|
format: The video format, defaults to ext (used for --get-format)
|
||
|
thumbnail: Full URL to a video thumbnail image.
|
||
|
description: One-line video description.
|
||
|
uploader: Full name of the video uploader.
|
||
|
upload_date: Video upload date (YYYYMMDD).
|
||
|
uploader_id: Nickname or id of the video uploader.
|
||
|
location: Physical location of the video.
|
||
|
player_url: SWF Player URL (used for rtmpdump).
|
||
|
subtitles: The subtitle file contents.
|
||
|
urlhandle: [internal] The urlHandle to be used to download the file,
|
||
|
like returned by urllib.request.urlopen
|
||
|
|
||
|
The fields should all be Unicode strings.
|
||
|
|
||
|
Subclasses of this one should re-define the _real_initialize() and
|
||
|
_real_extract() methods and define a _VALID_URL regexp.
|
||
|
Probably, they should also be added to the list of extractors.
|
||
|
|
||
|
_real_extract() must return a *list* of information dictionaries as
|
||
|
described above.
|
||
|
|
||
|
Finally, the _WORKING attribute should be set to False for broken IEs
|
||
|
in order to warn the users and skip the tests.
|
||
|
"""
|
||
|
|
||
|
_ready = False
|
||
|
_downloader = None
|
||
|
_WORKING = True
|
||
|
|
||
|
def __init__(self, downloader=None):
|
||
|
"""Constructor. Receives an optional downloader."""
|
||
|
self._ready = False
|
||
|
self.set_downloader(downloader)
|
||
|
|
||
|
@classmethod
|
||
|
def suitable(cls, url):
|
||
|
"""Receives a URL and returns True if suitable for this IE."""
|
||
|
return re.match(cls._VALID_URL, url) is not None
|
||
|
|
||
|
@classmethod
|
||
|
def working(cls):
|
||
|
"""Getter method for _WORKING."""
|
||
|
return cls._WORKING
|
||
|
|
||
|
def initialize(self):
|
||
|
"""Initializes an instance (authentication, etc)."""
|
||
|
if not self._ready:
|
||
|
self._real_initialize()
|
||
|
self._ready = True
|
||
|
|
||
|
def extract(self, url):
|
||
|
"""Extracts URL information and returns it in list of dicts."""
|
||
|
self.initialize()
|
||
|
return self._real_extract(url)
|
||
|
|
||
|
def set_downloader(self, downloader):
|
||
|
"""Sets the downloader for this IE."""
|
||
|
self._downloader = downloader
|
||
|
|
||
|
def _real_initialize(self):
|
||
|
"""Real initialization process. Redefine in subclasses."""
|
||
|
pass
|
||
|
|
||
|
def _real_extract(self, url):
|
||
|
"""Real extraction process. Redefine in subclasses."""
|
||
|
pass
|
||
|
|
||
|
@property
|
||
|
def IE_NAME(self):
|
||
|
return type(self).__name__[:-2]
|
||
|
|
||
|
def _request_webpage(self, url_or_request, video_id, note=None, errnote=None):
|
||
|
""" Returns the response handle """
|
||
|
if note is None:
|
||
|
self.report_download_webpage(video_id)
|
||
|
elif note is not False:
|
||
|
self.to_screen(u'%s: %s' % (video_id, note))
|
||
|
try:
|
||
|
return compat_urllib_request.urlopen(url_or_request)
|
||
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||
|
if errnote is None:
|
||
|
errnote = u'Unable to download webpage'
|
||
|
raise ExtractorError(u'%s: %s' % (errnote, compat_str(err)), sys.exc_info()[2])
|
||
|
|
||
|
def _download_webpage_handle(self, url_or_request, video_id, note=None, errnote=None):
|
||
|
""" Returns a tuple (page content as string, URL handle) """
|
||
|
urlh = self._request_webpage(url_or_request, video_id, note, errnote)
|
||
|
content_type = urlh.headers.get('Content-Type', '')
|
||
|
m = re.match(r'[a-zA-Z0-9_.-]+/[a-zA-Z0-9_.-]+\s*;\s*charset=(.+)', content_type)
|
||
|
if m:
|
||
|
encoding = m.group(1)
|
||
|
else:
|
||
|
encoding = 'utf-8'
|
||
|
webpage_bytes = urlh.read()
|
||
|
if self._downloader.params.get('dump_intermediate_pages', False):
|
||
|
try:
|
||
|
url = url_or_request.get_full_url()
|
||
|
except AttributeError:
|
||
|
url = url_or_request
|
||
|
self.to_screen(u'Dumping request to ' + url)
|
||
|
dump = base64.b64encode(webpage_bytes).decode('ascii')
|
||
|
self._downloader.to_screen(dump)
|
||
|
content = webpage_bytes.decode(encoding, 'replace')
|
||
|
return (content, urlh)
|
||
|
|
||
|
def _download_webpage(self, url_or_request, video_id, note=None, errnote=None):
|
||
|
""" Returns the data of the page as a string """
|
||
|
return self._download_webpage_handle(url_or_request, video_id, note, errnote)[0]
|
||
|
|
||
|
def to_screen(self, msg):
|
||
|
"""Print msg to screen, prefixing it with '[ie_name]'"""
|
||
|
self._downloader.to_screen(u'[%s] %s' % (self.IE_NAME, msg))
|
||
|
|
||
|
def report_extraction(self, id_or_name):
|
||
|
"""Report information extraction."""
|
||
|
self.to_screen(u'%s: Extracting information' % id_or_name)
|
||
|
|
||
|
def report_download_webpage(self, video_id):
|
||
|
"""Report webpage download."""
|
||
|
self.to_screen(u'%s: Downloading webpage' % video_id)
|
||
|
|
||
|
def report_age_confirmation(self):
|
||
|
"""Report attempt to confirm age."""
|
||
|
self.to_screen(u'Confirming age')
|
||
|
|
||
|
#Methods for following #608
|
||
|
#They set the correct value of the '_type' key
|
||
|
def video_result(self, video_info):
|
||
|
"""Returns a video"""
|
||
|
video_info['_type'] = 'video'
|
||
|
return video_info
|
||
|
def url_result(self, url, ie=None):
|
||
|
"""Returns a url that points to a page that should be processed"""
|
||
|
#TODO: ie should be the class used for getting the info
|
||
|
video_info = {'_type': 'url',
|
||
|
'url': url,
|
||
|
'ie_key': ie}
|
||
|
return video_info
|
||
|
def playlist_result(self, entries, playlist_id=None, playlist_title=None):
|
||
|
"""Returns a playlist"""
|
||
|
video_info = {'_type': 'playlist',
|
||
|
'entries': entries}
|
||
|
if playlist_id:
|
||
|
video_info['id'] = playlist_id
|
||
|
if playlist_title:
|
||
|
video_info['title'] = playlist_title
|
||
|
return video_info
|
||
|
|
||
|
def _search_regex(self, pattern, string, name, default=None, fatal=True, flags=0):
|
||
|
"""
|
||
|
Perform a regex search on the given string, using a single or a list of
|
||
|
patterns returning the first matching group.
|
||
|
In case of failure return a default value or raise a WARNING or a
|
||
|
ExtractorError, depending on fatal, specifying the field name.
|
||
|
"""
|
||
|
if isinstance(pattern, (str, compat_str, compiled_regex_type)):
|
||
|
mobj = re.search(pattern, string, flags)
|
||
|
else:
|
||
|
for p in pattern:
|
||
|
mobj = re.search(p, string, flags)
|
||
|
if mobj: break
|
||
|
|
||
|
if sys.stderr.isatty() and os.name != 'nt':
|
||
|
_name = u'\033[0;34m%s\033[0m' % name
|
||
|
else:
|
||
|
_name = name
|
||
|
|
||
|
if mobj:
|
||
|
# return the first matching group
|
||
|
return next(g for g in mobj.groups() if g is not None)
|
||
|
elif default is not None:
|
||
|
return default
|
||
|
elif fatal:
|
||
|
raise ExtractorError(u'Unable to extract %s' % _name)
|
||
|
else:
|
||
|
self._downloader.report_warning(u'unable to extract %s; '
|
||
|
u'please report this issue on GitHub.' % _name)
|
||
|
return None
|
||
|
|
||
|
def _html_search_regex(self, pattern, string, name, default=None, fatal=True, flags=0):
|
||
|
"""
|
||
|
Like _search_regex, but strips HTML tags and unescapes entities.
|
||
|
"""
|
||
|
res = self._search_regex(pattern, string, name, default, fatal, flags)
|
||
|
if res:
|
||
|
return clean_html(res).strip()
|
||
|
else:
|
||
|
return res
|
||
|
|
||
|
class SearchInfoExtractor(InfoExtractor):
|
||
|
"""
|
||
|
Base class for paged search queries extractors.
|
||
|
They accept urls in the format _SEARCH_KEY(|all|[0-9]):{query}
|
||
|
Instances should define _SEARCH_KEY and _MAX_RESULTS.
|
||
|
"""
|
||
|
|
||
|
@classmethod
|
||
|
def _make_valid_url(cls):
|
||
|
return r'%s(?P<prefix>|[1-9][0-9]*|all):(?P<query>[\s\S]+)' % cls._SEARCH_KEY
|
||
|
|
||
|
@classmethod
|
||
|
def suitable(cls, url):
|
||
|
return re.match(cls._make_valid_url(), url) is not None
|
||
|
|
||
|
def _real_extract(self, query):
|
||
|
mobj = re.match(self._make_valid_url(), query)
|
||
|
if mobj is None:
|
||
|
raise ExtractorError(u'Invalid search query "%s"' % query)
|
||
|
|
||
|
prefix = mobj.group('prefix')
|
||
|
query = mobj.group('query')
|
||
|
if prefix == '':
|
||
|
return self._get_n_results(query, 1)
|
||
|
elif prefix == 'all':
|
||
|
return self._get_n_results(query, self._MAX_RESULTS)
|
||
|
else:
|
||
|
n = int(prefix)
|
||
|
if n <= 0:
|
||
|
raise ExtractorError(u'invalid download number %s for query "%s"' % (n, query))
|
||
|
elif n > self._MAX_RESULTS:
|
||
|
self._downloader.report_warning(u'%s returns max %i results (you requested %i)' % (self._SEARCH_KEY, self._MAX_RESULTS, n))
|
||
|
n = self._MAX_RESULTS
|
||
|
return self._get_n_results(query, n)
|
||
|
|
||
|
def _get_n_results(self, query, n):
|
||
|
"""Get a specified number of results for a query"""
|
||
|
raise NotImplementedError("This method must be implemented by sublclasses")
|