yt-dlp/youtube_dl/extractor/qqmusic.py

241 lines
8.0 KiB
Python
Raw Normal View History

2015-03-17 07:03:29 -04:00
# coding: utf-8
from __future__ import unicode_literals
2015-03-18 01:56:02 -04:00
import random
import time
2015-03-18 02:59:33 -04:00
import re
2015-03-18 01:56:02 -04:00
2015-03-17 07:03:29 -04:00
from .common import InfoExtractor
2015-03-18 13:47:07 -04:00
from ..utils import (
strip_jsonp,
unescapeHTML,
js_to_json,
2015-03-18 13:47:07 -04:00
)
2015-03-18 02:59:33 -04:00
from ..compat import compat_urllib_request
2015-03-17 07:03:29 -04:00
class QQMusicIE(InfoExtractor):
IE_NAME = 'qqmusic'
2015-03-17 07:03:29 -04:00
_VALID_URL = r'http://y.qq.com/#type=song&mid=(?P<id>[0-9A-Za-z]+)'
_TESTS = [{
'url': 'http://y.qq.com/#type=song&mid=004295Et37taLD',
'md5': 'bed90b6db2a7a7a7e11bc585f471f63a',
'info_dict': {
'id': '004295Et37taLD',
'ext': 'm4a',
'title': '可惜没如果',
'upload_date': '20141227',
'creator': '林俊杰',
'description': 'md5:d327722d0361576fde558f1ac68a7065',
2015-03-17 07:03:29 -04:00
}
}]
2015-03-18 01:56:02 -04:00
# Reference: m_r_GetRUin() in top_player.js
# http://imgcache.gtimg.cn/music/portal_v3/y/top_player.js
@staticmethod
def m_r_get_ruin():
curMs = int(time.time() * 1000) % 1000
return int(round(random.random() * 2147483647) * curMs % 1E10)
2015-03-17 07:03:29 -04:00
def _real_extract(self, url):
mid = self._match_id(url)
detail_info_page = self._download_webpage(
'http://s.plcloud.music.qq.com/fcgi-bin/fcg_yqq_song_detail_info.fcg?songmid=%s&play=0' % mid,
2015-03-18 02:59:33 -04:00
mid, note='Download song detail info',
errnote='Unable to get song detail info', encoding='gbk')
2015-03-17 07:03:29 -04:00
song_name = self._html_search_regex(
r"songname:\s*'([^']+)'", detail_info_page, 'song name')
publish_time = self._html_search_regex(
r'发行时间:(\d{4}-\d{2}-\d{2})', detail_info_page,
'publish time', default=None)
if publish_time:
publish_time = publish_time.replace('-', '')
2015-03-17 07:03:29 -04:00
singer = self._html_search_regex(
r"singer:\s*'([^']+)", detail_info_page, 'singer', default=None)
lrc_content = self._html_search_regex(
r'<div class="content" id="lrc_content"[^<>]*>([^<>]+)</div>',
detail_info_page, 'LRC lyrics', default=None)
if lrc_content:
lrc_content = lrc_content.replace('\\n', '\n')
2015-03-17 07:03:29 -04:00
2015-03-18 01:56:02 -04:00
guid = self.m_r_get_ruin()
2015-03-17 07:03:29 -04:00
vkey = self._download_json(
'http://base.music.qq.com/fcgi-bin/fcg_musicexpress.fcg?json=3&guid=%s' % guid,
mid, note='Retrieve vkey', errnote='Unable to get vkey',
transform_source=strip_jsonp)['key']
song_url = 'http://cc.stream.qqmusic.qq.com/C200%s.m4a?vkey=%s&guid=%s&fromtag=0' % (mid, vkey, guid)
return {
'id': mid,
'url': song_url,
'title': song_name,
'upload_date': publish_time,
'creator': singer,
'description': lrc_content,
2015-03-17 07:03:29 -04:00
}
2015-03-18 02:59:33 -04:00
2015-03-18 13:47:07 -04:00
class QQPlaylistBaseIE(InfoExtractor):
@staticmethod
def qq_static_url(category, mid):
return 'http://y.qq.com/y/static/%s/%s/%s/%s.html' % (category, mid[-2], mid[-1], mid)
@classmethod
def get_entries_from_page(cls, page):
entries = []
for item in re.findall(r'class="data"[^<>]*>([^<>]+)</', page):
song_mid = unescapeHTML(item).split('|')[-5]
entries.append(cls.url_result(
'http://y.qq.com/#type=song&mid=' + song_mid, 'QQMusic',
song_mid))
2015-03-18 13:47:07 -04:00
return entries
class QQMusicSingerIE(QQPlaylistBaseIE):
IE_NAME = 'qqmusic:singer'
2015-03-18 02:59:33 -04:00
_VALID_URL = r'http://y.qq.com/#type=singer&mid=(?P<id>[0-9A-Za-z]+)'
_TEST = {
'url': 'http://y.qq.com/#type=singer&mid=001BLpXF2DyJe2',
'info_dict': {
'id': '001BLpXF2DyJe2',
'title': '林俊杰',
'description': 'md5:2a222d89ba4455a3af19940c0481bb78',
},
'playlist_count': 12,
}
def _real_extract(self, url):
mid = self._match_id(url)
singer_page = self._download_webpage(
2015-03-18 13:47:07 -04:00
self.qq_static_url('singer', mid), mid, 'Download singer page')
2015-03-18 02:59:33 -04:00
2015-03-18 13:47:07 -04:00
entries = self.get_entries_from_page(singer_page)
2015-03-18 02:59:33 -04:00
singer_name = self._html_search_regex(
r"singername\s*:\s*'([^']+)'", singer_page, 'singer name',
default=None)
singer_id = self._html_search_regex(
r"singerid\s*:\s*'([0-9]+)'", singer_page, 'singer id',
default=None)
singer_desc = None
if singer_id:
req = compat_urllib_request.Request(
'http://s.plcloud.music.qq.com/fcgi-bin/fcg_get_singer_desc.fcg?utf8=1&outCharset=utf-8&format=xml&singerid=%s' % singer_id)
req.add_header(
'Referer', 'http://s.plcloud.music.qq.com/xhr_proxy_utf8.html')
singer_desc_page = self._download_xml(
2015-03-18 13:47:07 -04:00
req, mid, 'Donwload singer description XML')
2015-03-18 02:59:33 -04:00
singer_desc = singer_desc_page.find('./data/info/desc').text
return self.playlist_result(entries, mid, singer_name, singer_desc)
2015-03-18 13:47:07 -04:00
class QQMusicAlbumIE(QQPlaylistBaseIE):
IE_NAME = 'qqmusic:album'
2015-03-18 13:47:07 -04:00
_VALID_URL = r'http://y.qq.com/#type=album&mid=(?P<id>[0-9A-Za-z]+)'
_TEST = {
'url': 'http://y.qq.com/#type=album&mid=000gXCTb2AhRR1&play=0',
'info_dict': {
'id': '000gXCTb2AhRR1',
'title': '我们都是这样长大的',
'description': 'md5:d216c55a2d4b3537fe4415b8767d74d6',
},
'playlist_count': 4,
}
def _real_extract(self, url):
mid = self._match_id(url)
album_page = self._download_webpage(
self.qq_static_url('album', mid), mid, 'Download album page')
entries = self.get_entries_from_page(album_page)
album_name = self._html_search_regex(
r"albumname\s*:\s*'([^']+)',", album_page, 'album name',
default=None)
album_detail = self._html_search_regex(
r'<div class="album_detail close_detail">\s*<p>((?:[^<>]+(?:<br />)?)+)</p>',
album_page, 'album details', default=None)
return self.playlist_result(entries, mid, album_name, album_detail)
class QQMusicToplistIE(QQPlaylistBaseIE):
IE_NAME = 'qqmusic:toplist'
_VALID_URL = r'http://y\.qq\.com/#type=toplist&p=(?P<id>(top|global)_[0-9]+)'
2015-05-14 11:25:43 -04:00
_TESTS = [{
'url': 'http://y.qq.com/#type=toplist&p=global_12',
'info_dict': {
'id': 'global_12',
'title': 'itunes榜',
},
'playlist_count': 10,
}, {
'url': 'http://y.qq.com/#type=toplist&p=top_6',
'info_dict': {
'id': 'top_6',
'title': 'QQ音乐巅峰榜·欧美',
},
'playlist_count': 100,
}, {
'url': 'http://y.qq.com/#type=toplist&p=global_5',
'info_dict': {
'id': 'global_5',
'title': '韩国mnet排行榜',
},
'playlist_count': 50,
}]
@staticmethod
def strip_qq_jsonp(code):
return js_to_json(re.sub(r'^MusicJsonCallback\((.*?)\)/\*.+?\*/$', r'\1', code))
2015-05-14 11:25:43 -04:00
def _real_extract(self, url):
list_id = self._match_id(url)
2015-05-14 11:28:42 -04:00
list_type, num_id = list_id.split("_")
2015-05-12 10:41:37 -04:00
list_page = self._download_webpage(
2015-05-14 11:25:43 -04:00
"http://y.qq.com/y/static/toplist/index/%s.html" % list_id,
2015-05-12 10:41:37 -04:00
list_id, 'Download toplist page')
entries = []
if list_type == 'top':
jsonp_url = "http://y.qq.com/y/static/toplist/json/top/%s/1.js" % num_id
2015-05-12 13:37:56 -04:00
else:
jsonp_url = "http://y.qq.com/y/static/toplist/json/global/%s/1_1.js" % num_id
2015-05-14 11:25:43 -04:00
2015-05-14 11:28:42 -04:00
toplist_json = self._download_json(
2015-05-14 11:25:43 -04:00
jsonp_url, list_id, note='Retrieve toplist json',
errnote='Unable to get toplist json', transform_source=self.strip_qq_jsonp)
2015-05-14 11:25:43 -04:00
2015-05-14 11:28:42 -04:00
for song in toplist_json['l']:
s = song['s']
song_mid = s.split("|")[20]
entries.append(self.url_result(
'http://y.qq.com/#type=song&mid=' + song_mid, 'QQMusic',
song_mid))
list_name = self._html_search_regex(
r'<h2 id="top_name">([^\']+)</h2>', list_page, 'top list name',
default=None)
2015-05-14 11:28:42 -04:00
return self.playlist_result(entries, list_id, list_name)