You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
417 lines
15 KiB
417 lines
15 KiB
from __future__ import unicode_literals |
|
|
|
import random |
|
import re |
|
import time |
|
|
|
from .common import InfoExtractor |
|
from ..compat import ( |
|
compat_str, |
|
compat_urlparse, |
|
) |
|
from ..utils import ( |
|
ExtractorError, |
|
float_or_none, |
|
int_or_none, |
|
KNOWN_EXTENSIONS, |
|
parse_filesize, |
|
str_or_none, |
|
try_get, |
|
unescapeHTML, |
|
update_url_query, |
|
unified_strdate, |
|
unified_timestamp, |
|
url_or_none, |
|
) |
|
|
|
|
|
class BandcampIE(InfoExtractor): |
|
_VALID_URL = r'https?://[^/]+\.bandcamp\.com/track/(?P<title>[^/?#&]+)' |
|
_TESTS = [{ |
|
'url': 'http://youtube-dl.bandcamp.com/track/youtube-dl-test-song', |
|
'md5': 'c557841d5e50261777a6585648adf439', |
|
'info_dict': { |
|
'id': '1812978515', |
|
'ext': 'mp3', |
|
'title': "youtube-dl \"'/\\\u00e4\u21ad - youtube-dl test song \"'/\\\u00e4\u21ad", |
|
'duration': 9.8485, |
|
}, |
|
'_skip': 'There is a limit of 200 free downloads / month for the test song' |
|
}, { |
|
# free download |
|
'url': 'http://benprunty.bandcamp.com/track/lanius-battle', |
|
'md5': '853e35bf34aa1d6fe2615ae612564b36', |
|
'info_dict': { |
|
'id': '2650410135', |
|
'ext': 'aiff', |
|
'title': 'Ben Prunty - Lanius (Battle)', |
|
'thumbnail': r're:^https?://.*\.jpg$', |
|
'uploader': 'Ben Prunty', |
|
'timestamp': 1396508491, |
|
'upload_date': '20140403', |
|
'release_date': '20140403', |
|
'duration': 260.877, |
|
'track': 'Lanius (Battle)', |
|
'track_number': 1, |
|
'track_id': '2650410135', |
|
'artist': 'Ben Prunty', |
|
'album': 'FTL: Advanced Edition Soundtrack', |
|
}, |
|
}, { |
|
# no free download, mp3 128 |
|
'url': 'https://relapsealumni.bandcamp.com/track/hail-to-fire', |
|
'md5': 'fec12ff55e804bb7f7ebeb77a800c8b7', |
|
'info_dict': { |
|
'id': '2584466013', |
|
'ext': 'mp3', |
|
'title': 'Mastodon - Hail to Fire', |
|
'thumbnail': r're:^https?://.*\.jpg$', |
|
'uploader': 'Mastodon', |
|
'timestamp': 1322005399, |
|
'upload_date': '20111122', |
|
'release_date': '20040207', |
|
'duration': 120.79, |
|
'track': 'Hail to Fire', |
|
'track_number': 5, |
|
'track_id': '2584466013', |
|
'artist': 'Mastodon', |
|
'album': 'Call of the Mastodon', |
|
}, |
|
}] |
|
|
|
def _real_extract(self, url): |
|
mobj = re.match(self._VALID_URL, url) |
|
title = mobj.group('title') |
|
webpage = self._download_webpage(url, title) |
|
thumbnail = self._html_search_meta('og:image', webpage, default=None) |
|
|
|
track_id = None |
|
track = None |
|
track_number = None |
|
duration = None |
|
|
|
formats = [] |
|
track_info = self._parse_json( |
|
self._search_regex( |
|
r'trackinfo\s*:\s*\[\s*({.+?})\s*\]\s*,\s*?\n', |
|
webpage, 'track info', default='{}'), title) |
|
if track_info: |
|
file_ = track_info.get('file') |
|
if isinstance(file_, dict): |
|
for format_id, format_url in file_.items(): |
|
if not url_or_none(format_url): |
|
continue |
|
ext, abr_str = format_id.split('-', 1) |
|
formats.append({ |
|
'format_id': format_id, |
|
'url': self._proto_relative_url(format_url, 'http:'), |
|
'ext': ext, |
|
'vcodec': 'none', |
|
'acodec': ext, |
|
'abr': int_or_none(abr_str), |
|
}) |
|
track = track_info.get('title') |
|
track_id = str_or_none(track_info.get('track_id') or track_info.get('id')) |
|
track_number = int_or_none(track_info.get('track_num')) |
|
duration = float_or_none(track_info.get('duration')) |
|
|
|
def extract(key): |
|
return self._search_regex( |
|
r'\b%s\s*["\']?\s*:\s*(["\'])(?P<value>(?:(?!\1).)+)\1' % key, |
|
webpage, key, default=None, group='value') |
|
|
|
artist = extract('artist') |
|
album = extract('album_title') |
|
timestamp = unified_timestamp( |
|
extract('publish_date') or extract('album_publish_date')) |
|
release_date = unified_strdate(extract('album_release_date')) |
|
|
|
download_link = self._search_regex( |
|
r'freeDownloadPage\s*:\s*(["\'])(?P<url>(?:(?!\1).)+)\1', webpage, |
|
'download link', default=None, group='url') |
|
if download_link: |
|
track_id = self._search_regex( |
|
r'(?ms)var TralbumData = .*?[{,]\s*id: (?P<id>\d+),?$', |
|
webpage, 'track id') |
|
|
|
download_webpage = self._download_webpage( |
|
download_link, track_id, 'Downloading free downloads page') |
|
|
|
blob = self._parse_json( |
|
self._search_regex( |
|
r'data-blob=(["\'])(?P<blob>{.+?})\1', download_webpage, |
|
'blob', group='blob'), |
|
track_id, transform_source=unescapeHTML) |
|
|
|
info = try_get( |
|
blob, (lambda x: x['digital_items'][0], |
|
lambda x: x['download_items'][0]), dict) |
|
if info: |
|
downloads = info.get('downloads') |
|
if isinstance(downloads, dict): |
|
if not track: |
|
track = info.get('title') |
|
if not artist: |
|
artist = info.get('artist') |
|
if not thumbnail: |
|
thumbnail = info.get('thumb_url') |
|
|
|
download_formats = {} |
|
download_formats_list = blob.get('download_formats') |
|
if isinstance(download_formats_list, list): |
|
for f in blob['download_formats']: |
|
name, ext = f.get('name'), f.get('file_extension') |
|
if all(isinstance(x, compat_str) for x in (name, ext)): |
|
download_formats[name] = ext.strip('.') |
|
|
|
for format_id, f in downloads.items(): |
|
format_url = f.get('url') |
|
if not format_url: |
|
continue |
|
# Stat URL generation algorithm is reverse engineered from |
|
# download_*_bundle_*.js |
|
stat_url = update_url_query( |
|
format_url.replace('/download/', '/statdownload/'), { |
|
'.rand': int(time.time() * 1000 * random.random()), |
|
}) |
|
format_id = f.get('encoding_name') or format_id |
|
stat = self._download_json( |
|
stat_url, track_id, 'Downloading %s JSON' % format_id, |
|
transform_source=lambda s: s[s.index('{'):s.rindex('}') + 1], |
|
fatal=False) |
|
if not stat: |
|
continue |
|
retry_url = url_or_none(stat.get('retry_url')) |
|
if not retry_url: |
|
continue |
|
formats.append({ |
|
'url': self._proto_relative_url(retry_url, 'http:'), |
|
'ext': download_formats.get(format_id), |
|
'format_id': format_id, |
|
'format_note': f.get('description'), |
|
'filesize': parse_filesize(f.get('size_mb')), |
|
'vcodec': 'none', |
|
}) |
|
|
|
self._sort_formats(formats) |
|
|
|
title = '%s - %s' % (artist, track) if artist else track |
|
|
|
if not duration: |
|
duration = float_or_none(self._html_search_meta( |
|
'duration', webpage, default=None)) |
|
|
|
return { |
|
'id': track_id, |
|
'title': title, |
|
'thumbnail': thumbnail, |
|
'uploader': artist, |
|
'timestamp': timestamp, |
|
'release_date': release_date, |
|
'duration': duration, |
|
'track': track, |
|
'track_number': track_number, |
|
'track_id': track_id, |
|
'artist': artist, |
|
'album': album, |
|
'formats': formats, |
|
} |
|
|
|
|
|
class BandcampAlbumIE(InfoExtractor): |
|
IE_NAME = 'Bandcamp:album' |
|
_VALID_URL = r'https?://(?:(?P<subdomain>[^.]+)\.)?bandcamp\.com(?:/album/(?P<album_id>[^/?#&]+))?' |
|
|
|
_TESTS = [{ |
|
'url': 'http://blazo.bandcamp.com/album/jazz-format-mixtape-vol-1', |
|
'playlist': [ |
|
{ |
|
'md5': '39bc1eded3476e927c724321ddf116cf', |
|
'info_dict': { |
|
'id': '1353101989', |
|
'ext': 'mp3', |
|
'title': 'Intro', |
|
} |
|
}, |
|
{ |
|
'md5': '1a2c32e2691474643e912cc6cd4bffaa', |
|
'info_dict': { |
|
'id': '38097443', |
|
'ext': 'mp3', |
|
'title': 'Kero One - Keep It Alive (Blazo remix)', |
|
} |
|
}, |
|
], |
|
'info_dict': { |
|
'title': 'Jazz Format Mixtape vol.1', |
|
'id': 'jazz-format-mixtape-vol-1', |
|
'uploader_id': 'blazo', |
|
}, |
|
'params': { |
|
'playlistend': 2 |
|
}, |
|
'skip': 'Bandcamp imposes download limits.' |
|
}, { |
|
'url': 'http://nightbringer.bandcamp.com/album/hierophany-of-the-open-grave', |
|
'info_dict': { |
|
'title': 'Hierophany of the Open Grave', |
|
'uploader_id': 'nightbringer', |
|
'id': 'hierophany-of-the-open-grave', |
|
}, |
|
'playlist_mincount': 9, |
|
}, { |
|
'url': 'http://dotscale.bandcamp.com', |
|
'info_dict': { |
|
'title': 'Loom', |
|
'id': 'dotscale', |
|
'uploader_id': 'dotscale', |
|
}, |
|
'playlist_mincount': 7, |
|
}, { |
|
# with escaped quote in title |
|
'url': 'https://jstrecords.bandcamp.com/album/entropy-ep', |
|
'info_dict': { |
|
'title': '"Entropy" EP', |
|
'uploader_id': 'jstrecords', |
|
'id': 'entropy-ep', |
|
}, |
|
'playlist_mincount': 3, |
|
}, { |
|
# not all tracks have songs |
|
'url': 'https://insulters.bandcamp.com/album/we-are-the-plague', |
|
'info_dict': { |
|
'id': 'we-are-the-plague', |
|
'title': 'WE ARE THE PLAGUE', |
|
'uploader_id': 'insulters', |
|
}, |
|
'playlist_count': 2, |
|
}] |
|
|
|
@classmethod |
|
def suitable(cls, url): |
|
return (False |
|
if BandcampWeeklyIE.suitable(url) or BandcampIE.suitable(url) |
|
else super(BandcampAlbumIE, cls).suitable(url)) |
|
|
|
def _real_extract(self, url): |
|
mobj = re.match(self._VALID_URL, url) |
|
uploader_id = mobj.group('subdomain') |
|
album_id = mobj.group('album_id') |
|
playlist_id = album_id or uploader_id |
|
webpage = self._download_webpage(url, playlist_id) |
|
track_elements = re.findall( |
|
r'(?s)<div[^>]*>(.*?<a[^>]+href="([^"]+?)"[^>]+itemprop="url"[^>]*>.*?)</div>', webpage) |
|
if not track_elements: |
|
raise ExtractorError('The page doesn\'t contain any tracks') |
|
# Only tracks with duration info have songs |
|
entries = [ |
|
self.url_result( |
|
compat_urlparse.urljoin(url, t_path), |
|
ie=BandcampIE.ie_key(), |
|
video_title=self._search_regex( |
|
r'<span\b[^>]+\bitemprop=["\']name["\'][^>]*>([^<]+)', |
|
elem_content, 'track title', fatal=False)) |
|
for elem_content, t_path in track_elements |
|
if self._html_search_meta('duration', elem_content, default=None)] |
|
|
|
title = self._html_search_regex( |
|
r'album_title\s*:\s*"((?:\\.|[^"\\])+?)"', |
|
webpage, 'title', fatal=False) |
|
if title: |
|
title = title.replace(r'\"', '"') |
|
return { |
|
'_type': 'playlist', |
|
'uploader_id': uploader_id, |
|
'id': playlist_id, |
|
'title': title, |
|
'entries': entries, |
|
} |
|
|
|
|
|
class BandcampWeeklyIE(InfoExtractor): |
|
IE_NAME = 'Bandcamp:weekly' |
|
_VALID_URL = r'https?://(?:www\.)?bandcamp\.com/?\?(?:.*?&)?show=(?P<id>\d+)' |
|
_TESTS = [{ |
|
'url': 'https://bandcamp.com/?show=224', |
|
'md5': 'b00df799c733cf7e0c567ed187dea0fd', |
|
'info_dict': { |
|
'id': '224', |
|
'ext': 'opus', |
|
'title': 'BC Weekly April 4th 2017 - Magic Moments', |
|
'description': 'md5:5d48150916e8e02d030623a48512c874', |
|
'duration': 5829.77, |
|
'release_date': '20170404', |
|
'series': 'Bandcamp Weekly', |
|
'episode': 'Magic Moments', |
|
'episode_number': 208, |
|
'episode_id': '224', |
|
} |
|
}, { |
|
'url': 'https://bandcamp.com/?blah/blah@&show=228', |
|
'only_matching': True |
|
}] |
|
|
|
def _real_extract(self, url): |
|
video_id = self._match_id(url) |
|
webpage = self._download_webpage(url, video_id) |
|
|
|
blob = self._parse_json( |
|
self._search_regex( |
|
r'data-blob=(["\'])(?P<blob>{.+?})\1', webpage, |
|
'blob', group='blob'), |
|
video_id, transform_source=unescapeHTML) |
|
|
|
show = blob['bcw_show'] |
|
|
|
# This is desired because any invalid show id redirects to `bandcamp.com` |
|
# which happens to expose the latest Bandcamp Weekly episode. |
|
show_id = int_or_none(show.get('show_id')) or int_or_none(video_id) |
|
|
|
formats = [] |
|
for format_id, format_url in show['audio_stream'].items(): |
|
if not url_or_none(format_url): |
|
continue |
|
for known_ext in KNOWN_EXTENSIONS: |
|
if known_ext in format_id: |
|
ext = known_ext |
|
break |
|
else: |
|
ext = None |
|
formats.append({ |
|
'format_id': format_id, |
|
'url': format_url, |
|
'ext': ext, |
|
'vcodec': 'none', |
|
}) |
|
self._sort_formats(formats) |
|
|
|
title = show.get('audio_title') or 'Bandcamp Weekly' |
|
subtitle = show.get('subtitle') |
|
if subtitle: |
|
title += ' - %s' % subtitle |
|
|
|
episode_number = None |
|
seq = blob.get('bcw_seq') |
|
|
|
if seq and isinstance(seq, list): |
|
try: |
|
episode_number = next( |
|
int_or_none(e.get('episode_number')) |
|
for e in seq |
|
if isinstance(e, dict) and int_or_none(e.get('id')) == show_id) |
|
except StopIteration: |
|
pass |
|
|
|
return { |
|
'id': video_id, |
|
'title': title, |
|
'description': show.get('desc') or show.get('short_desc'), |
|
'duration': float_or_none(show.get('audio_duration')), |
|
'is_live': False, |
|
'release_date': unified_strdate(show.get('published_date')), |
|
'series': 'Bandcamp Weekly', |
|
'episode': show.get('subtitle'), |
|
'episode_number': episode_number, |
|
'episode_id': compat_str(video_id), |
|
'formats': formats |
|
}
|
|
|