You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
202 lines
7.5 KiB
202 lines
7.5 KiB
from __future__ import unicode_literals |
|
|
|
import datetime |
|
import re |
|
|
|
from .common import InfoExtractor |
|
from ..compat import compat_urlparse |
|
from ..utils import ( |
|
ExtractorError, |
|
InAdvancePagedList, |
|
orderedSet, |
|
str_to_int, |
|
unified_strdate, |
|
) |
|
|
|
|
|
class MotherlessIE(InfoExtractor): |
|
_VALID_URL = r'https?://(?:www\.)?motherless\.com/(?:g/[a-z0-9_]+/)?(?P<id>[A-Z0-9]+)' |
|
_TESTS = [{ |
|
'url': 'http://motherless.com/AC3FFE1', |
|
'md5': '310f62e325a9fafe64f68c0bccb6e75f', |
|
'info_dict': { |
|
'id': 'AC3FFE1', |
|
'ext': 'mp4', |
|
'title': 'Fucked in the ass while playing PS3', |
|
'categories': ['Gaming', 'anal', 'reluctant', 'rough', 'Wife'], |
|
'upload_date': '20100913', |
|
'uploader_id': 'famouslyfuckedup', |
|
'thumbnail': r're:http://.*\.jpg', |
|
'age_limit': 18, |
|
} |
|
}, { |
|
'url': 'http://motherless.com/532291B', |
|
'md5': 'bc59a6b47d1f958e61fbd38a4d31b131', |
|
'info_dict': { |
|
'id': '532291B', |
|
'ext': 'mp4', |
|
'title': 'Amazing girl playing the omegle game, PERFECT!', |
|
'categories': ['Amateur', 'webcam', 'omegle', 'pink', 'young', 'masturbate', 'teen', |
|
'game', 'hairy'], |
|
'upload_date': '20140622', |
|
'uploader_id': 'Sulivana7x', |
|
'thumbnail': r're:http://.*\.jpg', |
|
'age_limit': 18, |
|
}, |
|
'skip': '404', |
|
}, { |
|
'url': 'http://motherless.com/g/cosplay/633979F', |
|
'md5': '0b2a43f447a49c3e649c93ad1fafa4a0', |
|
'info_dict': { |
|
'id': '633979F', |
|
'ext': 'mp4', |
|
'title': 'Turtlette', |
|
'categories': ['superheroine heroine superher'], |
|
'upload_date': '20140827', |
|
'uploader_id': 'shade0230', |
|
'thumbnail': r're:http://.*\.jpg', |
|
'age_limit': 18, |
|
} |
|
}, { |
|
# no keywords |
|
'url': 'http://motherless.com/8B4BBC1', |
|
'only_matching': True, |
|
}] |
|
|
|
def _real_extract(self, url): |
|
video_id = self._match_id(url) |
|
webpage = self._download_webpage(url, video_id) |
|
|
|
if any(p in webpage for p in ( |
|
'<title>404 - MOTHERLESS.COM<', |
|
">The page you're looking for cannot be found.<")): |
|
raise ExtractorError('Video %s does not exist' % video_id, expected=True) |
|
|
|
if '>The content you are trying to view is for friends only.' in webpage: |
|
raise ExtractorError('Video %s is for friends only' % video_id, expected=True) |
|
|
|
title = self._html_search_regex( |
|
r'id="view-upload-title">\s+([^<]+)<', webpage, 'title') |
|
video_url = self._html_search_regex( |
|
r'setup\(\{\s+"file".+: "([^"]+)",', webpage, 'video URL') |
|
age_limit = self._rta_search(webpage) |
|
view_count = str_to_int(self._html_search_regex( |
|
r'<strong>Views</strong>\s+([^<]+)<', |
|
webpage, 'view count', fatal=False)) |
|
like_count = str_to_int(self._html_search_regex( |
|
r'<strong>Favorited</strong>\s+([^<]+)<', |
|
webpage, 'like count', fatal=False)) |
|
|
|
upload_date = self._html_search_regex( |
|
r'<strong>Uploaded</strong>\s+([^<]+)<', webpage, 'upload date') |
|
if 'Ago' in upload_date: |
|
days = int(re.search(r'([0-9]+)', upload_date).group(1)) |
|
upload_date = (datetime.datetime.now() - datetime.timedelta(days=days)).strftime('%Y%m%d') |
|
else: |
|
upload_date = unified_strdate(upload_date) |
|
|
|
comment_count = webpage.count('class="media-comment-contents"') |
|
uploader_id = self._html_search_regex( |
|
r'"thumb-member-username">\s+<a href="/m/([^"]+)"', |
|
webpage, 'uploader_id') |
|
|
|
categories = self._html_search_meta('keywords', webpage, default=None) |
|
if categories: |
|
categories = [cat.strip() for cat in categories.split(',')] |
|
|
|
return { |
|
'id': video_id, |
|
'title': title, |
|
'upload_date': upload_date, |
|
'uploader_id': uploader_id, |
|
'thumbnail': self._og_search_thumbnail(webpage), |
|
'categories': categories, |
|
'view_count': view_count, |
|
'like_count': like_count, |
|
'comment_count': comment_count, |
|
'age_limit': age_limit, |
|
'url': video_url, |
|
} |
|
|
|
|
|
class MotherlessGroupIE(InfoExtractor): |
|
_VALID_URL = 'https?://(?:www\.)?motherless\.com/gv?/(?P<id>[a-z0-9_]+)' |
|
_TESTS = [{ |
|
'url': 'http://motherless.com/g/movie_scenes', |
|
'info_dict': { |
|
'id': 'movie_scenes', |
|
'title': 'Movie Scenes', |
|
'description': 'Hot and sexy scenes from "regular" movies... ' |
|
'Beautiful actresses fully nude... A looot of ' |
|
'skin! :)Enjoy!', |
|
}, |
|
'playlist_mincount': 662, |
|
}, { |
|
'url': 'http://motherless.com/gv/sex_must_be_funny', |
|
'info_dict': { |
|
'id': 'sex_must_be_funny', |
|
'title': 'Sex must be funny', |
|
'description': 'Sex can be funny. Wide smiles,laugh, games, fun of ' |
|
'any kind!' |
|
}, |
|
'playlist_mincount': 9, |
|
}] |
|
|
|
@classmethod |
|
def suitable(cls, url): |
|
return (False if MotherlessIE.suitable(url) |
|
else super(MotherlessGroupIE, cls).suitable(url)) |
|
|
|
def _extract_entries(self, webpage, base): |
|
entries = [] |
|
for mobj in re.finditer( |
|
r'href="(?P<href>/[^"]+)"[^>]*>(?:\s*<img[^>]+alt="[^-]+-\s(?P<title>[^"]+)")?', |
|
webpage): |
|
video_url = compat_urlparse.urljoin(base, mobj.group('href')) |
|
if not MotherlessIE.suitable(video_url): |
|
continue |
|
video_id = MotherlessIE._match_id(video_url) |
|
title = mobj.group('title') |
|
entries.append(self.url_result( |
|
video_url, ie=MotherlessIE.ie_key(), video_id=video_id, |
|
video_title=title)) |
|
# Alternative fallback |
|
if not entries: |
|
entries = [ |
|
self.url_result( |
|
compat_urlparse.urljoin(base, '/' + video_id), |
|
ie=MotherlessIE.ie_key(), video_id=video_id) |
|
for video_id in orderedSet(re.findall( |
|
r'data-codename=["\']([A-Z0-9]+)', webpage))] |
|
return entries |
|
|
|
def _real_extract(self, url): |
|
group_id = self._match_id(url) |
|
page_url = compat_urlparse.urljoin(url, '/gv/%s' % group_id) |
|
webpage = self._download_webpage(page_url, group_id) |
|
title = self._search_regex( |
|
r'<title>([\w\s]+\w)\s+-', webpage, 'title', fatal=False) |
|
description = self._html_search_meta( |
|
'description', webpage, fatal=False) |
|
page_count = self._int(self._search_regex( |
|
r'(\d+)</(?:a|span)><(?:a|span)[^>]+>\s*NEXT', |
|
webpage, 'page_count'), 'page_count') |
|
PAGE_SIZE = 80 |
|
|
|
def _get_page(idx): |
|
webpage = self._download_webpage( |
|
page_url, group_id, query={'page': idx + 1}, |
|
note='Downloading page %d/%d' % (idx + 1, page_count) |
|
) |
|
for entry in self._extract_entries(webpage, url): |
|
yield entry |
|
|
|
playlist = InAdvancePagedList(_get_page, page_count, PAGE_SIZE) |
|
|
|
return { |
|
'_type': 'playlist', |
|
'id': group_id, |
|
'title': title, |
|
'description': description, |
|
'entries': playlist |
|
}
|
|
|