import itertools import re import urllib.parse from .common import InfoExtractor from ..utils import ( int_or_none, mimetype2ext, remove_end, strip_or_none, unified_strdate, url_or_none, urljoin, ) class IwaraBaseIE(InfoExtractor): _BASE_REGEX = r'(?Phttps?://(?:www\.|ecchi\.)?iwara\.tv)' def _extract_playlist(self, base_url, webpage): for path in re.findall(r'class="title">\s*[a-zA-Z0-9]+)' _TESTS = [{ 'url': 'http://iwara.tv/videos/amVwUl1EHpAD9RD', # md5 is unstable 'info_dict': { 'id': 'amVwUl1EHpAD9RD', 'ext': 'mp4', 'title': '【MMD R-18】ガールフレンド carry_me_off', 'age_limit': 18, 'thumbnail': 'https://i.iwara.tv/sites/default/files/videos/thumbnails/7951/thumbnail-7951_0001.png', 'uploader': 'Reimu丨Action', 'upload_date': '20150828', 'description': 'md5:1d4905ce48c66c9299c617f08e106e0f', }, }, { 'url': 'http://ecchi.iwara.tv/videos/Vb4yf2yZspkzkBO', 'md5': '7e5f1f359cd51a027ba4a7b7710a50f0', 'info_dict': { 'id': '0B1LvuHnL-sRFNXB1WHNqbGw4SXc', 'ext': 'mp4', 'title': '[3D Hentai] Kyonyu × Genkai × Emaki Shinobi Girls.mp4', 'age_limit': 18, }, 'add_ie': ['GoogleDrive'], }, { 'url': 'http://www.iwara.tv/videos/nawkaumd6ilezzgq', # md5 is unstable 'info_dict': { 'id': '6liAP9s2Ojc', 'ext': 'mp4', 'age_limit': 18, 'title': '[MMD] Do It Again Ver.2 [1080p 60FPS] (Motion,Camera,Wav+DL)', 'description': 'md5:590c12c0df1443d833fbebe05da8c47a', 'upload_date': '20160910', 'uploader': 'aMMDsork', 'uploader_id': 'UCVOFyOSCyFkXTYYHITtqB7A', }, 'add_ie': ['Youtube'], }] def _real_extract(self, url): video_id = self._match_id(url) webpage, urlh = self._download_webpage_handle(url, video_id) hostname = urllib.parse.urlparse(urlh.geturl()).hostname # ecchi is 'sexy' in Japanese age_limit = 18 if hostname.split('.')[0] == 'ecchi' else 0 video_data = self._download_json('http://www.iwara.tv/api/video/%s' % video_id, video_id) if not video_data: iframe_url = self._html_search_regex( r']+src=([\'"])(?P[^\'"]+)\1', webpage, 'iframe URL', group='url') return { '_type': 'url_transparent', 'url': iframe_url, 'age_limit': age_limit, } title = remove_end(self._html_extract_title(webpage), ' | Iwara') thumbnail = self._html_search_regex( r'poster=[\'"]([^\'"]+)', webpage, 'thumbnail', default=None) uploader = self._html_search_regex( r'class="username">([^<]+)', webpage, 'uploader', fatal=False) upload_date = unified_strdate(self._html_search_regex( r'作成日:([^\s]+)', webpage, 'upload_date', fatal=False)) description = strip_or_none(self._search_regex( r'

(.+?(?=[^/?#&]+)' IE_NAME = 'iwara:playlist' _TESTS = [{ 'url': 'https://ecchi.iwara.tv/playlist/best-enf', 'info_dict': { 'title': 'Best enf', 'uploader': 'Jared98112', 'id': 'best-enf', }, 'playlist_mincount': 1097, }, { # urlencoded 'url': 'https://ecchi.iwara.tv/playlist/%E3%83%97%E3%83%AC%E3%82%A4%E3%83%AA%E3%82%B9%E3%83%88-2', 'info_dict': { 'id': 'プレイリスト-2', 'title': 'プレイリスト', 'uploader': 'mainyu', }, 'playlist_mincount': 91, }] def _real_extract(self, url): playlist_id, base_url = self._match_valid_url(url).group('id', 'base_url') playlist_id = urllib.parse.unquote(playlist_id) webpage = self._download_webpage(url, playlist_id) return { '_type': 'playlist', 'id': playlist_id, 'title': self._html_search_regex(r'class="title"[^>]*>([^<]+)', webpage, 'title', fatal=False), 'uploader': self._html_search_regex(r'

([^<]+)', webpage, 'uploader', fatal=False), 'entries': self._extract_playlist(base_url, webpage), } class IwaraUserIE(IwaraBaseIE): _VALID_URL = fr'{IwaraBaseIE._BASE_REGEX}/users/(?P[^/?#&]+)' IE_NAME = 'iwara:user' _TESTS = [{ 'note': 'number of all videos page is just 1 page. less than 40 videos', 'url': 'https://ecchi.iwara.tv/users/infinityyukarip', 'info_dict': { 'title': 'Uploaded videos from Infinity_YukariP', 'id': 'infinityyukarip', 'uploader': 'Infinity_YukariP', 'uploader_id': 'infinityyukarip', }, 'playlist_mincount': 39, }, { 'note': 'no even all videos page. probably less than 10 videos', 'url': 'https://ecchi.iwara.tv/users/mmd-quintet', 'info_dict': { 'title': 'Uploaded videos from mmd quintet', 'id': 'mmd-quintet', 'uploader': 'mmd quintet', 'uploader_id': 'mmd-quintet', }, 'playlist_mincount': 6, }, { 'note': 'has paging. more than 40 videos', 'url': 'https://ecchi.iwara.tv/users/theblackbirdcalls', 'info_dict': { 'title': 'Uploaded videos from TheBlackbirdCalls', 'id': 'theblackbirdcalls', 'uploader': 'TheBlackbirdCalls', 'uploader_id': 'theblackbirdcalls', }, 'playlist_mincount': 420, }, { 'note': 'foreign chars in URL. there must be foreign characters in URL', 'url': 'https://ecchi.iwara.tv/users/ぶた丼', 'info_dict': { 'title': 'Uploaded videos from ぶた丼', 'id': 'ぶた丼', 'uploader': 'ぶた丼', 'uploader_id': 'ぶた丼', }, 'playlist_mincount': 170, }] def _entries(self, playlist_id, base_url): webpage = self._download_webpage( f'{base_url}/users/{playlist_id}', playlist_id) videos_url = self._search_regex(r'', webpage, 'all videos url', default=None) if not videos_url: yield from self._extract_playlist(base_url, webpage) return videos_url = urljoin(base_url, videos_url) for n in itertools.count(1): page = self._download_webpage( videos_url, playlist_id, note=f'Downloading playlist page {n}', query={'page': str(n - 1)} if n > 1 else {}) yield from self._extract_playlist( base_url, page) if f'page={n}' not in page: break def _real_extract(self, url): playlist_id, base_url = self._match_valid_url(url).group('id', 'base_url') playlist_id = urllib.parse.unquote(playlist_id) return self.playlist_result( self._entries(playlist_id, base_url), playlist_id)