# coding: utf-8 from __future__ import unicode_literals import itertools import random import string import time import json from .common import InfoExtractor from ..compat import ( compat_urllib_parse_unquote, compat_urllib_parse_urlparse ) from ..utils import ( ExtractorError, HEADRequest, get_first, int_or_none, join_nonempty, LazyList, srt_subtitles_timecode, str_or_none, traverse_obj, try_get, url_or_none, qualities, ) class TikTokBaseIE(InfoExtractor): _APP_VERSIONS = [('20.9.3', '293'), ('20.4.3', '243'), ('20.2.1', '221'), ('20.1.2', '212'), ('20.0.4', '204')] _WORKING_APP_VERSION = None _APP_NAME = 'trill' _AID = 1180 _API_HOSTNAME = 'api-h2.tiktokv.com' _UPLOADER_URL_FORMAT = 'https://www.tiktok.com/@%s' _WEBPAGE_HOST = 'https://www.tiktok.com/' QUALITIES = ('360p', '540p', '720p', '1080p') def _call_api_impl(self, ep, query, manifest_app_version, video_id, fatal=True, note='Downloading API JSON', errnote='Unable to download API page'): self._set_cookie(self._API_HOSTNAME, 'odin_tt', ''.join(random.choice('0123456789abcdef') for _ in range(160))) webpage_cookies = self._get_cookies(self._WEBPAGE_HOST) if webpage_cookies.get('sid_tt'): self._set_cookie(self._API_HOSTNAME, 'sid_tt', webpage_cookies['sid_tt'].value) return self._download_json( 'https://%s/aweme/v1/%s/' % (self._API_HOSTNAME, ep), video_id=video_id, fatal=fatal, note=note, errnote=errnote, headers={ 'User-Agent': f'com.ss.android.ugc.trill/{manifest_app_version} (Linux; U; Android 10; en_US; Pixel 4; Build/QQ3A.200805.001; Cronet/58.0.2991.0)', 'Accept': 'application/json', }, query=query) def _build_api_query(self, query, app_version, manifest_app_version): return { **query, 'version_name': app_version, 'version_code': manifest_app_version, 'build_number': app_version, 'manifest_version_code': manifest_app_version, 'update_version_code': manifest_app_version, 'openudid': ''.join(random.choice('0123456789abcdef') for _ in range(16)), 'uuid': ''.join([random.choice(string.digits) for _ in range(16)]), '_rticket': int(time.time() * 1000), 'ts': int(time.time()), 'device_brand': 'Google', 'device_type': 'Pixel 4', 'device_platform': 'android', 'resolution': '1080*1920', 'dpi': 420, 'os_version': '10', 'os_api': '29', 'carrier_region': 'US', 'sys_region': 'US', 'region': 'US', 'app_name': self._APP_NAME, 'app_language': 'en', 'language': 'en', 'timezone_name': 'America/New_York', 'timezone_offset': '-14400', 'channel': 'googleplay', 'ac': 'wifi', 'mcc_mnc': '310260', 'is_my_cn': 0, 'aid': self._AID, 'ssmix': 'a', 'as': 'a1qwert123', 'cp': 'cbfhckdckkde1', } def _call_api(self, ep, query, video_id, fatal=True, note='Downloading API JSON', errnote='Unable to download API page'): if not self._WORKING_APP_VERSION: app_version = self._configuration_arg('app_version', [''], ie_key=TikTokIE.ie_key())[0] manifest_app_version = self._configuration_arg('manifest_app_version', [''], ie_key=TikTokIE.ie_key())[0] if app_version and manifest_app_version: self._WORKING_APP_VERSION = (app_version, manifest_app_version) self.write_debug('Imported app version combo from extractor arguments') elif app_version or manifest_app_version: self.report_warning('Only one of the two required version params are passed as extractor arguments', only_once=True) if self._WORKING_APP_VERSION: app_version, manifest_app_version = self._WORKING_APP_VERSION real_query = self._build_api_query(query, app_version, manifest_app_version) return self._call_api_impl(ep, real_query, manifest_app_version, video_id, fatal, note, errnote) for count, (app_version, manifest_app_version) in enumerate(self._APP_VERSIONS, start=1): real_query = self._build_api_query(query, app_version, manifest_app_version) try: res = self._call_api_impl(ep, real_query, manifest_app_version, video_id, fatal, note, errnote) self._WORKING_APP_VERSION = (app_version, manifest_app_version) return res except ExtractorError as e: if isinstance(e.cause, json.JSONDecodeError) and e.cause.pos == 0: if count == len(self._APP_VERSIONS): if fatal: raise e else: self.report_warning(str(e.cause or e.msg)) return self.report_warning('%s. Retrying... (attempt %s of %s)' % (str(e.cause or e.msg), count, len(self._APP_VERSIONS))) continue raise e def _get_subtitles(self, aweme_detail, aweme_id): # TODO: Extract text positioning info subtitles = {} captions_info = traverse_obj( aweme_detail, ('interaction_stickers', ..., 'auto_video_caption_info', 'auto_captions', ...), expected_type=dict, default=[]) for caption in captions_info: caption_url = traverse_obj(caption, ('url', 'url_list', ...), expected_type=url_or_none, get_all=False) if not caption_url: continue caption_json = self._download_json( caption_url, aweme_id, note='Downloading captions', errnote='Unable to download captions', fatal=False) if not caption_json: continue subtitles.setdefault(caption.get('language', 'en'), []).append({ 'ext': 'srt', 'data': '\n\n'.join( f'{i + 1}\n{srt_subtitles_timecode(line["start_time"] / 1000)} --> {srt_subtitles_timecode(line["end_time"] / 1000)}\n{line["text"]}' for i, line in enumerate(caption_json['utterances']) if line.get('text')) }) return subtitles def _parse_aweme_video_app(self, aweme_detail): aweme_id = aweme_detail['aweme_id'] video_info = aweme_detail['video'] def parse_url_key(url_key): format_id, codec, res, bitrate = self._search_regex( r'v[^_]+_(?P(?P[^_]+)_(?P\d+p)_(?P\d+))', url_key, 'url key', default=(None, None, None, None), group=('id', 'codec', 'res', 'bitrate')) if not format_id: return {}, None return { 'format_id': format_id, 'vcodec': 'h265' if codec == 'bytevc1' else codec, 'tbr': int_or_none(bitrate, scale=1000) or None, 'quality': qualities(self.QUALITIES)(res), }, res known_resolutions = {} def extract_addr(addr, add_meta={}): parsed_meta, res = parse_url_key(addr.get('url_key', '')) if res: known_resolutions.setdefault(res, {}).setdefault('height', add_meta.get('height')) known_resolutions[res].setdefault('width', add_meta.get('width')) parsed_meta.update(known_resolutions.get(res, {})) add_meta.setdefault('height', int_or_none(res[:-1])) return [{ 'url': url, 'filesize': int_or_none(addr.get('data_size')), 'ext': 'mp4', 'acodec': 'aac', 'source_preference': -2 if 'aweme/v1' in url else -1, # Downloads from API might get blocked **add_meta, **parsed_meta, 'format_note': join_nonempty( add_meta.get('format_note'), '(API)' if 'aweme/v1' in url else None, delim=' ') } for url in addr.get('url_list') or []] # Hack: Add direct video links first to prioritize them when removing duplicate formats formats = [] if video_info.get('play_addr'): formats.extend(extract_addr(video_info['play_addr'], { 'format_id': 'play_addr', 'format_note': 'Direct video', 'vcodec': 'h265' if traverse_obj( video_info, 'is_bytevc1', 'is_h265') else 'h264', # TODO: Check for "direct iOS" videos, like https://www.tiktok.com/@cookierun_dev/video/7039716639834656002 'width': video_info.get('width'), 'height': video_info.get('height'), })) if video_info.get('download_addr'): formats.extend(extract_addr(video_info['download_addr'], { 'format_id': 'download_addr', 'format_note': 'Download video%s' % (', watermarked' if video_info.get('has_watermark') else ''), 'vcodec': 'h264', 'width': video_info.get('width'), 'height': video_info.get('height'), 'preference': -2 if video_info.get('has_watermark') else -1, })) if video_info.get('play_addr_h264'): formats.extend(extract_addr(video_info['play_addr_h264'], { 'format_id': 'play_addr_h264', 'format_note': 'Direct video', 'vcodec': 'h264', })) if video_info.get('play_addr_bytevc1'): formats.extend(extract_addr(video_info['play_addr_bytevc1'], { 'format_id': 'play_addr_bytevc1', 'format_note': 'Direct video', 'vcodec': 'h265', })) for bitrate in video_info.get('bit_rate', []): if bitrate.get('play_addr'): formats.extend(extract_addr(bitrate['play_addr'], { 'format_id': bitrate.get('gear_name'), 'format_note': 'Playback video', 'tbr': try_get(bitrate, lambda x: x['bit_rate'] / 1000), 'vcodec': 'h265' if traverse_obj( bitrate, 'is_bytevc1', 'is_h265') else 'h264', 'fps': bitrate.get('FPS'), })) self._remove_duplicate_formats(formats) auth_cookie = self._get_cookies(self._WEBPAGE_HOST).get('sid_tt') if auth_cookie: for f in formats: self._set_cookie(compat_urllib_parse_urlparse(f['url']).hostname, 'sid_tt', auth_cookie.value) self._sort_formats(formats, ('quality', 'codec', 'size', 'br')) thumbnails = [] for cover_id in ('cover', 'ai_dynamic_cover', 'animated_cover', 'ai_dynamic_cover_bak', 'origin_cover', 'dynamic_cover'): cover = video_info.get(cover_id) if cover: for cover_url in cover['url_list']: thumbnails.append({ 'id': cover_id, 'url': cover_url, }) stats_info = aweme_detail.get('statistics', {}) author_info = aweme_detail.get('author', {}) music_info = aweme_detail.get('music', {}) user_url = self._UPLOADER_URL_FORMAT % (traverse_obj(author_info, 'sec_uid', 'id', 'uid', 'unique_id', expected_type=str_or_none, get_all=False)) labels = traverse_obj(aweme_detail, ('hybrid_label', ..., 'text'), expected_type=str, default=[]) contained_music_track = traverse_obj( music_info, ('matched_song', 'title'), ('matched_pgc_sound', 'title'), expected_type=str) contained_music_author = traverse_obj( music_info, ('matched_song', 'author'), ('matched_pgc_sound', 'author'), 'author', expected_type=str) is_generic_og_trackname = music_info.get('is_original_sound') and music_info.get('title') == 'original sound - %s' % music_info.get('owner_handle') if is_generic_og_trackname: music_track, music_author = contained_music_track or 'original sound', contained_music_author else: music_track, music_author = music_info.get('title'), music_info.get('author') return { 'id': aweme_id, 'title': aweme_detail.get('desc'), 'description': aweme_detail.get('desc'), 'view_count': int_or_none(stats_info.get('play_count')), 'like_count': int_or_none(stats_info.get('digg_count')), 'repost_count': int_or_none(stats_info.get('share_count')), 'comment_count': int_or_none(stats_info.get('comment_count')), 'uploader': str_or_none(author_info.get('unique_id')), 'creator': str_or_none(author_info.get('nickname')), 'uploader_id': str_or_none(author_info.get('uid')), 'uploader_url': user_url, 'track': music_track, 'album': str_or_none(music_info.get('album')) or None, 'artist': music_author, 'timestamp': int_or_none(aweme_detail.get('create_time')), 'formats': formats, 'subtitles': self.extract_subtitles(aweme_detail, aweme_id), 'thumbnails': thumbnails, 'duration': int_or_none(traverse_obj(video_info, 'duration', ('download_addr', 'duration')), scale=1000), 'availability': self._availability( is_private='Private' in labels, needs_subscription='Friends only' in labels, is_unlisted='Followers only' in labels) } def _parse_aweme_video_web(self, aweme_detail, webpage_url): video_info = aweme_detail['video'] author_info = traverse_obj(aweme_detail, 'authorInfo', 'author', expected_type=dict, default={}) music_info = aweme_detail.get('music') or {} stats_info = aweme_detail.get('stats') or {} user_url = self._UPLOADER_URL_FORMAT % (traverse_obj(author_info, 'secUid', 'id', 'uid', 'uniqueId', expected_type=str_or_none, get_all=False) or aweme_detail.get('authorSecId')) formats = [] play_url = video_info.get('playAddr') width = video_info.get('width') height = video_info.get('height') if isinstance(play_url, str): formats = [{ 'url': self._proto_relative_url(play_url), 'ext': 'mp4', 'width': width, 'height': height, }] elif isinstance(play_url, list): formats = [{ 'url': self._proto_relative_url(url), 'ext': 'mp4', 'width': width, 'height': height, } for url in traverse_obj(play_url, (..., 'src'), expected_type=url_or_none, default=[]) if url] download_url = url_or_none(video_info.get('downloadAddr')) or traverse_obj(video_info, ('download', 'url'), expected_type=url_or_none) if download_url: formats.append({ 'format_id': 'download', 'url': self._proto_relative_url(download_url), 'ext': 'mp4', 'width': width, 'height': height, }) self._remove_duplicate_formats(formats) self._sort_formats(formats) thumbnails = [] for thumbnail_name in ('thumbnail', 'cover', 'dynamicCover', 'originCover'): if aweme_detail.get(thumbnail_name): thumbnails = [{ 'url': self._proto_relative_url(aweme_detail[thumbnail_name]), 'width': width, 'height': height }] return { 'id': traverse_obj(aweme_detail, 'id', 'awemeId', expected_type=str_or_none), 'title': aweme_detail.get('desc'), 'duration': try_get(aweme_detail, lambda x: x['video']['duration'], int), 'view_count': int_or_none(stats_info.get('playCount')), 'like_count': int_or_none(stats_info.get('diggCount')), 'repost_count': int_or_none(stats_info.get('shareCount')), 'comment_count': int_or_none(stats_info.get('commentCount')), 'timestamp': int_or_none(aweme_detail.get('createTime')), 'creator': str_or_none(author_info.get('nickname')), 'uploader': str_or_none(author_info.get('uniqueId') or aweme_detail.get('author')), 'uploader_id': str_or_none(author_info.get('id') or aweme_detail.get('authorId')), 'uploader_url': user_url, 'track': str_or_none(music_info.get('title')), 'album': str_or_none(music_info.get('album')) or None, 'artist': str_or_none(music_info.get('authorName')), 'formats': formats, 'thumbnails': thumbnails, 'description': str_or_none(aweme_detail.get('desc')), 'http_headers': { 'Referer': webpage_url } } class TikTokIE(TikTokBaseIE): _VALID_URL = r'https?://www\.tiktok\.com/@[\w\.-]+/video/(?P\d+)' _TESTS = [{ 'url': 'https://www.tiktok.com/@leenabhushan/video/6748451240264420610', 'md5': '736bb7a466c6f0a6afeb597da1e6f5b7', 'info_dict': { 'id': '6748451240264420610', 'ext': 'mp4', 'title': '#jassmanak #lehanga #leenabhushan', 'description': '#jassmanak #lehanga #leenabhushan', 'duration': 13, 'height': 1024, 'width': 576, 'uploader': 'leenabhushan', 'uploader_id': '6691488002098119685', 'uploader_url': 'https://www.tiktok.com/@MS4wLjABAAAA_Eb4t1vodM1IuTy_cvp9CY22RAb59xqrO0Xtz9CYQJvgXaDvZxYnZYRzDWhhgJmy', 'creator': 'facestoriesbyleenabh', 'thumbnail': r're:^https?://[\w\/\.\-]+(~[\w\-]+\.image)?', 'upload_date': '20191016', 'timestamp': 1571246252, 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, 'artist': 'Ysrbeats', 'album': 'Lehanga', 'track': 'Lehanga', } }, { 'url': 'https://www.tiktok.com/@patroxofficial/video/6742501081818877190?langCountry=en', 'md5': '6f3cf8cdd9b28cb8363fe0a9a160695b', 'info_dict': { 'id': '6742501081818877190', 'ext': 'mp4', 'title': 'md5:5e2a23877420bb85ce6521dbee39ba94', 'description': 'md5:5e2a23877420bb85ce6521dbee39ba94', 'duration': 27, 'height': 960, 'width': 540, 'uploader': 'patrox', 'uploader_id': '18702747', 'uploader_url': 'https://www.tiktok.com/@MS4wLjABAAAAiFnldaILebi5heDoVU6bn4jBWWycX6-9U3xuNPqZ8Ws', 'creator': 'patroX', 'thumbnail': r're:^https?://[\w\/\.\-]+(~[\w\-]+\.image)?', 'upload_date': '20190930', 'timestamp': 1569860870, 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, 'artist': 'Evan Todd, Jessica Keenan Wynn, Alice Lee, Barrett Wilbert Weed & Jon Eidson', 'track': 'Big Fun', } }, { # Banned audio, only available on the app 'url': 'https://www.tiktok.com/@barudakhb_/video/6984138651336838402', 'info_dict': { 'id': '6984138651336838402', 'ext': 'mp4', 'title': 'Balas @yolaaftwsr hayu yu ? #SquadRandom_ 🔥', 'description': 'Balas @yolaaftwsr hayu yu ? #SquadRandom_ 🔥', 'uploader': 'barudakhb_', 'creator': 'md5:29f238c49bc0c176cb3cef1a9cea9fa6', 'uploader_id': '6974687867511718913', 'uploader_url': 'https://www.tiktok.com/@MS4wLjABAAAAbhBwQC-R1iKoix6jDFsF-vBdfx2ABoDjaZrM9fX6arU3w71q3cOWgWuTXn1soZ7d', 'track': 'Boka Dance', 'artist': 'md5:29f238c49bc0c176cb3cef1a9cea9fa6', 'timestamp': 1626121503, 'duration': 18, 'thumbnail': r're:^https?://[\w\/\.\-]+(~[\w\-]+\.image)?', 'upload_date': '20210712', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, } }, { # Sponsored video, only available with feed workaround 'url': 'https://www.tiktok.com/@MS4wLjABAAAATh8Vewkn0LYM7Fo03iec3qKdeCUOcBIouRk1mkiag6h3o_pQu_dUXvZ2EZlGST7_/video/7042692929109986561', 'info_dict': { 'id': '7042692929109986561', 'ext': 'mp4', 'title': 'Slap and Run!', 'description': 'Slap and Run!', 'uploader': 'user440922249', 'creator': 'Slap And Run', 'uploader_id': '7036055384943690754', 'uploader_url': 'https://www.tiktok.com/@MS4wLjABAAAATh8Vewkn0LYM7Fo03iec3qKdeCUOcBIouRk1mkiag6h3o_pQu_dUXvZ2EZlGST7_', 'track': 'Promoted Music', 'timestamp': 1639754738, 'duration': 30, 'thumbnail': r're:^https?://[\w\/\.\-]+(~[\w\-]+\.image)?', 'upload_date': '20211217', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, }, 'expected_warnings': ['Video not available'] }, { # Video without title and description 'url': 'https://www.tiktok.com/@pokemonlife22/video/7059698374567611694', 'info_dict': { 'id': '7059698374567611694', 'ext': 'mp4', 'title': 'tiktok video #7059698374567611694', 'description': '', 'uploader': 'pokemonlife22', 'creator': 'Pokemon', 'uploader_id': '6820838815978423302', 'uploader_url': 'https://www.tiktok.com/@MS4wLjABAAAA0tF1nBwQVVMyrGu3CqttkNgM68Do1OXUFuCY0CRQk8fEtSVDj89HqoqvbSTmUP2W', 'track': 'original sound', 'timestamp': 1643714123, 'duration': 6, 'thumbnail': r're:^https?://[\w\/\.\-]+(~[\w\-]+\.image)?', 'upload_date': '20220201', 'artist': 'Pokemon', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, }, 'expected_warnings': ['Video not available', 'Creating a generic title'] }, { # Auto-captions available 'url': 'https://www.tiktok.com/@hankgreen1/video/7047596209028074758', 'only_matching': True }] def _extract_aweme_app(self, aweme_id): try: aweme_detail = self._call_api('aweme/detail', {'aweme_id': aweme_id}, aweme_id, note='Downloading video details', errnote='Unable to download video details').get('aweme_detail') if not aweme_detail: raise ExtractorError('Video not available', video_id=aweme_id) except ExtractorError as e: self.report_warning(f'{e}; Retrying with feed workaround') feed_list = self._call_api('feed', {'aweme_id': aweme_id}, aweme_id, note='Downloading video feed', errnote='Unable to download video feed').get('aweme_list') or [] aweme_detail = next((aweme for aweme in feed_list if str(aweme.get('aweme_id')) == aweme_id), None) if not aweme_detail: raise ExtractorError('Unable to find video in feed', video_id=aweme_id) return self._parse_aweme_video_app(aweme_detail) def _real_extract(self, url): video_id = self._match_id(url) try: return self._extract_aweme_app(video_id) except ExtractorError as e: self.report_warning(f'{e}; Retrying with webpage') # If we only call once, we get a 403 when downlaoding the video. self._download_webpage(url, video_id) webpage = self._download_webpage(url, video_id, note='Downloading video webpage') next_data = self._search_nextjs_data(webpage, video_id, default='{}') if next_data: status = traverse_obj(next_data, ('props', 'pageProps', 'statusCode'), expected_type=int) or 0 video_data = traverse_obj(next_data, ('props', 'pageProps', 'itemInfo', 'itemStruct'), expected_type=dict) else: sigi_json = self._search_regex( r'>\s*window\[[\'"]SIGI_STATE[\'"]\]\s*=\s*(?P{.+});', webpage, 'sigi data', group='sigi_state') sigi_data = self._parse_json(sigi_json, video_id) status = traverse_obj(sigi_data, ('VideoPage', 'statusCode'), expected_type=int) or 0 video_data = traverse_obj(sigi_data, ('ItemModule', video_id), expected_type=dict) if status == 0: return self._parse_aweme_video_web(video_data, url) elif status == 10216: raise ExtractorError('This video is private', expected=True) raise ExtractorError('Video not available', video_id=video_id) class TikTokUserIE(TikTokBaseIE): IE_NAME = 'tiktok:user' _VALID_URL = r'https?://(?:www\.)?tiktok\.com/@(?P[\w\.-]+)/?(?:$|[#?])' _TESTS = [{ 'url': 'https://tiktok.com/@corgibobaa?lang=en', 'playlist_mincount': 45, 'info_dict': { 'id': '6935371178089399301', 'title': 'corgibobaa', 'thumbnail': r're:https://.+_1080x1080\.webp' }, 'expected_warnings': ['Retrying'] }, { 'url': 'https://www.tiktok.com/@6820838815978423302', 'playlist_mincount': 5, 'info_dict': { 'id': '6820838815978423302', 'title': '6820838815978423302', 'thumbnail': r're:https://.+_1080x1080\.webp' }, 'expected_warnings': ['Retrying'] }, { 'url': 'https://www.tiktok.com/@meme', 'playlist_mincount': 593, 'info_dict': { 'id': '79005827461758976', 'title': 'meme', 'thumbnail': r're:https://.+_1080x1080\.webp' }, 'expected_warnings': ['Retrying'] }] r''' # TODO: Fix by adding _signature to api_url def _entries(self, webpage, user_id, username): secuid = self._search_regex(r'\"secUid\":\"(?P[^\"]+)', webpage, username) verifyfp_cookie = self._get_cookies('https://www.tiktok.com').get('s_v_web_id') if not verifyfp_cookie: raise ExtractorError('Improper cookies (missing s_v_web_id).', expected=True) api_url = f'https://m.tiktok.com/api/post/item_list/?aid=1988&cookie_enabled=true&count=30&verifyFp={verifyfp_cookie.value}&secUid={secuid}&cursor=' cursor = '0' for page in itertools.count(): data_json = self._download_json(api_url + cursor, username, note='Downloading Page %d' % page) for video in data_json.get('itemList', []): video_id = video['id'] video_url = f'https://www.tiktok.com/@{user_id}/video/{video_id}' yield self._url_result(video_url, 'TikTok', video_id, str_or_none(video.get('desc'))) if not data_json.get('hasMore'): break cursor = data_json['cursor'] ''' def _video_entries_api(self, webpage, user_id, username): query = { 'user_id': user_id, 'count': 21, 'max_cursor': 0, 'min_cursor': 0, 'retry_type': 'no_retry', 'device_id': ''.join(random.choice(string.digits) for _ in range(19)), # Some endpoints don't like randomized device_id, so it isn't directly set in _call_api. } max_retries = self.get_param('extractor_retries', 3) for page in itertools.count(1): for retries in itertools.count(): try: post_list = self._call_api('aweme/post', query, username, note='Downloading user video list page %d%s' % (page, f' (attempt {retries})' if retries != 0 else ''), errnote='Unable to download user video list') except ExtractorError as e: if isinstance(e.cause, json.JSONDecodeError) and e.cause.pos == 0 and retries != max_retries: self.report_warning('%s. Retrying...' % str(e.cause or e.msg)) continue raise break yield from post_list.get('aweme_list', []) if not post_list.get('has_more'): break query['max_cursor'] = post_list['max_cursor'] def _entries_api(self, user_id, videos): for video in videos: yield { **self._parse_aweme_video_app(video), 'extractor_key': TikTokIE.ie_key(), 'extractor': 'TikTok', 'webpage_url': f'https://tiktok.com/@{user_id}/video/{video["aweme_id"]}', } def _real_extract(self, url): user_name = self._match_id(url) webpage = self._download_webpage(url, user_name, headers={ 'User-Agent': 'facebookexternalhit/1.1 (+http://www.facebook.com/externalhit_uatext.php)' }) user_id = self._html_search_regex(r'snssdk\d*://user/profile/(\d+)', webpage, 'user ID', default=None) or user_name videos = LazyList(self._video_entries_api(webpage, user_id, user_name)) thumbnail = traverse_obj(videos, (0, 'author', 'avatar_larger', 'url_list', 0)) return self.playlist_result(self._entries_api(user_id, videos), user_id, user_name, thumbnail=thumbnail) class TikTokBaseListIE(TikTokBaseIE): def _entries(self, list_id, display_id): query = { self._QUERY_NAME: list_id, 'cursor': 0, 'count': 20, 'type': 5, 'device_id': ''.join(random.choice(string.digits) for i in range(19)) } max_retries = self.get_param('extractor_retries', 3) for page in itertools.count(1): for retries in itertools.count(): try: post_list = self._call_api(self._API_ENDPOINT, query, display_id, note='Downloading video list page %d%s' % (page, f' (attempt {retries})' if retries != 0 else ''), errnote='Unable to download video list') except ExtractorError as e: if isinstance(e.cause, json.JSONDecodeError) and e.cause.pos == 0 and retries != max_retries: self.report_warning('%s. Retrying...' % str(e.cause or e.msg)) continue raise break for video in post_list.get('aweme_list', []): yield { **self._parse_aweme_video_app(video), 'extractor_key': TikTokIE.ie_key(), 'extractor': 'TikTok', 'webpage_url': f'https://tiktok.com/@_/video/{video["aweme_id"]}', } if not post_list.get('has_more'): break query['cursor'] = post_list['cursor'] def _real_extract(self, url): list_id = self._match_id(url) return self.playlist_result(self._entries(list_id, list_id), list_id) class TikTokSoundIE(TikTokBaseListIE): IE_NAME = 'tiktok:sound' _VALID_URL = r'https?://(?:www\.)?tiktok\.com/music/[\w\.-]+-(?P[\d]+)[/?#&]?' _QUERY_NAME = 'music_id' _API_ENDPOINT = 'music/aweme' _TESTS = [{ 'url': 'https://www.tiktok.com/music/Build-a-Btch-6956990112127585029?lang=en', 'playlist_mincount': 100, 'info_dict': { 'id': '6956990112127585029' }, 'expected_warnings': ['Retrying'] }, { # Actual entries are less than listed video count 'url': 'https://www.tiktok.com/music/jiefei-soap-remix-7036843036118469381', 'playlist_mincount': 2182, 'info_dict': { 'id': '7036843036118469381' }, 'expected_warnings': ['Retrying'] }] class TikTokEffectIE(TikTokBaseListIE): IE_NAME = 'tiktok:effect' _VALID_URL = r'https?://(?:www\.)?tiktok\.com/sticker/[\w\.-]+-(?P[\d]+)[/?#&]?' _QUERY_NAME = 'sticker_id' _API_ENDPOINT = 'sticker/aweme' _TESTS = [{ 'url': 'https://www.tiktok.com/sticker/MATERIAL-GWOOORL-1258156', 'playlist_mincount': 100, 'info_dict': { 'id': '1258156', }, 'expected_warnings': ['Retrying'] }, { # Different entries between mobile and web, depending on region 'url': 'https://www.tiktok.com/sticker/Elf-Friend-479565', 'only_matching': True }] class TikTokTagIE(TikTokBaseListIE): IE_NAME = 'tiktok:tag' _VALID_URL = r'https?://(?:www\.)?tiktok\.com/tag/(?P[^/?#&]+)' _QUERY_NAME = 'ch_id' _API_ENDPOINT = 'challenge/aweme' _TESTS = [{ 'url': 'https://tiktok.com/tag/hello2018', 'playlist_mincount': 39, 'info_dict': { 'id': '46294678', 'title': 'hello2018', }, 'expected_warnings': ['Retrying'] }, { 'url': 'https://tiktok.com/tag/fypシ?is_copy_url=0&is_from_webapp=v1', 'only_matching': True }] def _real_extract(self, url): display_id = self._match_id(url) webpage = self._download_webpage(url, display_id, headers={ 'User-Agent': 'facebookexternalhit/1.1 (+http://www.facebook.com/externalhit_uatext.php)' }) tag_id = self._html_search_regex(r'snssdk\d*://challenge/detail/(\d+)', webpage, 'tag ID') return self.playlist_result(self._entries(tag_id, display_id), tag_id, display_id) class DouyinIE(TikTokIE): _VALID_URL = r'https?://(?:www\.)?douyin\.com/video/(?P[0-9]+)' _TESTS = [{ 'url': 'https://www.douyin.com/video/6961737553342991651', 'md5': '10523312c8b8100f353620ac9dc8f067', 'info_dict': { 'id': '6961737553342991651', 'ext': 'mp4', 'title': '#杨超越 小小水手带你去远航❤️', 'uploader': '杨超越', 'upload_date': '20210513', 'timestamp': 1620905839, 'uploader_id': '110403406559', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, } }, { 'url': 'https://www.douyin.com/video/6982497745948921092', 'md5': 'd78408c984b9b5102904cf6b6bc2d712', 'info_dict': { 'id': '6982497745948921092', 'ext': 'mp4', 'title': '这个夏日和小羊@杨超越 一起遇见白色幻想', 'uploader': '杨超越工作室', 'upload_date': '20210708', 'timestamp': 1625739481, 'uploader_id': '408654318141572', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, } }, { 'url': 'https://www.douyin.com/video/6953975910773099811', 'md5': '72e882e24f75064c218b76c8b713c185', 'info_dict': { 'id': '6953975910773099811', 'ext': 'mp4', 'title': '#一起看海 出现在你的夏日里', 'uploader': '杨超越', 'upload_date': '20210422', 'timestamp': 1619098692, 'uploader_id': '110403406559', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, } }, { 'url': 'https://www.douyin.com/video/6950251282489675042', 'md5': 'b4db86aec367ef810ddd38b1737d2fed', 'info_dict': { 'id': '6950251282489675042', 'ext': 'mp4', 'title': '哈哈哈,成功了哈哈哈哈哈哈', 'uploader': '杨超越', 'upload_date': '20210412', 'timestamp': 1618231483, 'uploader_id': '110403406559', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, } }, { 'url': 'https://www.douyin.com/video/6963263655114722595', 'md5': '1abe1c477d05ee62efb40bf2329957cf', 'info_dict': { 'id': '6963263655114722595', 'ext': 'mp4', 'title': '#哪个爱豆的105度最甜 换个角度看看我哈哈', 'uploader': '杨超越', 'upload_date': '20210517', 'timestamp': 1621261163, 'uploader_id': '110403406559', 'view_count': int, 'like_count': int, 'repost_count': int, 'comment_count': int, } }] _APP_VERSIONS = [('9.6.0', '960')] _APP_NAME = 'aweme' _AID = 1128 _API_HOSTNAME = 'aweme.snssdk.com' _UPLOADER_URL_FORMAT = 'https://www.douyin.com/user/%s' _WEBPAGE_HOST = 'https://www.douyin.com/' def _real_extract(self, url): video_id = self._match_id(url) try: return self._extract_aweme_app(video_id) except ExtractorError as e: self.report_warning(f'{e}; Retrying with webpage') webpage = self._download_webpage(url, video_id) render_data_json = self._search_regex( r'', webpage, 'render data', default=None) if not render_data_json: # TODO: Run verification challenge code to generate signature cookies raise ExtractorError('Fresh cookies (not necessarily logged in) are needed') render_data = self._parse_json( render_data_json, video_id, transform_source=compat_urllib_parse_unquote) return self._parse_aweme_video_web(get_first(render_data, ('aweme', 'detail')), url) class TikTokVMIE(InfoExtractor): _VALID_URL = r'https?://(?:vm|vt)\.tiktok\.com/(?P\w+)' IE_NAME = 'vm.tiktok' _TESTS = [{ 'url': 'https://vm.tiktok.com/ZSe4FqkKd', 'info_dict': { 'id': '7023491746608712966', 'ext': 'mp4', 'title': 'md5:5607564db90271abbbf8294cca77eddd', 'description': 'md5:5607564db90271abbbf8294cca77eddd', 'duration': 11, 'upload_date': '20211026', 'uploader_id': '7007385080558846981', 'creator': 'Memes', 'artist': 'Memes', 'track': 'original sound', 'uploader': 'susmandem', 'timestamp': 1635284105, 'thumbnail': r're:https://.+\.webp.*', 'like_count': int, 'view_count': int, 'comment_count': int, 'repost_count': int, 'uploader_url': 'https://www.tiktok.com/@MS4wLjABAAAAXcNoOEOxVyBzuII_E--T0MeCrLP0ay1Sm6x_n3dluiWEoWZD0VlQOytwad4W0i0n', } }, { 'url': 'https://vt.tiktok.com/ZSe4FqkKd', 'only_matching': True, }] def _real_extract(self, url): return self.url_result(self._request_webpage( HEADRequest(url), self._match_id(url), headers={'User-Agent': 'facebookexternalhit/1.1'}).geturl(), TikTokIE)