mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-30 12:01:28 +00:00
[youtube] misc cleanup and bug fixes (#505)
* Update some `_extract_response` calls to keep them consistent * Cleanup continuation extraction related code using new API format * Improve `_extract_account_syncid` to support multiple parameters * Generalize `get_text` and related functions into one * Update `INNERTUBE_CONTEXT_CLIENT_NAME` with integer values Authored by: colethedj
This commit is contained in:
parent
314ee30548
commit
fe93e2c4cf
|
@ -5,6 +5,7 @@
|
||||||
import base64
|
import base64
|
||||||
import calendar
|
import calendar
|
||||||
import copy
|
import copy
|
||||||
|
import datetime
|
||||||
import hashlib
|
import hashlib
|
||||||
import itertools
|
import itertools
|
||||||
import json
|
import json
|
||||||
|
@ -54,7 +55,8 @@
|
||||||
update_url_query,
|
update_url_query,
|
||||||
url_or_none,
|
url_or_none,
|
||||||
urlencode_postdata,
|
urlencode_postdata,
|
||||||
urljoin
|
urljoin,
|
||||||
|
variadic
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@ -360,7 +362,7 @@ def _real_initialize(self):
|
||||||
'hl': 'en',
|
'hl': 'en',
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
'INNERTUBE_CONTEXT_CLIENT_NAME': 'ANDROID'
|
'INNERTUBE_CONTEXT_CLIENT_NAME': 3
|
||||||
},
|
},
|
||||||
'ANDROID_EMBEDDED_PLAYER': {
|
'ANDROID_EMBEDDED_PLAYER': {
|
||||||
'INNERTUBE_API_VERSION': 'v1',
|
'INNERTUBE_API_VERSION': 'v1',
|
||||||
|
@ -374,7 +376,7 @@ def _real_initialize(self):
|
||||||
'hl': 'en',
|
'hl': 'en',
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
'INNERTUBE_CONTEXT_CLIENT_NAME': 'ANDROID_EMBEDDED_PLAYER'
|
'INNERTUBE_CONTEXT_CLIENT_NAME': 55
|
||||||
},
|
},
|
||||||
'ANDROID_MUSIC': {
|
'ANDROID_MUSIC': {
|
||||||
'INNERTUBE_API_VERSION': 'v1',
|
'INNERTUBE_API_VERSION': 'v1',
|
||||||
|
@ -388,7 +390,7 @@ def _real_initialize(self):
|
||||||
'hl': 'en',
|
'hl': 'en',
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
'INNERTUBE_CONTEXT_CLIENT_NAME': 'ANDROID_MUSIC'
|
'INNERTUBE_CONTEXT_CLIENT_NAME': 21
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -498,20 +500,23 @@ def _extract_identity_token(self, webpage, item_id):
|
||||||
'identity token', default=None)
|
'identity token', default=None)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _extract_account_syncid(data):
|
def _extract_account_syncid(*args):
|
||||||
"""
|
"""
|
||||||
Extract syncId required to download private playlists of secondary channels
|
Extract syncId required to download private playlists of secondary channels
|
||||||
@param data Either response or ytcfg
|
@params response and/or ytcfg
|
||||||
"""
|
"""
|
||||||
sync_ids = (try_get(
|
for data in args:
|
||||||
data, (lambda x: x['responseContext']['mainAppWebResponseContext']['datasyncId'],
|
# ytcfg includes channel_syncid if on secondary channel
|
||||||
lambda x: x['DATASYNC_ID']), compat_str) or '').split("||")
|
delegated_sid = try_get(data, lambda x: x['DELEGATED_SESSION_ID'], compat_str)
|
||||||
if len(sync_ids) >= 2 and sync_ids[1]:
|
if delegated_sid:
|
||||||
# datasyncid is of the form "channel_syncid||user_syncid" for secondary channel
|
return delegated_sid
|
||||||
# and just "user_syncid||" for primary channel. We only want the channel_syncid
|
sync_ids = (try_get(
|
||||||
return sync_ids[0]
|
data, (lambda x: x['responseContext']['mainAppWebResponseContext']['datasyncId'],
|
||||||
# ytcfg includes channel_syncid if on secondary channel
|
lambda x: x['DATASYNC_ID']), compat_str) or '').split("||")
|
||||||
return data.get('DELEGATED_SESSION_ID')
|
if len(sync_ids) >= 2 and sync_ids[1]:
|
||||||
|
# datasyncid is of the form "channel_syncid||user_syncid" for secondary channel
|
||||||
|
# and just "user_syncid||" for primary channel. We only want the channel_syncid
|
||||||
|
return sync_ids[0]
|
||||||
|
|
||||||
def _extract_ytcfg(self, video_id, webpage):
|
def _extract_ytcfg(self, video_id, webpage):
|
||||||
if not webpage:
|
if not webpage:
|
||||||
|
@ -561,21 +566,6 @@ def _build_api_continuation_query(continuation, ctp=None):
|
||||||
query['clickTracking'] = {'clickTrackingParams': ctp}
|
query['clickTracking'] = {'clickTrackingParams': ctp}
|
||||||
return query
|
return query
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def _continuation_query_ajax_to_api(cls, continuation_query):
|
|
||||||
continuation = dict_get(continuation_query, ('continuation', 'ctoken'))
|
|
||||||
return cls._build_api_continuation_query(continuation, continuation_query.get('itct'))
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _build_continuation_query(continuation, ctp=None):
|
|
||||||
query = {
|
|
||||||
'ctoken': continuation,
|
|
||||||
'continuation': continuation,
|
|
||||||
}
|
|
||||||
if ctp:
|
|
||||||
query['itct'] = ctp
|
|
||||||
return query
|
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _extract_next_continuation_data(cls, renderer):
|
def _extract_next_continuation_data(cls, renderer):
|
||||||
next_continuation = try_get(
|
next_continuation = try_get(
|
||||||
|
@ -587,7 +577,7 @@ def _extract_next_continuation_data(cls, renderer):
|
||||||
if not continuation:
|
if not continuation:
|
||||||
return
|
return
|
||||||
ctp = next_continuation.get('clickTrackingParams')
|
ctp = next_continuation.get('clickTrackingParams')
|
||||||
return cls._build_continuation_query(continuation, ctp)
|
return cls._build_api_continuation_query(continuation, ctp)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _extract_continuation_ep_data(cls, continuation_ep: dict):
|
def _extract_continuation_ep_data(cls, continuation_ep: dict):
|
||||||
|
@ -597,16 +587,18 @@ def _extract_continuation_ep_data(cls, continuation_ep: dict):
|
||||||
if not continuation:
|
if not continuation:
|
||||||
return
|
return
|
||||||
ctp = continuation_ep.get('clickTrackingParams')
|
ctp = continuation_ep.get('clickTrackingParams')
|
||||||
return cls._build_continuation_query(continuation, ctp)
|
return cls._build_api_continuation_query(continuation, ctp)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _extract_continuation(cls, renderer):
|
def _extract_continuation(cls, renderer):
|
||||||
next_continuation = cls._extract_next_continuation_data(renderer)
|
next_continuation = cls._extract_next_continuation_data(renderer)
|
||||||
if next_continuation:
|
if next_continuation:
|
||||||
return next_continuation
|
return next_continuation
|
||||||
|
|
||||||
contents = []
|
contents = []
|
||||||
for key in ('contents', 'items'):
|
for key in ('contents', 'items'):
|
||||||
contents.extend(try_get(renderer, lambda x: x[key], list) or [])
|
contents.extend(try_get(renderer, lambda x: x[key], list) or [])
|
||||||
|
|
||||||
for content in contents:
|
for content in contents:
|
||||||
if not isinstance(content, dict):
|
if not isinstance(content, dict):
|
||||||
continue
|
continue
|
||||||
|
@ -618,8 +610,8 @@ def _extract_continuation(cls, renderer):
|
||||||
if continuation:
|
if continuation:
|
||||||
return continuation
|
return continuation
|
||||||
|
|
||||||
@staticmethod
|
@classmethod
|
||||||
def _extract_alerts(data):
|
def _extract_alerts(cls, data):
|
||||||
for alert_dict in try_get(data, lambda x: x['alerts'], list) or []:
|
for alert_dict in try_get(data, lambda x: x['alerts'], list) or []:
|
||||||
if not isinstance(alert_dict, dict):
|
if not isinstance(alert_dict, dict):
|
||||||
continue
|
continue
|
||||||
|
@ -627,11 +619,7 @@ def _extract_alerts(data):
|
||||||
alert_type = alert.get('type')
|
alert_type = alert.get('type')
|
||||||
if not alert_type:
|
if not alert_type:
|
||||||
continue
|
continue
|
||||||
message = try_get(alert, lambda x: x['text']['simpleText'], compat_str) or ''
|
message = cls._get_text(alert.get('text'))
|
||||||
if message:
|
|
||||||
yield alert_type, message
|
|
||||||
for run in try_get(alert, lambda x: x['text']['runs'], list) or []:
|
|
||||||
message += try_get(run, lambda x: x['text'], compat_str)
|
|
||||||
if message:
|
if message:
|
||||||
yield alert_type, message
|
yield alert_type, message
|
||||||
|
|
||||||
|
@ -661,18 +649,23 @@ def _extract_badges(self, renderer: dict):
|
||||||
return badges
|
return badges
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _join_text_entries(runs):
|
def _get_text(data, getter=None, max_runs=None):
|
||||||
text = None
|
for get in variadic(getter):
|
||||||
for run in runs:
|
d = try_get(data, get) if get is not None else data
|
||||||
if not isinstance(run, dict):
|
text = try_get(d, lambda x: x['simpleText'], compat_str)
|
||||||
continue
|
if text:
|
||||||
sub_text = try_get(run, lambda x: x['text'], compat_str)
|
return text
|
||||||
if sub_text:
|
runs = try_get(d, lambda x: x['runs'], list) or []
|
||||||
if not text:
|
if not runs and isinstance(d, list):
|
||||||
text = sub_text
|
runs = d
|
||||||
continue
|
|
||||||
text += sub_text
|
def get_runs(runs):
|
||||||
return text
|
for run in runs[:min(len(runs), max_runs or len(runs))]:
|
||||||
|
yield try_get(run, lambda x: x['text'], compat_str) or ''
|
||||||
|
|
||||||
|
text = ''.join(get_runs(runs))
|
||||||
|
if text:
|
||||||
|
return text
|
||||||
|
|
||||||
def _extract_response(self, item_id, query, note='Downloading API JSON', headers=None,
|
def _extract_response(self, item_id, query, note='Downloading API JSON', headers=None,
|
||||||
ytcfg=None, check_get_keys=None, ep='browse', fatal=True, api_hostname=None,
|
ytcfg=None, check_get_keys=None, ep='browse', fatal=True, api_hostname=None,
|
||||||
|
@ -736,24 +729,16 @@ def is_music_url(url):
|
||||||
|
|
||||||
def _extract_video(self, renderer):
|
def _extract_video(self, renderer):
|
||||||
video_id = renderer.get('videoId')
|
video_id = renderer.get('videoId')
|
||||||
title = try_get(
|
title = self._get_text(renderer.get('title'))
|
||||||
renderer,
|
description = self._get_text(renderer.get('descriptionSnippet'))
|
||||||
(lambda x: x['title']['runs'][0]['text'],
|
duration = parse_duration(self._get_text(renderer.get('lengthText')))
|
||||||
lambda x: x['title']['simpleText']), compat_str)
|
view_count_text = self._get_text(renderer.get('viewCountText')) or ''
|
||||||
description = try_get(
|
|
||||||
renderer, lambda x: x['descriptionSnippet']['runs'][0]['text'],
|
|
||||||
compat_str)
|
|
||||||
duration = parse_duration(try_get(
|
|
||||||
renderer, lambda x: x['lengthText']['simpleText'], compat_str))
|
|
||||||
view_count_text = try_get(
|
|
||||||
renderer, lambda x: x['viewCountText']['simpleText'], compat_str) or ''
|
|
||||||
view_count = str_to_int(self._search_regex(
|
view_count = str_to_int(self._search_regex(
|
||||||
r'^([\d,]+)', re.sub(r'\s', '', view_count_text),
|
r'^([\d,]+)', re.sub(r'\s', '', view_count_text),
|
||||||
'view count', default=None))
|
'view count', default=None))
|
||||||
uploader = try_get(
|
|
||||||
renderer,
|
uploader = self._get_text(renderer, (lambda x: x['ownerText'], lambda x: x['shortBylineText']))
|
||||||
(lambda x: x['ownerText']['runs'][0]['text'],
|
|
||||||
lambda x: x['shortBylineText']['runs'][0]['text']), compat_str)
|
|
||||||
return {
|
return {
|
||||||
'_type': 'url',
|
'_type': 'url',
|
||||||
'ie_key': YoutubeIE.ie_key(),
|
'ie_key': YoutubeIE.ie_key(),
|
||||||
|
@ -2004,15 +1989,18 @@ def _extract_comment(self, comment_renderer, parent=None):
|
||||||
comment_id = comment_renderer.get('commentId')
|
comment_id = comment_renderer.get('commentId')
|
||||||
if not comment_id:
|
if not comment_id:
|
||||||
return
|
return
|
||||||
comment_text_runs = try_get(comment_renderer, lambda x: x['contentText']['runs']) or []
|
|
||||||
text = self._join_text_entries(comment_text_runs) or ''
|
text = self._get_text(comment_renderer.get('contentText'))
|
||||||
comment_time_text = try_get(comment_renderer, lambda x: x['publishedTimeText']['runs']) or []
|
|
||||||
time_text = self._join_text_entries(comment_time_text)
|
|
||||||
# note: timestamp is an estimate calculated from the current time and time_text
|
# note: timestamp is an estimate calculated from the current time and time_text
|
||||||
timestamp = calendar.timegm(self.parse_time_text(time_text).timetuple())
|
time_text = self._get_text(comment_renderer.get('publishedTimeText')) or ''
|
||||||
author = try_get(comment_renderer, lambda x: x['authorText']['simpleText'], compat_str)
|
time_text_dt = self.parse_time_text(time_text)
|
||||||
|
if isinstance(time_text_dt, datetime.datetime):
|
||||||
|
timestamp = calendar.timegm(time_text_dt.timetuple())
|
||||||
|
author = self._get_text(comment_renderer.get('authorText'))
|
||||||
author_id = try_get(comment_renderer,
|
author_id = try_get(comment_renderer,
|
||||||
lambda x: x['authorEndpoint']['browseEndpoint']['browseId'], compat_str)
|
lambda x: x['authorEndpoint']['browseEndpoint']['browseId'], compat_str)
|
||||||
|
|
||||||
votes = parse_count(try_get(comment_renderer, (lambda x: x['voteCount']['simpleText'],
|
votes = parse_count(try_get(comment_renderer, (lambda x: x['voteCount']['simpleText'],
|
||||||
lambda x: x['likeCount']), compat_str)) or 0
|
lambda x: x['likeCount']), compat_str)) or 0
|
||||||
author_thumbnail = try_get(comment_renderer,
|
author_thumbnail = try_get(comment_renderer,
|
||||||
|
@ -2043,13 +2031,12 @@ def extract_header(contents):
|
||||||
_continuation = None
|
_continuation = None
|
||||||
for content in contents:
|
for content in contents:
|
||||||
comments_header_renderer = try_get(content, lambda x: x['commentsHeaderRenderer'])
|
comments_header_renderer = try_get(content, lambda x: x['commentsHeaderRenderer'])
|
||||||
expected_comment_count = try_get(comments_header_renderer,
|
expected_comment_count = parse_count(self._get_text(
|
||||||
(lambda x: x['countText']['runs'][0]['text'],
|
comments_header_renderer, (lambda x: x['countText'], lambda x: x['commentsCount']), max_runs=1))
|
||||||
lambda x: x['commentsCount']['runs'][0]['text']),
|
|
||||||
compat_str)
|
|
||||||
if expected_comment_count:
|
if expected_comment_count:
|
||||||
comment_counts[1] = str_to_int(expected_comment_count)
|
comment_counts[1] = expected_comment_count
|
||||||
self.to_screen('Downloading ~%d comments' % str_to_int(expected_comment_count))
|
self.to_screen('Downloading ~%d comments' % expected_comment_count)
|
||||||
_total_comments = comment_counts[1]
|
_total_comments = comment_counts[1]
|
||||||
sort_mode_str = self._configuration_arg('comment_sort', [''])[0]
|
sort_mode_str = self._configuration_arg('comment_sort', [''])[0]
|
||||||
comment_sort_index = int(sort_mode_str != 'top') # 1 = new, 0 = top
|
comment_sort_index = int(sort_mode_str != 'top') # 1 = new, 0 = top
|
||||||
|
@ -2110,10 +2097,10 @@ def extract_thread(contents):
|
||||||
comment_counts = [0, 0, 0]
|
comment_counts = [0, 0, 0]
|
||||||
|
|
||||||
continuation = self._extract_continuation(root_continuation_data)
|
continuation = self._extract_continuation(root_continuation_data)
|
||||||
if continuation and len(continuation['ctoken']) < 27:
|
if continuation and len(continuation['continuation']) < 27:
|
||||||
self.write_debug('Detected old API continuation token. Generating new API compatible token.')
|
self.write_debug('Detected old API continuation token. Generating new API compatible token.')
|
||||||
continuation_token = self._generate_comment_continuation(video_id)
|
continuation_token = self._generate_comment_continuation(video_id)
|
||||||
continuation = self._build_continuation_query(continuation_token, None)
|
continuation = self._build_api_continuation_query(continuation_token, None)
|
||||||
|
|
||||||
visitor_data = None
|
visitor_data = None
|
||||||
is_first_continuation = parent is None
|
is_first_continuation = parent is None
|
||||||
|
@ -2135,7 +2122,7 @@ def extract_thread(contents):
|
||||||
page_num, comment_prog_str)
|
page_num, comment_prog_str)
|
||||||
|
|
||||||
response = self._extract_response(
|
response = self._extract_response(
|
||||||
item_id=None, query=self._continuation_query_ajax_to_api(continuation),
|
item_id=None, query=continuation,
|
||||||
ep='next', ytcfg=ytcfg, headers=headers, note=note_prefix,
|
ep='next', ytcfg=ytcfg, headers=headers, note=note_prefix,
|
||||||
check_get_keys=('onResponseReceivedEndpoints', 'continuationContents'))
|
check_get_keys=('onResponseReceivedEndpoints', 'continuationContents'))
|
||||||
if not response:
|
if not response:
|
||||||
|
@ -2298,9 +2285,7 @@ def _real_extract(self, url):
|
||||||
|
|
||||||
ytcfg = self._extract_ytcfg(video_id, webpage) or self._get_default_ytcfg()
|
ytcfg = self._extract_ytcfg(video_id, webpage) or self._get_default_ytcfg()
|
||||||
identity_token = self._extract_identity_token(webpage, video_id)
|
identity_token = self._extract_identity_token(webpage, video_id)
|
||||||
syncid = self._extract_account_syncid(ytcfg)
|
|
||||||
session_index = self._extract_session_index(ytcfg)
|
session_index = self._extract_session_index(ytcfg)
|
||||||
headers = self._generate_api_headers(ytcfg, identity_token, syncid, session_index=session_index)
|
|
||||||
player_url = self._extract_player_url(ytcfg, webpage)
|
player_url = self._extract_player_url(ytcfg, webpage)
|
||||||
|
|
||||||
player_client = self._configuration_arg('player_client', [''])[0]
|
player_client = self._configuration_arg('player_client', [''])[0]
|
||||||
|
@ -2308,17 +2293,14 @@ def _real_extract(self, url):
|
||||||
self.report_warning(f'Invalid player_client {player_client} given. Falling back to android client.')
|
self.report_warning(f'Invalid player_client {player_client} given. Falling back to android client.')
|
||||||
force_mobile_client = player_client != 'web'
|
force_mobile_client = player_client != 'web'
|
||||||
player_skip = self._configuration_arg('player_skip')
|
player_skip = self._configuration_arg('player_skip')
|
||||||
|
player_response = None
|
||||||
|
if webpage:
|
||||||
|
player_response = self._extract_yt_initial_variable(
|
||||||
|
webpage, self._YT_INITIAL_PLAYER_RESPONSE_RE,
|
||||||
|
video_id, 'initial player response')
|
||||||
|
|
||||||
def get_text(x):
|
syncid = self._extract_account_syncid(ytcfg, player_response)
|
||||||
if not x:
|
headers = self._generate_api_headers(ytcfg, identity_token, syncid, session_index=session_index)
|
||||||
return
|
|
||||||
text = x.get('simpleText')
|
|
||||||
if text and isinstance(text, compat_str):
|
|
||||||
return text
|
|
||||||
runs = x.get('runs')
|
|
||||||
if not isinstance(runs, list):
|
|
||||||
return
|
|
||||||
return ''.join([r['text'] for r in runs if isinstance(r.get('text'), compat_str)])
|
|
||||||
|
|
||||||
ytm_streaming_data = {}
|
ytm_streaming_data = {}
|
||||||
if is_music_url:
|
if is_music_url:
|
||||||
|
@ -2352,12 +2334,6 @@ def get_text(x):
|
||||||
note='Downloading %sremix player API JSON' % ('android ' if force_mobile_client else ''))
|
note='Downloading %sremix player API JSON' % ('android ' if force_mobile_client else ''))
|
||||||
ytm_streaming_data = try_get(ytm_player_response, lambda x: x['streamingData'], dict) or {}
|
ytm_streaming_data = try_get(ytm_player_response, lambda x: x['streamingData'], dict) or {}
|
||||||
|
|
||||||
player_response = None
|
|
||||||
if webpage:
|
|
||||||
player_response = self._extract_yt_initial_variable(
|
|
||||||
webpage, self._YT_INITIAL_PLAYER_RESPONSE_RE,
|
|
||||||
video_id, 'initial player response')
|
|
||||||
|
|
||||||
if not player_response or force_mobile_client:
|
if not player_response or force_mobile_client:
|
||||||
sts = self._extract_signature_timestamp(video_id, player_url, ytcfg, fatal=False)
|
sts = self._extract_signature_timestamp(video_id, player_url, ytcfg, fatal=False)
|
||||||
yt_client = 'WEB'
|
yt_client = 'WEB'
|
||||||
|
@ -2456,7 +2432,7 @@ def get_text(x):
|
||||||
lambda x: x['microformat']['playerMicroformatRenderer'],
|
lambda x: x['microformat']['playerMicroformatRenderer'],
|
||||||
dict) or {}
|
dict) or {}
|
||||||
video_title = video_details.get('title') \
|
video_title = video_details.get('title') \
|
||||||
or get_text(microformat.get('title')) \
|
or self._get_text(microformat.get('title')) \
|
||||||
or search_meta(['og:title', 'twitter:title', 'title'])
|
or search_meta(['og:title', 'twitter:title', 'title'])
|
||||||
video_description = video_details.get('shortDescription')
|
video_description = video_details.get('shortDescription')
|
||||||
|
|
||||||
|
@ -2635,10 +2611,10 @@ def feed_entry(name):
|
||||||
playability_status,
|
playability_status,
|
||||||
lambda x: x['errorScreen']['playerErrorMessageRenderer'],
|
lambda x: x['errorScreen']['playerErrorMessageRenderer'],
|
||||||
dict) or {}
|
dict) or {}
|
||||||
reason = get_text(pemr.get('reason')) or playability_status.get('reason')
|
reason = self._get_text(pemr.get('reason')) or playability_status.get('reason')
|
||||||
subreason = pemr.get('subreason')
|
subreason = pemr.get('subreason')
|
||||||
if subreason:
|
if subreason:
|
||||||
subreason = clean_html(get_text(subreason))
|
subreason = clean_html(self._get_text(subreason))
|
||||||
if subreason == 'The uploader has not made this video available in your country.':
|
if subreason == 'The uploader has not made this video available in your country.':
|
||||||
countries = microformat.get('availableCountries')
|
countries = microformat.get('availableCountries')
|
||||||
if not countries:
|
if not countries:
|
||||||
|
@ -2785,9 +2761,7 @@ def process_language(container, base_url, lang_code, sub_name, query):
|
||||||
continue
|
continue
|
||||||
process_language(
|
process_language(
|
||||||
automatic_captions, base_url, translation_language_code,
|
automatic_captions, base_url, translation_language_code,
|
||||||
try_get(translation_language, (
|
self._get_text(translation_language.get('languageName'), max_runs=1),
|
||||||
lambda x: x['languageName']['simpleText'],
|
|
||||||
lambda x: x['languageName']['runs'][0]['text'])),
|
|
||||||
{'tlang': translation_language_code})
|
{'tlang': translation_language_code})
|
||||||
info['automatic_captions'] = automatic_captions
|
info['automatic_captions'] = automatic_captions
|
||||||
info['subtitles'] = subtitles
|
info['subtitles'] = subtitles
|
||||||
|
@ -2855,7 +2829,7 @@ def process_language(container, base_url, lang_code, sub_name, query):
|
||||||
|
|
||||||
def chapter_time(mmlir):
|
def chapter_time(mmlir):
|
||||||
return parse_duration(
|
return parse_duration(
|
||||||
get_text(mmlir.get('timeDescription')))
|
self._get_text(mmlir.get('timeDescription')))
|
||||||
|
|
||||||
chapters = []
|
chapters = []
|
||||||
for next_num, content in enumerate(contents, start=1):
|
for next_num, content in enumerate(contents, start=1):
|
||||||
|
@ -2869,7 +2843,7 @@ def chapter_time(mmlir):
|
||||||
chapters.append({
|
chapters.append({
|
||||||
'start_time': start_time,
|
'start_time': start_time,
|
||||||
'end_time': end_time,
|
'end_time': end_time,
|
||||||
'title': get_text(mmlir.get('title')),
|
'title': self._get_text(mmlir.get('title')),
|
||||||
})
|
})
|
||||||
if chapters:
|
if chapters:
|
||||||
break
|
break
|
||||||
|
@ -2885,7 +2859,7 @@ def chapter_time(mmlir):
|
||||||
if vpir:
|
if vpir:
|
||||||
stl = vpir.get('superTitleLink')
|
stl = vpir.get('superTitleLink')
|
||||||
if stl:
|
if stl:
|
||||||
stl = get_text(stl)
|
stl = self._get_text(stl)
|
||||||
if try_get(
|
if try_get(
|
||||||
vpir,
|
vpir,
|
||||||
lambda x: x['superTitleIcon']['iconType']) == 'LOCATION_PIN':
|
lambda x: x['superTitleIcon']['iconType']) == 'LOCATION_PIN':
|
||||||
|
@ -2925,7 +2899,7 @@ def chapter_time(mmlir):
|
||||||
})
|
})
|
||||||
vsir = content.get('videoSecondaryInfoRenderer')
|
vsir = content.get('videoSecondaryInfoRenderer')
|
||||||
if vsir:
|
if vsir:
|
||||||
info['channel'] = get_text(try_get(
|
info['channel'] = self._get_text(try_get(
|
||||||
vsir,
|
vsir,
|
||||||
lambda x: x['owner']['videoOwnerRenderer']['title'],
|
lambda x: x['owner']['videoOwnerRenderer']['title'],
|
||||||
dict))
|
dict))
|
||||||
|
@ -2943,8 +2917,8 @@ def chapter_time(mmlir):
|
||||||
mrr_title = mrr.get('title')
|
mrr_title = mrr.get('title')
|
||||||
if not mrr_title:
|
if not mrr_title:
|
||||||
continue
|
continue
|
||||||
mrr_title = get_text(mrr['title'])
|
mrr_title = self._get_text(mrr['title'])
|
||||||
mrr_contents_text = get_text(mrr['contents'][0])
|
mrr_contents_text = self._get_text(mrr['contents'][0])
|
||||||
if mrr_title == 'License':
|
if mrr_title == 'License':
|
||||||
info['license'] = mrr_contents_text
|
info['license'] = mrr_contents_text
|
||||||
elif not multiple_songs:
|
elif not multiple_songs:
|
||||||
|
@ -3515,9 +3489,8 @@ def _grid_entries(self, grid_renderer):
|
||||||
renderer = self._extract_basic_item_renderer(item)
|
renderer = self._extract_basic_item_renderer(item)
|
||||||
if not isinstance(renderer, dict):
|
if not isinstance(renderer, dict):
|
||||||
continue
|
continue
|
||||||
title = try_get(
|
title = self._get_text(renderer.get('title'))
|
||||||
renderer, (lambda x: x['title']['runs'][0]['text'],
|
|
||||||
lambda x: x['title']['simpleText']), compat_str)
|
|
||||||
# playlist
|
# playlist
|
||||||
playlist_id = renderer.get('playlistId')
|
playlist_id = renderer.get('playlistId')
|
||||||
if playlist_id:
|
if playlist_id:
|
||||||
|
@ -3534,8 +3507,6 @@ def _grid_entries(self, grid_renderer):
|
||||||
# channel
|
# channel
|
||||||
channel_id = renderer.get('channelId')
|
channel_id = renderer.get('channelId')
|
||||||
if channel_id:
|
if channel_id:
|
||||||
title = try_get(
|
|
||||||
renderer, lambda x: x['title']['simpleText'], compat_str)
|
|
||||||
yield self.url_result(
|
yield self.url_result(
|
||||||
'https://www.youtube.com/channel/%s' % channel_id,
|
'https://www.youtube.com/channel/%s' % channel_id,
|
||||||
ie=YoutubeTabIE.ie_key(), video_title=title)
|
ie=YoutubeTabIE.ie_key(), video_title=title)
|
||||||
|
@ -3578,8 +3549,7 @@ def _shelf_entries(self, shelf_renderer, skip_channels=False):
|
||||||
# will not work
|
# will not work
|
||||||
if skip_channels and '/channels?' in shelf_url:
|
if skip_channels and '/channels?' in shelf_url:
|
||||||
return
|
return
|
||||||
title = try_get(
|
title = self._get_text(shelf_renderer, lambda x: x['title'])
|
||||||
shelf_renderer, lambda x: x['title']['runs'][0]['text'], compat_str)
|
|
||||||
yield self.url_result(shelf_url, video_title=title)
|
yield self.url_result(shelf_url, video_title=title)
|
||||||
# Shelf may not contain shelf URL, fallback to extraction from content
|
# Shelf may not contain shelf URL, fallback to extraction from content
|
||||||
for entry in self._shelf_entries_from_content(shelf_renderer):
|
for entry in self._shelf_entries_from_content(shelf_renderer):
|
||||||
|
@ -3718,20 +3688,15 @@ def extract_entries(parent_renderer): # this needs to called again for continua
|
||||||
for entry in extract_entries(parent_renderer):
|
for entry in extract_entries(parent_renderer):
|
||||||
yield entry
|
yield entry
|
||||||
continuation = continuation_list[0]
|
continuation = continuation_list[0]
|
||||||
context = self._extract_context(ytcfg)
|
visitor_data = None
|
||||||
visitor_data = try_get(context, lambda x: x['client']['visitorData'], compat_str)
|
|
||||||
|
|
||||||
for page_num in itertools.count(1):
|
for page_num in itertools.count(1):
|
||||||
if not continuation:
|
if not continuation:
|
||||||
break
|
break
|
||||||
query = {
|
|
||||||
'continuation': continuation['continuation'],
|
|
||||||
'clickTracking': {'clickTrackingParams': continuation['itct']}
|
|
||||||
}
|
|
||||||
headers = self._generate_api_headers(ytcfg, identity_token, account_syncid, visitor_data)
|
headers = self._generate_api_headers(ytcfg, identity_token, account_syncid, visitor_data)
|
||||||
response = self._extract_response(
|
response = self._extract_response(
|
||||||
item_id='%s page %s' % (item_id, page_num),
|
item_id='%s page %s' % (item_id, page_num),
|
||||||
query=query, headers=headers, ytcfg=ytcfg,
|
query=continuation, headers=headers, ytcfg=ytcfg,
|
||||||
check_get_keys=('continuationContents', 'onResponseReceivedActions', 'onResponseReceivedEndpoints'))
|
check_get_keys=('continuationContents', 'onResponseReceivedActions', 'onResponseReceivedEndpoints'))
|
||||||
|
|
||||||
if not response:
|
if not response:
|
||||||
|
@ -3877,21 +3842,20 @@ def _extract_from_tabs(self, item_id, webpage, data, tabs):
|
||||||
'channel': metadata['uploader'],
|
'channel': metadata['uploader'],
|
||||||
'channel_id': metadata['uploader_id'],
|
'channel_id': metadata['uploader_id'],
|
||||||
'channel_url': metadata['uploader_url']})
|
'channel_url': metadata['uploader_url']})
|
||||||
|
ytcfg = self._extract_ytcfg(item_id, webpage)
|
||||||
return self.playlist_result(
|
return self.playlist_result(
|
||||||
self._entries(
|
self._entries(
|
||||||
selected_tab, playlist_id,
|
selected_tab, playlist_id,
|
||||||
self._extract_identity_token(webpage, item_id),
|
self._extract_identity_token(webpage, item_id),
|
||||||
self._extract_account_syncid(data),
|
self._extract_account_syncid(ytcfg, data), ytcfg),
|
||||||
self._extract_ytcfg(item_id, webpage)),
|
|
||||||
**metadata)
|
**metadata)
|
||||||
|
|
||||||
def _extract_mix_playlist(self, playlist, playlist_id, data, webpage):
|
def _extract_mix_playlist(self, playlist, playlist_id, data, webpage):
|
||||||
first_id = last_id = None
|
first_id = last_id = None
|
||||||
ytcfg = self._extract_ytcfg(playlist_id, webpage)
|
ytcfg = self._extract_ytcfg(playlist_id, webpage)
|
||||||
headers = self._generate_api_headers(
|
headers = self._generate_api_headers(
|
||||||
ytcfg, account_syncid=self._extract_account_syncid(data),
|
ytcfg, account_syncid=self._extract_account_syncid(ytcfg, data),
|
||||||
identity_token=self._extract_identity_token(webpage, item_id=playlist_id),
|
identity_token=self._extract_identity_token(webpage, item_id=playlist_id))
|
||||||
visitor_data=try_get(self._extract_context(ytcfg), lambda x: x['client']['visitorData'], compat_str))
|
|
||||||
for page_num in itertools.count(1):
|
for page_num in itertools.count(1):
|
||||||
videos = list(self._playlist_entries(playlist))
|
videos = list(self._playlist_entries(playlist))
|
||||||
if not videos:
|
if not videos:
|
||||||
|
@ -3916,9 +3880,7 @@ def _extract_mix_playlist(self, playlist, playlist_id, data, webpage):
|
||||||
}
|
}
|
||||||
response = self._extract_response(
|
response = self._extract_response(
|
||||||
item_id='%s page %d' % (playlist_id, page_num),
|
item_id='%s page %d' % (playlist_id, page_num),
|
||||||
query=query,
|
query=query, ep='next', headers=headers, ytcfg=ytcfg,
|
||||||
ep='next',
|
|
||||||
headers=headers,
|
|
||||||
check_get_keys='contents'
|
check_get_keys='contents'
|
||||||
)
|
)
|
||||||
playlist = try_get(
|
playlist = try_get(
|
||||||
|
@ -3960,8 +3922,8 @@ def _extract_availability(self, data):
|
||||||
renderer_dict, lambda x: x['privacyDropdownItemRenderer']['isSelected'], bool) or False
|
renderer_dict, lambda x: x['privacyDropdownItemRenderer']['isSelected'], bool) or False
|
||||||
if not is_selected:
|
if not is_selected:
|
||||||
continue
|
continue
|
||||||
label = self._join_text_entries(
|
label = self._get_text(
|
||||||
try_get(renderer_dict, lambda x: x['privacyDropdownItemRenderer']['label']['runs'], list) or [])
|
try_get(renderer_dict, lambda x: x['privacyDropdownItemRenderer']['label'], dict) or [])
|
||||||
if label:
|
if label:
|
||||||
badge_labels.add(label.lower())
|
badge_labels.add(label.lower())
|
||||||
break
|
break
|
||||||
|
@ -4010,7 +3972,7 @@ def _reload_with_unavailable_videos(self, item_id, data, webpage):
|
||||||
|
|
||||||
ytcfg = self._extract_ytcfg(item_id, webpage)
|
ytcfg = self._extract_ytcfg(item_id, webpage)
|
||||||
headers = self._generate_api_headers(
|
headers = self._generate_api_headers(
|
||||||
ytcfg, account_syncid=self._extract_account_syncid(ytcfg),
|
ytcfg, account_syncid=self._extract_account_syncid(ytcfg, data),
|
||||||
identity_token=self._extract_identity_token(webpage, item_id=item_id),
|
identity_token=self._extract_identity_token(webpage, item_id=item_id),
|
||||||
visitor_data=try_get(
|
visitor_data=try_get(
|
||||||
self._extract_context(ytcfg), lambda x: x['client']['visitorData'], compat_str))
|
self._extract_context(ytcfg), lambda x: x['client']['visitorData'], compat_str))
|
||||||
|
@ -4020,7 +3982,7 @@ def _reload_with_unavailable_videos(self, item_id, data, webpage):
|
||||||
}
|
}
|
||||||
return self._extract_response(
|
return self._extract_response(
|
||||||
item_id=item_id, headers=headers, query=query,
|
item_id=item_id, headers=headers, query=query,
|
||||||
check_get_keys='contents', fatal=False,
|
check_get_keys='contents', fatal=False, ytcfg=ytcfg,
|
||||||
note='Downloading API JSON with unavailable videos')
|
note='Downloading API JSON with unavailable videos')
|
||||||
|
|
||||||
def _extract_webpage(self, url, item_id):
|
def _extract_webpage(self, url, item_id):
|
||||||
|
@ -4352,7 +4314,9 @@ def _entries(self, query, n):
|
||||||
if self._SEARCH_PARAMS:
|
if self._SEARCH_PARAMS:
|
||||||
data['params'] = self._SEARCH_PARAMS
|
data['params'] = self._SEARCH_PARAMS
|
||||||
total = 0
|
total = 0
|
||||||
|
continuation = {}
|
||||||
for page_num in itertools.count(1):
|
for page_num in itertools.count(1):
|
||||||
|
data.update(continuation)
|
||||||
search = self._extract_response(
|
search = self._extract_response(
|
||||||
item_id='query "%s" page %s' % (query, page_num), ep='search', query=data,
|
item_id='query "%s" page %s' % (query, page_num), ep='search', query=data,
|
||||||
check_get_keys=('contents', 'onResponseReceivedCommands')
|
check_get_keys=('contents', 'onResponseReceivedCommands')
|
||||||
|
@ -4370,13 +4334,10 @@ def _entries(self, query, n):
|
||||||
# Youtube sometimes adds promoted content to searches,
|
# Youtube sometimes adds promoted content to searches,
|
||||||
# changing the index location of videos and token.
|
# changing the index location of videos and token.
|
||||||
# So we search through all entries till we find them.
|
# So we search through all entries till we find them.
|
||||||
continuation_token = None
|
continuation = None
|
||||||
for slr_content in slr_contents:
|
for slr_content in slr_contents:
|
||||||
if continuation_token is None:
|
if not continuation:
|
||||||
continuation_token = try_get(
|
continuation = self._extract_continuation({'contents': [slr_content]})
|
||||||
slr_content,
|
|
||||||
lambda x: x['continuationItemRenderer']['continuationEndpoint']['continuationCommand']['token'],
|
|
||||||
compat_str)
|
|
||||||
|
|
||||||
isr_contents = try_get(
|
isr_contents = try_get(
|
||||||
slr_content,
|
slr_content,
|
||||||
|
@ -4399,9 +4360,8 @@ def _entries(self, query, n):
|
||||||
if total == n:
|
if total == n:
|
||||||
return
|
return
|
||||||
|
|
||||||
if not continuation_token:
|
if not continuation:
|
||||||
break
|
break
|
||||||
data['continuation'] = continuation_token
|
|
||||||
|
|
||||||
def _get_n_results(self, query, n):
|
def _get_n_results(self, query, n):
|
||||||
"""Get a specified number of results for a query"""
|
"""Get a specified number of results for a query"""
|
||||||
|
|
Loading…
Reference in a new issue