mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-18 14:19:18 +00:00
More calls to trouble changed to report_error
This commit is contained in:
parent
9edb0916f4
commit
613bf66939
|
@ -406,10 +406,10 @@ def prepare_filename(self, info_dict):
|
||||||
filename = self.params['outtmpl'] % template_dict
|
filename = self.params['outtmpl'] % template_dict
|
||||||
return filename
|
return filename
|
||||||
except KeyError as err:
|
except KeyError as err:
|
||||||
self.trouble(u'ERROR: Erroneous output template')
|
self.report_error(u'Erroneous output template')
|
||||||
return None
|
return None
|
||||||
except ValueError as err:
|
except ValueError as err:
|
||||||
self.trouble(u'ERROR: Insufficient system charset ' + repr(preferredencoding()))
|
self.report_error(u'Insufficient system charset ' + repr(preferredencoding()))
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def _match_entry(self, info_dict):
|
def _match_entry(self, info_dict):
|
||||||
|
@ -468,16 +468,16 @@ def extract_info(self, url, download = True, ie_name = None):
|
||||||
results.append(self.process_ie_result(ie_result, download))
|
results.append(self.process_ie_result(ie_result, download))
|
||||||
return results
|
return results
|
||||||
except ExtractorError as de: # An error we somewhat expected
|
except ExtractorError as de: # An error we somewhat expected
|
||||||
self.trouble(u'ERROR: ' + compat_str(de), de.format_traceback())
|
self.report_error(compat_str(de), de.format_traceback())
|
||||||
break
|
break
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
if self.params.get('ignoreerrors', False):
|
if self.params.get('ignoreerrors', False):
|
||||||
self.trouble(u'ERROR: ' + compat_str(e), tb=compat_str(traceback.format_exc()))
|
self.report_error(compat_str(e), tb=compat_str(traceback.format_exc()))
|
||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
raise
|
raise
|
||||||
if not suitable_found:
|
if not suitable_found:
|
||||||
self.trouble(u'ERROR: no suitable InfoExtractor: %s' % url)
|
self.report_error(u'no suitable InfoExtractor: %s' % url)
|
||||||
|
|
||||||
def process_ie_result(self, ie_result, download = True):
|
def process_ie_result(self, ie_result, download = True):
|
||||||
"""
|
"""
|
||||||
|
@ -636,7 +636,7 @@ def process_info(self, info_dict):
|
||||||
with io.open(encodeFilename(sub_filename), 'w', encoding='utf-8') as subfile:
|
with io.open(encodeFilename(sub_filename), 'w', encoding='utf-8') as subfile:
|
||||||
subfile.write(sub)
|
subfile.write(sub)
|
||||||
except (OSError, IOError):
|
except (OSError, IOError):
|
||||||
self.trouble(u'ERROR: Cannot write subtitles file ' + descfn)
|
self.report_error(u'Cannot write subtitles file ' + descfn)
|
||||||
return
|
return
|
||||||
if self.params.get('onlysubtitles', False):
|
if self.params.get('onlysubtitles', False):
|
||||||
return
|
return
|
||||||
|
@ -683,7 +683,7 @@ def download(self, url_list):
|
||||||
#It also downloads the videos
|
#It also downloads the videos
|
||||||
videos = self.extract_info(url)
|
videos = self.extract_info(url)
|
||||||
except UnavailableVideoError:
|
except UnavailableVideoError:
|
||||||
self.trouble(u'\nERROR: unable to download video')
|
self.report_error(u'unable to download video')
|
||||||
except MaxDownloadsReached:
|
except MaxDownloadsReached:
|
||||||
self.to_screen(u'[info] Maximum number of downloaded files reached.')
|
self.to_screen(u'[info] Maximum number of downloaded files reached.')
|
||||||
raise
|
raise
|
||||||
|
|
|
@ -1224,7 +1224,7 @@ def grep_webpage(self, url, regex, regexFlags, matchTuples):
|
||||||
|
|
||||||
for (i, key, err) in matchTuples:
|
for (i, key, err) in matchTuples:
|
||||||
if mobj.group(i) is None:
|
if mobj.group(i) is None:
|
||||||
self._downloader.trouble(err)
|
self._downloader.report_error(err)
|
||||||
return
|
return
|
||||||
else:
|
else:
|
||||||
info[key] = mobj.group(i)
|
info[key] = mobj.group(i)
|
||||||
|
@ -1238,7 +1238,7 @@ def extractLiveStream(self, url):
|
||||||
r'src="(.*?/videothek_js.*?\.js)',
|
r'src="(.*?/videothek_js.*?\.js)',
|
||||||
0,
|
0,
|
||||||
[
|
[
|
||||||
(1, 'url', u'ERROR: Invalid URL: %s' % url)
|
(1, 'url', u'Invalid URL: %s' % url)
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
http_host = url.split('/')[2]
|
http_host = url.split('/')[2]
|
||||||
|
@ -1250,9 +1250,9 @@ def extractLiveStream(self, url):
|
||||||
'(rtmp://.*?)\'',
|
'(rtmp://.*?)\'',
|
||||||
re.DOTALL,
|
re.DOTALL,
|
||||||
[
|
[
|
||||||
(1, 'path', u'ERROR: could not extract video path: %s' % url),
|
(1, 'path', u'could not extract video path: %s' % url),
|
||||||
(2, 'player', u'ERROR: could not extract video player: %s' % url),
|
(2, 'player', u'could not extract video player: %s' % url),
|
||||||
(3, 'url', u'ERROR: could not extract video url: %s' % url)
|
(3, 'url', u'could not extract video url: %s' % url)
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
video_url = u'%s/%s' % (info.get('url'), info.get('path'))
|
video_url = u'%s/%s' % (info.get('url'), info.get('path'))
|
||||||
|
@ -1264,7 +1264,7 @@ def extractPlus7Stream(self, url):
|
||||||
r'param name="movie".*?videorefFileUrl=(http[^\'"&]*)',
|
r'param name="movie".*?videorefFileUrl=(http[^\'"&]*)',
|
||||||
0,
|
0,
|
||||||
[
|
[
|
||||||
(1, 'url', u'ERROR: Invalid URL: %s' % url)
|
(1, 'url', u'Invalid URL: %s' % url)
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
next_url = compat_urllib_parse.unquote(info.get('url'))
|
next_url = compat_urllib_parse.unquote(info.get('url'))
|
||||||
|
@ -1273,7 +1273,7 @@ def extractPlus7Stream(self, url):
|
||||||
r'<video lang="%s" ref="(http[^\'"&]*)' % video_lang,
|
r'<video lang="%s" ref="(http[^\'"&]*)' % video_lang,
|
||||||
0,
|
0,
|
||||||
[
|
[
|
||||||
(1, 'url', u'ERROR: Could not find <video> tag: %s' % url)
|
(1, 'url', u'Could not find <video> tag: %s' % url)
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
next_url = compat_urllib_parse.unquote(info.get('url'))
|
next_url = compat_urllib_parse.unquote(info.get('url'))
|
||||||
|
@ -1286,10 +1286,10 @@ def extractPlus7Stream(self, url):
|
||||||
'<url quality="hd">(.*?)</url>',
|
'<url quality="hd">(.*?)</url>',
|
||||||
re.DOTALL,
|
re.DOTALL,
|
||||||
[
|
[
|
||||||
(1, 'id', u'ERROR: could not extract video id: %s' % url),
|
(1, 'id', u'could not extract video id: %s' % url),
|
||||||
(2, 'title', u'ERROR: could not extract video title: %s' % url),
|
(2, 'title', u'could not extract video title: %s' % url),
|
||||||
(3, 'date', u'ERROR: could not extract video date: %s' % url),
|
(3, 'date', u'could not extract video date: %s' % url),
|
||||||
(4, 'url', u'ERROR: could not extract video url: %s' % url)
|
(4, 'url', u'could not extract video url: %s' % url)
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -1518,7 +1518,7 @@ def _get_n_results(self, query, n):
|
||||||
api_response = json.loads(data)['data']
|
api_response = json.loads(data)['data']
|
||||||
|
|
||||||
if not 'items' in api_response:
|
if not 'items' in api_response:
|
||||||
self._downloader.trouble(u'[youtube] No video results')
|
self._downloader.report_error(u'[youtube] No video results')
|
||||||
return
|
return
|
||||||
|
|
||||||
new_ids = list(video['id'] for video in api_response['items'])
|
new_ids = list(video['id'] for video in api_response['items'])
|
||||||
|
@ -2787,7 +2787,7 @@ def report_resolve(self, video_id):
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
if mobj is None:
|
if mobj is None:
|
||||||
self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
|
self._downloader.report_error(u'invalid URL: %s' % url)
|
||||||
return
|
return
|
||||||
|
|
||||||
# extract uploader (which is in the url)
|
# extract uploader (which is in the url)
|
||||||
|
@ -2805,14 +2805,14 @@ def _real_extract(self, url):
|
||||||
info_json_bytes = compat_urllib_request.urlopen(request).read()
|
info_json_bytes = compat_urllib_request.urlopen(request).read()
|
||||||
info_json = info_json_bytes.decode('utf-8')
|
info_json = info_json_bytes.decode('utf-8')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % compat_str(err))
|
self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err))
|
||||||
return
|
return
|
||||||
|
|
||||||
videos = []
|
videos = []
|
||||||
info = json.loads(info_json)
|
info = json.loads(info_json)
|
||||||
if 'errors' in info:
|
if 'errors' in info:
|
||||||
for err in info['errors']:
|
for err in info['errors']:
|
||||||
self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % compat_str(err['error_message']))
|
self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err['error_message']))
|
||||||
return
|
return
|
||||||
|
|
||||||
for track in info['tracks']:
|
for track in info['tracks']:
|
||||||
|
@ -2825,7 +2825,7 @@ def _real_extract(self, url):
|
||||||
stream_json_bytes = compat_urllib_request.urlopen(request).read()
|
stream_json_bytes = compat_urllib_request.urlopen(request).read()
|
||||||
stream_json = stream_json_bytes.decode('utf-8')
|
stream_json = stream_json_bytes.decode('utf-8')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
self._downloader.trouble(u'ERROR: unable to download stream definitions: %s' % compat_str(err))
|
self._downloader.report_error(u'unable to download stream definitions: %s' % compat_str(err))
|
||||||
return
|
return
|
||||||
|
|
||||||
streams = json.loads(stream_json)
|
streams = json.loads(stream_json)
|
||||||
|
@ -3173,7 +3173,7 @@ def _real_extract(self, url):
|
||||||
format = ext + '-' + rendition.attrib['width'] + 'x' + rendition.attrib['height'] + '_' + rendition.attrib['bitrate']
|
format = ext + '-' + rendition.attrib['width'] + 'x' + rendition.attrib['height'] + '_' + rendition.attrib['bitrate']
|
||||||
video_url = rendition.find('./src').text
|
video_url = rendition.find('./src').text
|
||||||
except KeyError:
|
except KeyError:
|
||||||
self._downloader.trouble('Invalid rendition field.')
|
self._downloader.report_error('Invalid rendition field.')
|
||||||
return
|
return
|
||||||
|
|
||||||
info = {
|
info = {
|
||||||
|
@ -3620,7 +3620,7 @@ def _real_extract(self, url):
|
||||||
if not m:
|
if not m:
|
||||||
m = re.search(r'<title>(?P<title>[^<]+?)</title>', webpage)
|
m = re.search(r'<title>(?P<title>[^<]+?)</title>', webpage)
|
||||||
if not m:
|
if not m:
|
||||||
self._downloader.trouble(u'Cannot find video title')
|
self._downloader.report_error(u'Cannot find video title')
|
||||||
title = clean_html(m.group('title'))
|
title = clean_html(m.group('title'))
|
||||||
|
|
||||||
m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage)
|
m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage)
|
||||||
|
@ -3726,7 +3726,7 @@ def _real_extract(self, url):
|
||||||
else:
|
else:
|
||||||
ext = 'flv'
|
ext = 'flv'
|
||||||
else:
|
else:
|
||||||
self._downloader.trouble(u'ERROR: Cannot find video url for %s' % video_id)
|
self._downloader.report_error(u'Cannot find video url for %s' % video_id)
|
||||||
return
|
return
|
||||||
|
|
||||||
_title = r"""<title>(.*)</title>"""
|
_title = r"""<title>(.*)</title>"""
|
||||||
|
@ -4246,7 +4246,7 @@ class LiveLeakIE(InfoExtractor):
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
mobj = re.match(self._VALID_URL, url)
|
mobj = re.match(self._VALID_URL, url)
|
||||||
if mobj is None:
|
if mobj is None:
|
||||||
self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
|
self._downloader.report_error(u'invalid URL: %s' % url)
|
||||||
return
|
return
|
||||||
|
|
||||||
video_id = mobj.group('video_id')
|
video_id = mobj.group('video_id')
|
||||||
|
@ -4261,7 +4261,7 @@ def _real_extract(self, url):
|
||||||
|
|
||||||
m = re.search(r'<meta property="og:title" content="(?P<title>.*?)"', webpage)
|
m = re.search(r'<meta property="og:title" content="(?P<title>.*?)"', webpage)
|
||||||
if not m:
|
if not m:
|
||||||
self._downloader.trouble(u'Cannot find video title')
|
self._downloader.report_error(u'Cannot find video title')
|
||||||
title = unescapeHTML(m.group('title')).replace('LiveLeak.com -', '').strip()
|
title = unescapeHTML(m.group('title')).replace('LiveLeak.com -', '').strip()
|
||||||
|
|
||||||
m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage)
|
m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage)
|
||||||
|
|
Loading…
Reference in a new issue