mirror of
https://github.com/SickGear/SickGear.git
synced 2024-12-11 21:53:37 +00:00
285 lines
9.7 KiB
Python
285 lines
9.7 KiB
Python
import time
|
|
import logging
|
|
|
|
from ..exceptions import (
|
|
ConnectTimeoutError,
|
|
MaxRetryError,
|
|
ProtocolError,
|
|
ReadTimeoutError,
|
|
ResponseError,
|
|
)
|
|
from ..packages import six
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
class Retry(object):
|
|
""" Retry configuration.
|
|
|
|
Each retry attempt will create a new Retry object with updated values, so
|
|
they can be safely reused.
|
|
|
|
Retries can be defined as a default for a pool::
|
|
|
|
retries = Retry(connect=5, read=2, redirect=5)
|
|
http = PoolManager(retries=retries)
|
|
response = http.request('GET', 'http://example.com/')
|
|
|
|
Or per-request (which overrides the default for the pool)::
|
|
|
|
response = http.request('GET', 'http://example.com/', retries=Retry(10))
|
|
|
|
Retries can be disabled by passing ``False``::
|
|
|
|
response = http.request('GET', 'http://example.com/', retries=False)
|
|
|
|
Errors will be wrapped in :class:`~urllib3.exceptions.MaxRetryError` unless
|
|
retries are disabled, in which case the causing exception will be raised.
|
|
|
|
:param int total:
|
|
Total number of retries to allow. Takes precedence over other counts.
|
|
|
|
Set to ``None`` to remove this constraint and fall back on other
|
|
counts. It's a good idea to set this to some sensibly-high value to
|
|
account for unexpected edge cases and avoid infinite retry loops.
|
|
|
|
Set to ``0`` to fail on the first retry.
|
|
|
|
Set to ``False`` to disable and imply ``raise_on_redirect=False``.
|
|
|
|
:param int connect:
|
|
How many connection-related errors to retry on.
|
|
|
|
These are errors raised before the request is sent to the remote server,
|
|
which we assume has not triggered the server to process the request.
|
|
|
|
Set to ``0`` to fail on the first retry of this type.
|
|
|
|
:param int read:
|
|
How many times to retry on read errors.
|
|
|
|
These errors are raised after the request was sent to the server, so the
|
|
request may have side-effects.
|
|
|
|
Set to ``0`` to fail on the first retry of this type.
|
|
|
|
:param int redirect:
|
|
How many redirects to perform. Limit this to avoid infinite redirect
|
|
loops.
|
|
|
|
A redirect is a HTTP response with a status code 301, 302, 303, 307 or
|
|
308.
|
|
|
|
Set to ``0`` to fail on the first retry of this type.
|
|
|
|
Set to ``False`` to disable and imply ``raise_on_redirect=False``.
|
|
|
|
:param iterable method_whitelist:
|
|
Set of uppercased HTTP method verbs that we should retry on.
|
|
|
|
By default, we only retry on methods which are considered to be
|
|
indempotent (multiple requests with the same parameters end with the
|
|
same state). See :attr:`Retry.DEFAULT_METHOD_WHITELIST`.
|
|
|
|
:param iterable status_forcelist:
|
|
A set of HTTP status codes that we should force a retry on.
|
|
|
|
By default, this is disabled with ``None``.
|
|
|
|
:param float backoff_factor:
|
|
A backoff factor to apply between attempts. urllib3 will sleep for::
|
|
|
|
{backoff factor} * (2 ^ ({number of total retries} - 1))
|
|
|
|
seconds. If the backoff_factor is 0.1, then :func:`.sleep` will sleep
|
|
for [0.1s, 0.2s, 0.4s, ...] between retries. It will never be longer
|
|
than :attr:`Retry.MAX_BACKOFF`.
|
|
|
|
By default, backoff is disabled (set to 0).
|
|
|
|
:param bool raise_on_redirect: Whether, if the number of redirects is
|
|
exhausted, to raise a MaxRetryError, or to return a response with a
|
|
response code in the 3xx range.
|
|
"""
|
|
|
|
DEFAULT_METHOD_WHITELIST = frozenset([
|
|
'HEAD', 'GET', 'PUT', 'DELETE', 'OPTIONS', 'TRACE'])
|
|
|
|
#: Maximum backoff time.
|
|
BACKOFF_MAX = 120
|
|
|
|
def __init__(self, total=10, connect=None, read=None, redirect=None,
|
|
method_whitelist=DEFAULT_METHOD_WHITELIST, status_forcelist=None,
|
|
backoff_factor=0, raise_on_redirect=True, _observed_errors=0):
|
|
|
|
self.total = total
|
|
self.connect = connect
|
|
self.read = read
|
|
|
|
if redirect is False or total is False:
|
|
redirect = 0
|
|
raise_on_redirect = False
|
|
|
|
self.redirect = redirect
|
|
self.status_forcelist = status_forcelist or set()
|
|
self.method_whitelist = method_whitelist
|
|
self.backoff_factor = backoff_factor
|
|
self.raise_on_redirect = raise_on_redirect
|
|
self._observed_errors = _observed_errors # TODO: use .history instead?
|
|
|
|
def new(self, **kw):
|
|
params = dict(
|
|
total=self.total,
|
|
connect=self.connect, read=self.read, redirect=self.redirect,
|
|
method_whitelist=self.method_whitelist,
|
|
status_forcelist=self.status_forcelist,
|
|
backoff_factor=self.backoff_factor,
|
|
raise_on_redirect=self.raise_on_redirect,
|
|
_observed_errors=self._observed_errors,
|
|
)
|
|
params.update(kw)
|
|
return type(self)(**params)
|
|
|
|
@classmethod
|
|
def from_int(cls, retries, redirect=True, default=None):
|
|
""" Backwards-compatibility for the old retries format."""
|
|
if retries is None:
|
|
retries = default if default is not None else cls.DEFAULT
|
|
|
|
if isinstance(retries, Retry):
|
|
return retries
|
|
|
|
redirect = bool(redirect) and None
|
|
new_retries = cls(retries, redirect=redirect)
|
|
log.debug("Converted retries value: %r -> %r" % (retries, new_retries))
|
|
return new_retries
|
|
|
|
def get_backoff_time(self):
|
|
""" Formula for computing the current backoff
|
|
|
|
:rtype: float
|
|
"""
|
|
if self._observed_errors <= 1:
|
|
return 0
|
|
|
|
backoff_value = self.backoff_factor * (2 ** (self._observed_errors - 1))
|
|
return min(self.BACKOFF_MAX, backoff_value)
|
|
|
|
def sleep(self):
|
|
""" Sleep between retry attempts using an exponential backoff.
|
|
|
|
By default, the backoff factor is 0 and this method will return
|
|
immediately.
|
|
"""
|
|
backoff = self.get_backoff_time()
|
|
if backoff <= 0:
|
|
return
|
|
time.sleep(backoff)
|
|
|
|
def _is_connection_error(self, err):
|
|
""" Errors when we're fairly sure that the server did not receive the
|
|
request, so it should be safe to retry.
|
|
"""
|
|
return isinstance(err, ConnectTimeoutError)
|
|
|
|
def _is_read_error(self, err):
|
|
""" Errors that occur after the request has been started, so we should
|
|
assume that the server began processing it.
|
|
"""
|
|
return isinstance(err, (ReadTimeoutError, ProtocolError))
|
|
|
|
def is_forced_retry(self, method, status_code):
|
|
""" Is this method/status code retryable? (Based on method/codes whitelists)
|
|
"""
|
|
if self.method_whitelist and method.upper() not in self.method_whitelist:
|
|
return False
|
|
|
|
return self.status_forcelist and status_code in self.status_forcelist
|
|
|
|
def is_exhausted(self):
|
|
""" Are we out of retries? """
|
|
retry_counts = (self.total, self.connect, self.read, self.redirect)
|
|
retry_counts = list(filter(None, retry_counts))
|
|
if not retry_counts:
|
|
return False
|
|
|
|
return min(retry_counts) < 0
|
|
|
|
def increment(self, method=None, url=None, response=None, error=None, _pool=None, _stacktrace=None):
|
|
""" Return a new Retry object with incremented retry counters.
|
|
|
|
:param response: A response object, or None, if the server did not
|
|
return a response.
|
|
:type response: :class:`~urllib3.response.HTTPResponse`
|
|
:param Exception error: An error encountered during the request, or
|
|
None if the response was received successfully.
|
|
|
|
:return: A new ``Retry`` object.
|
|
"""
|
|
if self.total is False and error:
|
|
# Disabled, indicate to re-raise the error.
|
|
raise six.reraise(type(error), error, _stacktrace)
|
|
|
|
total = self.total
|
|
if total is not None:
|
|
total -= 1
|
|
|
|
_observed_errors = self._observed_errors
|
|
connect = self.connect
|
|
read = self.read
|
|
redirect = self.redirect
|
|
cause = 'unknown'
|
|
|
|
if error and self._is_connection_error(error):
|
|
# Connect retry?
|
|
if connect is False:
|
|
raise six.reraise(type(error), error, _stacktrace)
|
|
elif connect is not None:
|
|
connect -= 1
|
|
_observed_errors += 1
|
|
|
|
elif error and self._is_read_error(error):
|
|
# Read retry?
|
|
if read is False:
|
|
raise six.reraise(type(error), error, _stacktrace)
|
|
elif read is not None:
|
|
read -= 1
|
|
_observed_errors += 1
|
|
|
|
elif response and response.get_redirect_location():
|
|
# Redirect retry?
|
|
if redirect is not None:
|
|
redirect -= 1
|
|
cause = 'too many redirects'
|
|
|
|
else:
|
|
# Incrementing because of a server error like a 500 in
|
|
# status_forcelist and a the given method is in the whitelist
|
|
_observed_errors += 1
|
|
cause = ResponseError.GENERIC_ERROR
|
|
if response and response.status:
|
|
cause = ResponseError.SPECIFIC_ERROR.format(
|
|
status_code=response.status)
|
|
|
|
new_retry = self.new(
|
|
total=total,
|
|
connect=connect, read=read, redirect=redirect,
|
|
_observed_errors=_observed_errors)
|
|
|
|
if new_retry.is_exhausted():
|
|
raise MaxRetryError(_pool, url, error or ResponseError(cause))
|
|
|
|
log.debug("Incremented Retry for (url='%s'): %r" % (url, new_retry))
|
|
|
|
return new_retry
|
|
|
|
|
|
def __repr__(self):
|
|
return ('{cls.__name__}(total={self.total}, connect={self.connect}, '
|
|
'read={self.read}, redirect={self.redirect})').format(
|
|
cls=type(self), self=self)
|
|
|
|
|
|
# For backwards compatibility (equivalent to pre-v1.9):
|
|
Retry.DEFAULT = Retry(3)
|