2023-01-12 01:04:47 +00:00
|
|
|
# encoding:utf-8
|
|
|
|
# ---------------
|
|
|
|
# functions are placed here to remove cyclic import issues from placement in helpers
|
|
|
|
#
|
|
|
|
import ast
|
|
|
|
import codecs
|
|
|
|
import datetime
|
|
|
|
import getpass
|
|
|
|
import hashlib
|
|
|
|
import io
|
|
|
|
import logging
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import socket
|
|
|
|
import stat
|
|
|
|
import subprocess
|
|
|
|
import sys
|
|
|
|
import tempfile
|
|
|
|
import threading
|
|
|
|
import time
|
|
|
|
import traceback
|
|
|
|
import unicodedata
|
|
|
|
|
|
|
|
from exceptions_helper import ex, ConnectionSkipException
|
|
|
|
from json_helper import json_loads
|
2023-01-06 11:47:44 +00:00
|
|
|
from cachecontrol import CacheControl, caches
|
Add try_date for use in attempting conversion of unspecified date format types birthdate, deathdate, aired.
Change tweaks, remove one warn, order imports.
Change tidy up of other warnings, mismatched types, typos, a strange arg issue with deepcopy, and PEP8.
Bug fix, bad key in get_item for TVInfoSocialIDs.
Fix ambiguities of `show` used in sg versus external uses.
Change add cast, crew type mappings
Change only take Main Actors, Hosts, Interviewers, Presenters.
2021-09-05 19:10:58 +00:00
|
|
|
from lib.dateutil.parser import parser
|
2023-01-12 01:04:47 +00:00
|
|
|
# from lib.tmdbsimple.configuration import Configuration
|
|
|
|
# from lib.tmdbsimple.genres import Genres
|
|
|
|
from cfscrape import CloudflareScraper
|
|
|
|
from send2trash import send2trash
|
|
|
|
|
|
|
|
# noinspection PyPep8Naming
|
2023-02-10 14:15:50 +00:00
|
|
|
from encodingKludge import SYS_ENCODING
|
2023-01-12 01:04:47 +00:00
|
|
|
import requests
|
|
|
|
|
2023-02-11 18:02:58 +00:00
|
|
|
from _23 import decode_bytes, html_unescape, list_range, \
|
|
|
|
Popen, scandir, urlparse, urlsplit, urlunparse
|
2023-01-12 01:04:47 +00:00
|
|
|
from six import integer_types, iteritems, iterkeys, itervalues, moves, PY2, string_types, text_type
|
|
|
|
|
2023-09-08 12:07:06 +00:00
|
|
|
ACCEPT_ENCODING = "gzip,deflate"
|
|
|
|
try:
|
|
|
|
try:
|
|
|
|
import brotlicffi as _unused_module_brotli # noqa: F401
|
|
|
|
except ImportError:
|
|
|
|
import brotli as _unused_module_brotli # noqa: F401
|
|
|
|
except ImportError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
ACCEPT_ENCODING += ",br"
|
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
import zipfile
|
|
|
|
# py7z hardwired removed, see comment below
|
|
|
|
py7zr = None
|
Add try_date for use in attempting conversion of unspecified date format types birthdate, deathdate, aired.
Change tweaks, remove one warn, order imports.
Change tidy up of other warnings, mismatched types, typos, a strange arg issue with deepcopy, and PEP8.
Bug fix, bad key in get_item for TVInfoSocialIDs.
Fix ambiguities of `show` used in sg versus external uses.
Change add cast, crew type mappings
Change only take Main Actors, Hosts, Interviewers, Presenters.
2021-09-05 19:10:58 +00:00
|
|
|
tz_p = parser()
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
# noinspection PyUnreachableCode
|
|
|
|
if False:
|
|
|
|
from _23 import DirEntry
|
|
|
|
from lxml_etree import etree
|
|
|
|
try:
|
|
|
|
# py7z hardwired removed because Python 3.9 interpreter crashes with a process kill signal 9 when memory is
|
|
|
|
# low/exhausted during a native 7z compress action on Linux. Therefore, the native functions cannot be trusted.
|
|
|
|
# `import` moved to this non-runtime scope to preserve code resolution in case reinstated at a later PY release
|
|
|
|
# noinspection PyUnresolvedReferences,PyPackageRequirements
|
|
|
|
import py7zr
|
|
|
|
except ImportError:
|
|
|
|
py7zr = None
|
|
|
|
# sickgear is strictly used here for resolution, this is only possible because
|
|
|
|
# this section is not used at runtime which would create circular reference issues
|
|
|
|
# noinspection PyPep8Naming
|
|
|
|
from sickgear import db, notifiers as NOTIFIERS
|
|
|
|
# noinspection PyUnresolvedReferences
|
|
|
|
from typing import Any, AnyStr, Dict, Generator, NoReturn, integer_types, Iterable, Iterator, List, Optional, \
|
Change improve tmdb_api, trakt_api, and TVInfoShow object.
Add `spoken_languages` to tmdb API and TVInfoShow object.
Add `trailers`, `homepage` to trakt API and TVInfoShow object.
Add trakt episode data if returned from api.
Add trakt API methods.
- get_most_played
- get_most_watched
- get_most_collected
- get_anticipated
- get_recommended
- get_trending
- get_popular
- get_recommended_for_account
- get_new_shows
- get_new_seasons
- get_watchlisted_for_account
- get_similar
- hide_recommended_for_account (to hide/remove recommended shows for account)
- unhide_recommended_for_account
- list_hidden_recommended_for_account
Fix caching tmdb language list over different runtime instances.
Add episode_count and fix ti_show in tmdb_api person object.
Change set additional properties in get_person trakt_api.
Add tmdb API methods and tvinfo_base.
- get_recommended_for_show
- get_similar
---
fix supported language caching
improve print output (source name) of tvinfo_api_tests
fix tvinfo_api_tests data creation
---
Add code so that it runs with all_test
use mock today() and now() dates
add option to only get new urls mock data
try also to make object creation only when needed
fix person parser in tmdb_api
add search_person test in tvinfo_api_tests
restore mocked methods at the end of the tvinfo_api_tests to prevent other tests to fail when called via all_tests
switch gzip with better lzma compression for mock files (default lib in py3)
move mock files in test unit sub folder
---
Fix trakt method `get_recommended`.
Fix browse trakt tests in tvinfo_api_tests.
Change set episode id in trakt api.
---
Add test_browse_endpoints to tvinfo_api_tests.
---
Add enforce_type to sg_helpers.
Change use enforce str for overviews.
Change remove `if PY2` code sections
Add support for datetime.time in _make_airtime in tv.py
Refactor tvmaze_api show data setter.
Change test to not allow None for seriesname.
Add additional missing showdata with caller load_data().
Add load_data() to TVInfoShow.
Add guestcast, guestcrew to episodes in pytvmaze lib.
---
Change make seriesid of TVInfoShow a alias property of id.
Add tvinfo tests.
Add search tests.
Add show, person tests.
Change add trakt tests.
Change add tmdb search tests.
tvmaze_api exclude rating from mapping.
Allow None for seriesname.
Fix origin_countries in trakt_api search.
Fix show_type in tvmaze_api.
Fix airtime for episodes in tvmaze_api.
---
Change switch to property instead of legacy dict-like use for trakt search results.
Change optimize speed of get() function.
Fix make BaseTVinfoSeasonnotfound and BaseTVinfoAttributenotfound also a subclass of AttributeError and KeyError.
Change mock get() to work with and without default args just like dict get().
Change add language to tmdb_api search results.
Change improve person search by remote id, by getting the complete persons data when there is only 1 result.
Change trakt API search results to tvinfoshow.
Change search results to TVInfoShow objs in tvmaze_api.
Change simplify poster URL generation for search results.
Change search results to TVInfoShow objs.
Change add tvdb genre links to displayShow.
Change workaround for missing data in person data (series set to None).
Fix add show to characters of person if there is no name on IMDb (set to 'unknown name').
Change add config and icons for linkedin, reddit, wikidata, youtube.
Add TVInfoIDs, TVInfoSocialIDs to Trakt.
Add TVInfoIDs to tmdb_api.
Add TVInfoIDs to tvmaze_api.
add TVInfoIDs to imdb_api.
Change make character name '' if None.
Fix for 'unknown name' persons and characters.
Add contentrating.
Change fill in new fields to get_person results.
----
Change set new in/active dates to network.
Change add active_date, inactive_date to TVInfoNetwork class.
Change add default kwargs to tmdb discover method if no kwargs are set.
Change default: English language shows with first air date greater then today.
Change add slug field to returned data from discover.
Change add 'score' mapped to rating to discover returned results.
Fix valid_data for discover method.
Change add result_count to discover.
Change add _sanitise_image_uri to discover method.
Fix convert_person.
Change add missing _sanitise_image_uri for images in some places.
Fix crew.
Change return type of tvinfo base: discover to list tvinfoshow.
Fix people remote id search.
Change add tmdb person id search.
Change fix people endpoint fieldname changes.
Change add biography to person object.
Change move 401 expired token handling into TvdbAuth class.
Change get new token if old token is expired.
Change add raise error if episodes fallback fails to load data.
Change add break if no valid_data to absolute and alternative numberings.
Change add filter only networks.
Change add new required parameter meta=translations to get translated (includes the original language) show overviews.
Change add check if show is set for person compare.
Fix person update properties with no show set.
Change add person image.
Change add alternative episode orders.
Change add alt_ep_numbering to TVINFO_Show.
Change add old interface for dvd order.
Change add trakt slug tvinfo search test cases.
Change add mock for old tvdb get new token.
Change old lib to newer tvinfo data.
Fix person id (not available on old api).
Change more places to new TVInfoAPI interface.
2021-09-16 20:03:19 +00:00
|
|
|
Tuple, Type, Union
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
html_convert_fractions = {0: '', 25: '¼', 50: '½', 75: '¾', 100: 1}
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
PROG_DIR = os.path.join(os.path.dirname(os.path.normpath(os.path.abspath(__file__))), '..')
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
# Mapping error status codes to official W3C names
|
|
|
|
http_error_code = {
|
|
|
|
300: 'Multiple Choices',
|
|
|
|
301: 'Moved Permanently',
|
|
|
|
302: 'Found',
|
|
|
|
303: 'See Other',
|
|
|
|
304: 'Not Modified',
|
|
|
|
305: 'Use Proxy',
|
|
|
|
307: 'Temporary Redirect',
|
|
|
|
308: 'Permanent Redirect',
|
|
|
|
400: 'Bad Request',
|
|
|
|
401: 'Unauthorized',
|
|
|
|
402: 'Payment Required',
|
|
|
|
403: 'Forbidden',
|
|
|
|
404: 'Not Found',
|
|
|
|
405: 'Method Not Allowed',
|
|
|
|
406: 'Not Acceptable',
|
|
|
|
407: 'Proxy Authentication Required',
|
|
|
|
408: 'Request Timeout',
|
|
|
|
409: 'Conflict',
|
|
|
|
410: 'Gone',
|
|
|
|
411: 'Length Required',
|
|
|
|
412: 'Precondition Failed',
|
|
|
|
413: 'Request Entity Too Large',
|
|
|
|
414: 'Request-URI Too Long',
|
|
|
|
415: 'Unsupported Media Type',
|
|
|
|
416: 'Requested Range Not Satisfiable',
|
|
|
|
417: 'Expectation Failed',
|
|
|
|
429: 'Too Many Requests',
|
|
|
|
431: 'Request Header Fields Too Large',
|
|
|
|
444: 'No Response',
|
|
|
|
451: 'Unavailable For Legal Reasons',
|
|
|
|
500: 'Internal Server Error',
|
|
|
|
501: 'Not Implemented',
|
|
|
|
502: 'Bad Gateway',
|
|
|
|
503: 'Service Unavailable',
|
|
|
|
504: 'Gateway Timeout',
|
|
|
|
505: 'HTTP Version Not Supported',
|
|
|
|
511: 'Network Authentication Required'}
|
|
|
|
|
|
|
|
logger = logging.getLogger('sg.helper')
|
|
|
|
logger.addHandler(logging.NullHandler())
|
|
|
|
|
|
|
|
USER_AGENT = ''
|
|
|
|
CACHE_DIR = None
|
|
|
|
DATA_DIR = None
|
|
|
|
PROXY_SETTING = None
|
|
|
|
TRASH_REMOVE_SHOW = False
|
|
|
|
REMOVE_FILENAME_CHARS = None
|
|
|
|
MEMCACHE = {}
|
|
|
|
FLARESOLVERR_HOST = None
|
|
|
|
|
|
|
|
# noinspection PyRedeclaration
|
|
|
|
db = None
|
|
|
|
# noinspection PyRedeclaration
|
|
|
|
NOTIFIERS = None
|
|
|
|
|
|
|
|
|
|
|
|
class ConnectionFailTypes(object):
|
|
|
|
http = 1
|
|
|
|
connection = 2
|
|
|
|
connection_timeout = 3
|
|
|
|
timeout = 4
|
|
|
|
other = 5
|
|
|
|
limit = 6
|
|
|
|
nodata = 7
|
|
|
|
|
|
|
|
names = {http: 'http', timeout: 'timeout',
|
|
|
|
connection: 'connection', connection_timeout: 'connection_timeout',
|
|
|
|
nodata: 'nodata', other: 'other', limit: 'limit'}
|
|
|
|
|
|
|
|
def __init__(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class ConnectionFail(object):
|
|
|
|
def __init__(self, fail_type=ConnectionFailTypes.other, code=None, fail_time=None):
|
|
|
|
self.code = code
|
|
|
|
self.fail_type = fail_type
|
|
|
|
self.fail_time = (datetime.datetime.now(), fail_time)[isinstance(fail_time, datetime.datetime)]
|
|
|
|
|
|
|
|
|
|
|
|
class ConnectionFailDict(object):
|
|
|
|
def __init__(self):
|
|
|
|
self.domain_list = {} # type: Dict[AnyStr, ConnectionFailList]
|
|
|
|
self.lock = threading.RLock()
|
|
|
|
self.load_from_db()
|
|
|
|
|
|
|
|
def load_from_db(self):
|
|
|
|
if None is not db:
|
|
|
|
with self.lock:
|
|
|
|
my_db = db.DBConnection('cache.db')
|
2023-02-24 11:46:07 +00:00
|
|
|
if my_db.has_table('connection_fails'):
|
2023-01-12 01:04:47 +00:00
|
|
|
domains = my_db.select('SELECT DISTINCT domain_url from connection_fails')
|
|
|
|
for domain in domains:
|
|
|
|
self.domain_list[domain['domain_url']] = ConnectionFailList(domain['domain_url'])
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def get_domain(url):
|
|
|
|
# type: (AnyStr) -> Optional[AnyStr]
|
|
|
|
try:
|
|
|
|
return urlsplit(url).hostname.lower()
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def add_failure(self, url, fail_type):
|
|
|
|
# type: (AnyStr, ConnectionFail) -> None
|
|
|
|
host = self.get_domain(url)
|
|
|
|
if None is not host:
|
|
|
|
with self.lock:
|
|
|
|
self.domain_list.setdefault(host, ConnectionFailList(host)).add_fail(fail_type)
|
|
|
|
|
|
|
|
def inc_failure_count(self,
|
|
|
|
url, # type: AnyStr
|
|
|
|
*args, **kwargs):
|
|
|
|
host = self.get_domain(url)
|
|
|
|
if None is not host:
|
|
|
|
with self.lock:
|
|
|
|
if host in self.domain_list:
|
|
|
|
domain = self.domain_list[host]
|
|
|
|
fail_type = ('fail_type' in kwargs and kwargs['fail_type'].fail_type) or \
|
|
|
|
(isinstance(args, tuple) and isinstance(args[0], ConnectionFail) and args[0].fail_type)
|
|
|
|
# noinspection PyProtectedMember
|
|
|
|
if not isinstance(domain.failure_time, datetime.datetime) or \
|
|
|
|
fail_type != domain._last_fail_type or \
|
|
|
|
domain.fail_newest_delta() > datetime.timedelta(seconds=3):
|
|
|
|
domain.failure_count += 1
|
|
|
|
domain.failure_time = datetime.datetime.now()
|
|
|
|
domain._last_fail_type = fail_type
|
|
|
|
domain.add_fail(*args, **kwargs)
|
|
|
|
else:
|
|
|
|
logger.debug('%s: Not logging same failure within 3 seconds' % url)
|
|
|
|
|
|
|
|
def should_skip(self, url, log_warning=True, use_tmr_limit=True):
|
|
|
|
# type: (AnyStr, bool, bool) -> bool
|
|
|
|
host = self.get_domain(url)
|
|
|
|
if None is not host:
|
|
|
|
with self.lock:
|
|
|
|
if host in self.domain_list:
|
|
|
|
return self.domain_list[host].should_skip(log_warning=log_warning, use_tmr_limit=use_tmr_limit)
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
|
|
|
DOMAIN_FAILURES = ConnectionFailDict()
|
|
|
|
sp = 8
|
|
|
|
trakt_fail_times = {(i * sp) + m: s for m in range(1, 1 + sp) for i, s in
|
|
|
|
enumerate([(0, 5), (0, 15), (0, 30), (1, 0), (2, 0)])}
|
|
|
|
trakt_fail_times.update({i: s for i, s in enumerate([(3, 0), (6, 0), (12, 0), (24, 0)], len(trakt_fail_times))})
|
|
|
|
domain_fail_times = {'api.trakt.tv': trakt_fail_times}
|
|
|
|
default_fail_times = {1: (0, 15), 2: (0, 30), 3: (1, 0), 4: (2, 0), 5: (3, 0), 6: (6, 0), 7: (12, 0), 8: (24, 0)}
|
|
|
|
|
|
|
|
|
|
|
|
class ConnectionFailList(object):
|
|
|
|
def __init__(self, url):
|
|
|
|
# type: (AnyStr) -> None
|
|
|
|
self.url = url
|
|
|
|
self._fails = [] # type: List[ConnectionFail]
|
|
|
|
self.lock = threading.Lock()
|
|
|
|
self.clear_old()
|
|
|
|
self.load_list()
|
|
|
|
self.last_save = datetime.datetime.now() # type: datetime.datetime
|
|
|
|
self._failure_count = 0 # type: int
|
|
|
|
self._failure_time = None # type: Optional[datetime.datetime]
|
|
|
|
self._tmr_limit_count = 0 # type: int
|
|
|
|
self._tmr_limit_time = None # type: Optional[datetime.datetime]
|
|
|
|
self._tmr_limit_wait = None # type: Optional[datetime.timedelta]
|
|
|
|
self._last_fail_type = None # type: Optional[ConnectionFail]
|
|
|
|
self.has_limit = False # type: bool
|
|
|
|
self.fail_times = domain_fail_times.get(url, default_fail_times) # type: Dict[integer_types, Tuple[int, int]]
|
|
|
|
self._load_fail_values()
|
|
|
|
self.dirty = False # type: bool
|
|
|
|
|
|
|
|
@property
|
|
|
|
def failure_time(self):
|
|
|
|
# type: (...) -> Union[None, datetime.datetime]
|
|
|
|
return self._failure_time
|
|
|
|
|
|
|
|
@failure_time.setter
|
|
|
|
def failure_time(self, value):
|
|
|
|
if None is value or isinstance(value, datetime.datetime):
|
|
|
|
changed_val = self._failure_time != value
|
|
|
|
self._failure_time = value
|
|
|
|
if changed_val:
|
|
|
|
# noinspection PyCallByClass,PyTypeChecker
|
|
|
|
self._save_fail_value('failure_time', (_totimestamp(value), value)[None is value])
|
|
|
|
|
|
|
|
@property
|
|
|
|
def tmr_limit_count(self):
|
|
|
|
# type: (...) -> int
|
|
|
|
return self._tmr_limit_count
|
|
|
|
|
|
|
|
@tmr_limit_count.setter
|
|
|
|
def tmr_limit_count(self, value):
|
|
|
|
changed_val = self._tmr_limit_count != value
|
|
|
|
self._tmr_limit_count = value
|
|
|
|
if changed_val:
|
|
|
|
self._save_fail_value('tmr_limit_count', value)
|
|
|
|
|
|
|
|
def tmr_limit_update(self, period, unit, desc):
|
|
|
|
# type: (Optional[AnyStr], Optional[AnyStr], AnyStr) -> None
|
|
|
|
self.tmr_limit_time = datetime.datetime.now()
|
|
|
|
self.tmr_limit_count += 1
|
|
|
|
limit_set = False
|
|
|
|
if None not in (period, unit):
|
|
|
|
limit_set = True
|
|
|
|
if unit in ('s', 'sec', 'secs', 'seconds', 'second'):
|
|
|
|
self.tmr_limit_wait = datetime.timedelta(seconds=try_int(period))
|
|
|
|
elif unit in ('m', 'min', 'mins', 'minutes', 'minute'):
|
|
|
|
self.tmr_limit_wait = datetime.timedelta(minutes=try_int(period))
|
|
|
|
elif unit in ('h', 'hr', 'hrs', 'hours', 'hour'):
|
|
|
|
self.tmr_limit_wait = datetime.timedelta(hours=try_int(period))
|
|
|
|
elif unit in ('d', 'days', 'day'):
|
|
|
|
self.tmr_limit_wait = datetime.timedelta(days=try_int(period))
|
|
|
|
else:
|
|
|
|
limit_set = False
|
|
|
|
if not limit_set:
|
|
|
|
time_index = self.fail_time_index(base_limit=0)
|
|
|
|
self.tmr_limit_wait = self.wait_time(time_index)
|
|
|
|
logger.warning('Request limit reached. Waiting for %s until next retry. Message: %s' %
|
|
|
|
(self.tmr_limit_wait, desc or 'none found'))
|
|
|
|
|
|
|
|
@property
|
|
|
|
def tmr_limit_time(self):
|
|
|
|
# type: (...) -> Union[None, datetime.datetime]
|
|
|
|
return self._tmr_limit_time
|
|
|
|
|
|
|
|
@tmr_limit_time.setter
|
|
|
|
def tmr_limit_time(self, value):
|
|
|
|
if None is value or isinstance(value, datetime.datetime):
|
|
|
|
changed_val = self._tmr_limit_time != value
|
|
|
|
self._tmr_limit_time = value
|
|
|
|
if changed_val:
|
|
|
|
# noinspection PyCallByClass,PyTypeChecker
|
|
|
|
self._save_fail_value('tmr_limit_time', (_totimestamp(value), value)[None is value])
|
|
|
|
|
|
|
|
@property
|
|
|
|
def last_fail(self):
|
|
|
|
# type: (...) -> Optional[int]
|
|
|
|
try:
|
|
|
|
return sorted(self.fails, key=lambda x: x.fail_time, reverse=True)[0].fail_type
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
|
|
|
|
@property
|
|
|
|
def failure_count(self):
|
|
|
|
# type: (...) -> int
|
|
|
|
return self._failure_count
|
|
|
|
|
|
|
|
@failure_count.setter
|
|
|
|
def failure_count(self, value):
|
|
|
|
changed_val = self._failure_count != value
|
|
|
|
self._failure_count = value
|
|
|
|
if changed_val:
|
|
|
|
self._save_fail_value('failure_count', value)
|
|
|
|
|
|
|
|
def is_waiting(self):
|
|
|
|
# type: (...) -> bool
|
|
|
|
return self.fail_newest_delta() < self.wait_time()
|
|
|
|
|
|
|
|
@property
|
|
|
|
def max_index(self):
|
|
|
|
# type: (...) -> int
|
|
|
|
return len(self.fail_times)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def tmr_limit_wait(self):
|
|
|
|
# type: (...) -> Optional[datetime.timedelta]
|
|
|
|
return self._tmr_limit_wait
|
|
|
|
|
|
|
|
@tmr_limit_wait.setter
|
|
|
|
def tmr_limit_wait(self, value):
|
|
|
|
if isinstance(getattr(self, 'fails', None), ConnectionFailList) and isinstance(value, datetime.timedelta):
|
|
|
|
self.add_fail(ConnectionFail(fail_type=ConnectionFailTypes.limit))
|
|
|
|
changed_val = self._tmr_limit_wait != value
|
|
|
|
self._tmr_limit_wait = value
|
|
|
|
if changed_val:
|
|
|
|
if None is value:
|
|
|
|
self._save_fail_value('tmr_limit_wait', value)
|
|
|
|
elif isinstance(value, datetime.timedelta):
|
|
|
|
self._save_fail_value('tmr_limit_wait', value.total_seconds())
|
|
|
|
|
|
|
|
def fail_time_index(self, base_limit=2):
|
|
|
|
# type: (int) -> int
|
|
|
|
i = max(self.failure_count - base_limit, 1)
|
|
|
|
if i not in self.fail_times:
|
|
|
|
i = list(self.fail_times)[-1]
|
|
|
|
return (i, self.max_index)[i >= self.max_index]
|
|
|
|
|
|
|
|
def valid_tmr_time(self):
|
|
|
|
# type: (...) -> bool
|
|
|
|
return isinstance(self.tmr_limit_wait, datetime.timedelta) and \
|
|
|
|
isinstance(self.tmr_limit_time, datetime.datetime)
|
|
|
|
|
|
|
|
def wait_time(self, time_index=None):
|
|
|
|
# type: (Optional[int]) -> datetime.timedelta
|
|
|
|
"""
|
|
|
|
Return a suitable wait time, selected by parameter, or based on the current failure count
|
|
|
|
|
|
|
|
:param time_index: A key value index into the fail_times dict, or selects using failure count if None
|
|
|
|
:return: Time
|
|
|
|
"""
|
|
|
|
if None is time_index:
|
|
|
|
time_index = self.fail_time_index()
|
|
|
|
return datetime.timedelta(hours=self.fail_times[time_index][0], minutes=self.fail_times[time_index][1])
|
|
|
|
|
|
|
|
def fail_newest_delta(self):
|
|
|
|
# type: (...) -> datetime.timedelta
|
|
|
|
"""
|
|
|
|
Return how long since most recent failure
|
|
|
|
:return: Period since most recent failure on record
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
return datetime.datetime.now() - self.failure_time
|
|
|
|
except (BaseException, Exception):
|
|
|
|
return datetime.timedelta(days=1000)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def get_next_try_time(self):
|
|
|
|
# type: (...) -> datetime.timedelta
|
|
|
|
n = None
|
|
|
|
h = datetime.timedelta(seconds=0)
|
|
|
|
f = datetime.timedelta(seconds=0)
|
|
|
|
if self.valid_tmr_time():
|
|
|
|
h = self.tmr_limit_time + self.tmr_limit_wait - datetime.datetime.now()
|
|
|
|
if 3 <= self.failure_count and isinstance(self.failure_time, datetime.datetime) and self.is_waiting():
|
|
|
|
h = self.failure_time + self.wait_time() - datetime.datetime.now()
|
|
|
|
if datetime.timedelta(seconds=0) < max((h, f)):
|
|
|
|
n = max((h, f))
|
|
|
|
return n
|
|
|
|
|
|
|
|
def retry_next(self):
|
|
|
|
if self.valid_tmr_time():
|
|
|
|
self.tmr_limit_time = datetime.datetime.now() - self.tmr_limit_wait
|
|
|
|
if 3 <= self.failure_count and isinstance(self.failure_time, datetime.datetime) and self.is_waiting():
|
|
|
|
self.failure_time = datetime.datetime.now() - self.wait_time()
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def fmt_delta(delta):
|
|
|
|
# type: (Union[datetime.datetime, datetime.timedelta]) -> AnyStr
|
|
|
|
return str(delta).rsplit('.')[0]
|
|
|
|
|
|
|
|
def should_skip(self, log_warning=True, use_tmr_limit=True):
|
|
|
|
# type: (bool, bool) -> bool
|
|
|
|
"""
|
|
|
|
Determine if a subsequent server request should be skipped. The result of this logic is based on most recent
|
|
|
|
server connection activity including, exhausted request limits, and counting connect failures to determine a
|
|
|
|
"cool down" period before recommending reconnection attempts; by returning False.
|
|
|
|
:param log_warning: Output to log if True (default) otherwise set False for no output.
|
|
|
|
:param use_tmr_limit: Setting this to False will ignore a tmr limit being reached and will instead return False.
|
|
|
|
:return: True for any known issue that would prevent a subsequent server connection, otherwise False.
|
|
|
|
"""
|
|
|
|
if self.valid_tmr_time():
|
|
|
|
time_left = self.tmr_limit_time + self.tmr_limit_wait - datetime.datetime.now()
|
|
|
|
if time_left > datetime.timedelta(seconds=0):
|
|
|
|
if log_warning:
|
|
|
|
logger.warning('%sToo many requests reached at %s, waiting for %s' % (
|
|
|
|
self.url, self.fmt_delta(self.tmr_limit_time), self.fmt_delta(time_left)))
|
|
|
|
return use_tmr_limit
|
|
|
|
else:
|
|
|
|
self.tmr_limit_time = None
|
|
|
|
self.tmr_limit_wait = None
|
|
|
|
if 3 <= self.failure_count:
|
|
|
|
if None is self.failure_time:
|
|
|
|
self.failure_time = datetime.datetime.now()
|
|
|
|
if self.is_waiting():
|
|
|
|
if log_warning:
|
|
|
|
time_left = self.wait_time() - self.fail_newest_delta()
|
|
|
|
logger.warning('Failed %s times, skipping domain %s for %s, '
|
|
|
|
'last failure at %s with fail type: %s' %
|
|
|
|
(self.failure_count, self.url, self.fmt_delta(time_left),
|
|
|
|
self.fmt_delta(self.failure_time), ConnectionFailTypes.names.get(
|
|
|
|
self.last_fail, ConnectionFailTypes.names[ConnectionFailTypes.other])))
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
@property
|
|
|
|
def fails(self):
|
|
|
|
# type: (...) -> List
|
|
|
|
return self._fails
|
|
|
|
|
|
|
|
@property
|
|
|
|
def fails_sorted(self):
|
|
|
|
# type: (...) -> List
|
|
|
|
fail_dict = {}
|
|
|
|
b_d = {'count': 0}
|
|
|
|
for e in self._fails:
|
|
|
|
fail_date = e.fail_time.date()
|
|
|
|
fail_hour = e.fail_time.time().hour
|
|
|
|
date_time = datetime.datetime.combine(fail_date, datetime.time(hour=fail_hour))
|
|
|
|
if ConnectionFailTypes.names[e.fail_type] not in fail_dict.get(date_time, {}):
|
|
|
|
default = {'date': str(fail_date), 'date_time': date_time,
|
|
|
|
'timestamp': try_int(_totimestamp(e.fail_time)), 'multirow': False}
|
|
|
|
for et in itervalues(ConnectionFailTypes.names):
|
|
|
|
default[et] = b_d.copy()
|
|
|
|
fail_dict.setdefault(date_time, default)[ConnectionFailTypes.names[e.fail_type]]['count'] = 1
|
|
|
|
else:
|
|
|
|
fail_dict[date_time][ConnectionFailTypes.names[e.fail_type]]['count'] += 1
|
|
|
|
if ConnectionFailTypes.http == e.fail_type:
|
|
|
|
if e.code in fail_dict[date_time].get(ConnectionFailTypes.names[e.fail_type],
|
|
|
|
{'code': {}}).get('code', {}):
|
|
|
|
fail_dict[date_time][ConnectionFailTypes.names[e.fail_type]]['code'][e.code] += 1
|
|
|
|
else:
|
|
|
|
fail_dict[date_time][ConnectionFailTypes.names[e.fail_type]].setdefault('code', {})[e.code] = 1
|
|
|
|
|
|
|
|
row_count = {}
|
|
|
|
for (k, v) in iteritems(fail_dict):
|
|
|
|
row_count.setdefault(v.get('date'), 0)
|
|
|
|
if v.get('date') in row_count:
|
|
|
|
row_count[v.get('date')] += 1
|
|
|
|
for (k, v) in iteritems(fail_dict):
|
|
|
|
if 1 < row_count.get(v.get('date')):
|
|
|
|
fail_dict[k]['multirow'] = True
|
|
|
|
|
|
|
|
fail_list = sorted([fail_dict[k] for k in iterkeys(fail_dict)], key=lambda y: y.get('date_time'), reverse=True)
|
|
|
|
|
|
|
|
totals = {}
|
|
|
|
for fail_date in set([fail.get('date') for fail in fail_list]):
|
|
|
|
daytotals = {}
|
|
|
|
for et in itervalues(ConnectionFailTypes.names):
|
|
|
|
daytotals.update({et: sum([x.get(et).get('count') for x in fail_list if fail_date == x.get('date')])})
|
|
|
|
totals.update({fail_date: daytotals})
|
|
|
|
for (fail_date, total) in iteritems(totals):
|
|
|
|
for i, item in enumerate(fail_list):
|
|
|
|
if fail_date == item.get('date'):
|
|
|
|
if item.get('multirow'):
|
|
|
|
fail_list[i:i] = [item.copy()]
|
|
|
|
for et in itervalues(ConnectionFailTypes.names):
|
|
|
|
fail_list[i][et] = {'count': total[et]}
|
|
|
|
if et == ConnectionFailTypes.names[ConnectionFailTypes.http]:
|
|
|
|
fail_list[i][et]['code'] = {}
|
|
|
|
break
|
|
|
|
|
|
|
|
return fail_list
|
|
|
|
|
|
|
|
def add_fail(self,
|
|
|
|
fail # type: ConnectionFail
|
|
|
|
):
|
|
|
|
if isinstance(fail, ConnectionFail):
|
|
|
|
with self.lock:
|
|
|
|
self.dirty = True
|
|
|
|
self._fails.append(fail)
|
|
|
|
logger.debug('Adding fail.%s for %s' % (ConnectionFailTypes.names.get(
|
|
|
|
fail.fail_type, ConnectionFailTypes.names[ConnectionFailTypes.other]), self.url))
|
|
|
|
self.save_list()
|
|
|
|
|
|
|
|
def _load_fail_values(self):
|
|
|
|
if None is not DATA_DIR:
|
|
|
|
my_db = db.DBConnection('cache.db')
|
2023-02-24 11:46:07 +00:00
|
|
|
if my_db.has_table('connection_fails_count'):
|
2023-01-12 01:04:47 +00:00
|
|
|
r = my_db.select('SELECT * FROM connection_fails_count WHERE domain_url = ?', [self.url])
|
|
|
|
if r:
|
|
|
|
self._failure_count = try_int(r[0]['failure_count'], 0)
|
|
|
|
if r[0]['failure_time']:
|
|
|
|
self._failure_time = datetime.datetime.fromtimestamp(r[0]['failure_time'])
|
|
|
|
else:
|
|
|
|
self._failure_time = None
|
|
|
|
self._tmr_limit_count = try_int(r[0]['tmr_limit_count'], 0)
|
|
|
|
if r[0]['tmr_limit_time']:
|
|
|
|
self._tmr_limit_time = datetime.datetime.fromtimestamp(r[0]['tmr_limit_time'])
|
|
|
|
else:
|
|
|
|
self._tmr_limit_time = None
|
|
|
|
if r[0]['tmr_limit_wait']:
|
|
|
|
self._tmr_limit_wait = datetime.timedelta(seconds=try_int(r[0]['tmr_limit_wait'], 0))
|
|
|
|
else:
|
|
|
|
self._tmr_limit_wait = None
|
|
|
|
self._last_fail_type = self.last_fail
|
|
|
|
|
|
|
|
def _save_fail_value(self, field, value):
|
|
|
|
my_db = db.DBConnection('cache.db')
|
2023-02-24 11:46:07 +00:00
|
|
|
if my_db.has_table('connection_fails_count'):
|
2023-01-12 01:04:47 +00:00
|
|
|
r = my_db.action('UPDATE connection_fails_count SET %s = ? WHERE domain_url = ?' % field,
|
|
|
|
[value, self.url])
|
|
|
|
if 0 == r.rowcount:
|
|
|
|
my_db.action('REPLACE INTO connection_fails_count (domain_url, %s) VALUES (?,?)' % field,
|
|
|
|
[self.url, value])
|
|
|
|
|
|
|
|
def save_list(self):
|
|
|
|
if self.dirty:
|
|
|
|
self.clear_old()
|
|
|
|
if None is not db:
|
|
|
|
with self.lock:
|
|
|
|
try:
|
|
|
|
my_db = db.DBConnection('cache.db')
|
|
|
|
cl = []
|
|
|
|
for f in self._fails:
|
|
|
|
cl.append(['INSERT OR IGNORE INTO connection_fails (domain_url, fail_type, fail_code, '
|
|
|
|
'fail_time) '
|
|
|
|
'VALUES (?,?,?,?)', [self.url, f.fail_type, f.code,
|
|
|
|
_totimestamp(f.fail_time)]])
|
|
|
|
self.dirty = False
|
|
|
|
if cl:
|
|
|
|
my_db.mass_action(cl)
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
self.last_save = datetime.datetime.now()
|
|
|
|
|
|
|
|
def load_list(self):
|
|
|
|
if None is not db:
|
|
|
|
with self.lock:
|
|
|
|
try:
|
|
|
|
my_db = db.DBConnection('cache.db')
|
2023-02-24 11:46:07 +00:00
|
|
|
if my_db.has_table('connection_fails'):
|
2023-01-12 01:04:47 +00:00
|
|
|
results = my_db.select('SELECT * FROM connection_fails WHERE domain_url = ?', [self.url])
|
|
|
|
self._fails = []
|
|
|
|
for r in results:
|
|
|
|
try:
|
|
|
|
self._fails.append(ConnectionFail(
|
|
|
|
fail_type=try_int(r['fail_type']), code=try_int(r['fail_code']),
|
|
|
|
fail_time=datetime.datetime.fromtimestamp(try_int(r['fail_time']))))
|
|
|
|
except (BaseException, Exception):
|
|
|
|
continue
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def clear_old(self):
|
|
|
|
if None is not db:
|
|
|
|
with self.lock:
|
|
|
|
try:
|
|
|
|
my_db = db.DBConnection('cache.db')
|
2023-02-24 11:46:07 +00:00
|
|
|
if my_db.has_table('connection_fails'):
|
2023-01-12 01:04:47 +00:00
|
|
|
# noinspection PyCallByClass,PyTypeChecker
|
|
|
|
time_limit = _totimestamp(datetime.datetime.now() - datetime.timedelta(days=28))
|
|
|
|
my_db.action('DELETE FROM connection_fails WHERE fail_time < ?', [time_limit])
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def _totimestamp(dt=None):
|
|
|
|
# type: (Optional[datetime.datetime]) -> integer_types
|
|
|
|
""" This function should only be used in this module due to its 1970s+ limitation as that's all we need here and
|
|
|
|
sgdatatime can't be used at this module level
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
if PY2:
|
|
|
|
import time
|
|
|
|
return int(time.mktime(dt.timetuple()))
|
|
|
|
return int(datetime.datetime.timestamp(dt))
|
|
|
|
except (BaseException, Exception):
|
|
|
|
return 0
|
|
|
|
|
|
|
|
|
|
|
|
def _log_failure_url(url, post_data=None, post_json=None):
|
|
|
|
# type: (AnyStr, Optional[AnyStr], Optional[AnyStr]) -> None
|
|
|
|
if DOMAIN_FAILURES.should_skip(url, log_warning=False):
|
|
|
|
post = []
|
|
|
|
if post_data:
|
|
|
|
post += [' .. Post params: [%s]' % '&'.join([post_data])]
|
|
|
|
if post_json:
|
|
|
|
post += [' .. Json params: [%s]' % '&'.join([post_json])]
|
|
|
|
logger.warning('Failure URL: %s%s' % (url, ''.join(post)))
|
|
|
|
|
|
|
|
|
|
|
|
# try to convert to int, if the value is not already int
|
|
|
|
def try_ord(c):
|
|
|
|
# type: (Union[int, chr]) -> int
|
|
|
|
if isinstance(c, int):
|
|
|
|
return c
|
|
|
|
return ord(c)
|
|
|
|
|
|
|
|
|
|
|
|
# try to convert to int, if it fails the default will be returned
|
|
|
|
def try_int(s, s_default=0):
|
|
|
|
try:
|
|
|
|
return int(s)
|
|
|
|
except (BaseException, Exception):
|
|
|
|
return s_default
|
|
|
|
|
|
|
|
|
Add try_date for use in attempting conversion of unspecified date format types birthdate, deathdate, aired.
Change tweaks, remove one warn, order imports.
Change tidy up of other warnings, mismatched types, typos, a strange arg issue with deepcopy, and PEP8.
Bug fix, bad key in get_item for TVInfoSocialIDs.
Fix ambiguities of `show` used in sg versus external uses.
Change add cast, crew type mappings
Change only take Main Actors, Hosts, Interviewers, Presenters.
2021-09-05 19:10:58 +00:00
|
|
|
def try_date(s, s_default=None):
|
|
|
|
# type: (AnyStr, Any) -> Optional[AnyStr]
|
|
|
|
"""
|
|
|
|
Convert string to a standard UTC date string
|
|
|
|
:param s:
|
|
|
|
:param s_default:
|
|
|
|
:return:
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
parse = tz_p.parse(clean_data(s))
|
|
|
|
return '%04d-%02d-%02d' % (parse.year, parse.month, parse.day)
|
|
|
|
except(BaseException, Exception):
|
|
|
|
return s_default
|
|
|
|
|
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
def _maybe_request_url(e, def_url=''):
|
|
|
|
return hasattr(e, 'request') and hasattr(e.request, 'url') and ' ' + e.request.url or def_url
|
|
|
|
|
|
|
|
|
|
|
|
def clean_data(data):
|
|
|
|
# type: (Any) -> Any
|
|
|
|
"""Cleans up strings, lists, dicts returned
|
|
|
|
|
|
|
|
Issues corrected:
|
|
|
|
- Replaces & with &
|
Add try_date for use in attempting conversion of unspecified date format types birthdate, deathdate, aired.
Change tweaks, remove one warn, order imports.
Change tidy up of other warnings, mismatched types, typos, a strange arg issue with deepcopy, and PEP8.
Bug fix, bad key in get_item for TVInfoSocialIDs.
Fix ambiguities of `show` used in sg versus external uses.
Change add cast, crew type mappings
Change only take Main Actors, Hosts, Interviewers, Presenters.
2021-09-05 19:10:58 +00:00
|
|
|
- Replace multiple spaces with one space
|
2023-01-12 01:04:47 +00:00
|
|
|
- Trailing whitespace
|
|
|
|
- Decode html entities
|
|
|
|
:param data: data
|
|
|
|
:type data: List or Dict or AnyStr
|
|
|
|
:return:
|
|
|
|
:rtype: List or Dict or AnyStr
|
|
|
|
"""
|
|
|
|
|
|
|
|
if isinstance(data, list):
|
|
|
|
return [clean_data(d) for d in data]
|
|
|
|
if isinstance(data, set):
|
|
|
|
return set(clean_data(d) for d in data)
|
|
|
|
if isinstance(data, dict):
|
|
|
|
return {k: clean_data(v) for k, v in iteritems(data)}
|
|
|
|
if isinstance(data, string_types):
|
Add try_date for use in attempting conversion of unspecified date format types birthdate, deathdate, aired.
Change tweaks, remove one warn, order imports.
Change tidy up of other warnings, mismatched types, typos, a strange arg issue with deepcopy, and PEP8.
Bug fix, bad key in get_item for TVInfoSocialIDs.
Fix ambiguities of `show` used in sg versus external uses.
Change add cast, crew type mappings
Change only take Main Actors, Hosts, Interviewers, Presenters.
2021-09-05 19:10:58 +00:00
|
|
|
return unicodedata.normalize('NFKD', re.sub(r' {2,}', ' ', html_unescape(data).strip().replace('&', '&')))
|
2023-01-12 01:04:47 +00:00
|
|
|
return data
|
|
|
|
|
|
|
|
|
|
|
|
def get_system_temp_dir():
|
|
|
|
"""
|
|
|
|
:return: Returns the [system temp dir]/tvdb_api-u501 (or tvdb_api-myuser)
|
|
|
|
:rtype: AnyStr
|
|
|
|
"""
|
|
|
|
if hasattr(os, 'getuid'):
|
|
|
|
uid = 'u%d' % (os.getuid())
|
|
|
|
else:
|
|
|
|
# For Windows
|
|
|
|
try:
|
|
|
|
uid = getpass.getuser()
|
|
|
|
except ImportError:
|
2023-02-10 14:15:50 +00:00
|
|
|
return os.path.join(tempfile.gettempdir(), 'SickGear')
|
2023-01-12 01:04:47 +00:00
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
return os.path.join(tempfile.gettempdir(), 'SickGear-%s' % uid)
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
|
|
|
|
def proxy_setting(setting, request_url, force=False):
|
|
|
|
"""
|
2023-02-13 21:00:11 +00:00
|
|
|
Returns a list of
|
|
|
|
a) proxy_setting address value or a PAC is fetched and parsed if proxy_setting starts with "PAC:" (case-insensitive)
|
|
|
|
b) True/False if "PAC" is found in the proxy_setting.
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
The PAC data parser is crude, javascript is not eval'd. The first "PROXY URL" found is extracted with a list
|
|
|
|
of "url_a_part.url_remaining", "url_b_part.url_remaining", "url_n_part.url_remaining" and so on.
|
|
|
|
Also, PAC data items are escaped for matching therefore regular expression items will not match a request_url.
|
|
|
|
|
|
|
|
If force is True or request_url contains a PAC parsed data item then the PAC proxy address is returned else False.
|
|
|
|
None is returned in the event of an error fetching PAC data.
|
|
|
|
|
|
|
|
"""
|
|
|
|
|
|
|
|
# check for "PAC" usage
|
|
|
|
match = re.search(r'^\s*PAC:\s*(.*)', setting, re.I)
|
|
|
|
if not match:
|
|
|
|
return setting, False
|
|
|
|
pac_url = match.group(1)
|
|
|
|
|
|
|
|
# prevent a recursive test with existing proxy setting when fetching PAC url
|
|
|
|
global PROXY_SETTING
|
|
|
|
proxy_setting_backup = PROXY_SETTING
|
|
|
|
PROXY_SETTING = ''
|
|
|
|
|
|
|
|
resp = ''
|
|
|
|
try:
|
|
|
|
resp = get_url(pac_url)
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
PROXY_SETTING = proxy_setting_backup
|
|
|
|
|
|
|
|
if not resp:
|
|
|
|
return None, False
|
|
|
|
|
|
|
|
proxy_address = None
|
|
|
|
request_url_match = False
|
|
|
|
parsed_url = urlparse(request_url)
|
|
|
|
netloc = parsed_url.netloc
|
2023-02-13 21:00:11 +00:00
|
|
|
for pac_data in re.finditer(r"""[^'"]*['"]([^.]+\.[^'"]*)['"]""", resp, re.I):
|
2023-01-12 01:04:47 +00:00
|
|
|
data = re.search(r"""PROXY\s+([^'"]+)""", pac_data.group(1), re.I)
|
|
|
|
if data:
|
|
|
|
if force:
|
|
|
|
return data.group(1), True
|
|
|
|
# noinspection PyUnresolvedReferences
|
|
|
|
proxy_address = (proxy_address, data.group(1))[None is proxy_address]
|
|
|
|
elif re.search(re.escape(pac_data.group(1)), netloc, re.I):
|
|
|
|
request_url_match = True
|
|
|
|
if None is not proxy_address:
|
|
|
|
break
|
|
|
|
|
|
|
|
if None is proxy_address:
|
|
|
|
return None, True
|
|
|
|
|
|
|
|
return (False, proxy_address)[request_url_match], True
|
|
|
|
|
|
|
|
|
|
|
|
def get_url(url, # type: AnyStr
|
|
|
|
post_data=None, # type: Optional
|
|
|
|
params=None, # type: Optional
|
|
|
|
headers=None, # type: Optional[Dict]
|
|
|
|
timeout=30, # type: int
|
|
|
|
session=None, # type: Optional[requests.Session]
|
|
|
|
parse_json=False, # type: bool
|
|
|
|
memcache_cookies=None, # type: dict
|
|
|
|
raise_status_code=False, # type: bool
|
|
|
|
raise_exceptions=False, # type: bool
|
|
|
|
as_binary=False, # type: bool
|
|
|
|
encoding=None, # type: Optional[AnyStr]
|
|
|
|
failure_monitor=True, # type: bool
|
|
|
|
use_tmr_limit=True, # type: bool
|
|
|
|
raise_skip_exception=False, # type: bool
|
|
|
|
exclude_client_http_codes=True, # type: bool
|
|
|
|
exclude_http_codes=(404, 429), # type: Tuple[integer_types]
|
|
|
|
exclude_no_data=True, # type: bool
|
|
|
|
use_method=None, # type: Optional[AnyStr]
|
|
|
|
return_response=False, # type: bool
|
|
|
|
**kwargs):
|
|
|
|
# type: (...) -> Optional[Union[AnyStr, bool, bytes, Dict, Tuple[Union[Dict, List], requests.Session], requests.Response]]
|
|
|
|
"""
|
|
|
|
Return data from a URI with a possible check for authentication prior to the data fetch.
|
|
|
|
Raised errors and no data in responses are tracked for making future logic decisions.
|
|
|
|
|
|
|
|
# param url_solver=sickgear.FLARESOLVERR_HOST must be passed if url is behind CF for use in cf_scrape/__init__.py
|
|
|
|
|
|
|
|
Returned data is either:
|
|
|
|
1) a byte-string retrieved from the URL provider.
|
|
|
|
2) a boolean if successfully used kwargs 'savefile' set to file pathname.
|
|
|
|
3) JSON dict if parse_json is True, and `Requests::session` when kwargs 'resp_sess' True.
|
|
|
|
4) `Requests::response`, and `Requests::session` when kwargs 'resp_sess' is True.
|
|
|
|
|
|
|
|
:param url: address to request fetch data from
|
|
|
|
:param post_data: if this or `post_json` is set, then request POST method is used to send this data
|
|
|
|
:param params:
|
|
|
|
:param headers: headers to add
|
|
|
|
:param timeout: timeout
|
|
|
|
:param session: optional session object
|
|
|
|
:param parse_json: return JSON Dict
|
|
|
|
:param memcache_cookies: memory persistent store for cookies
|
|
|
|
:param raise_status_code: raise exception for status codes
|
|
|
|
:param raise_exceptions: raise exceptions
|
|
|
|
:param as_binary: return bytes instead of text
|
|
|
|
:param encoding: overwrite encoding return header if as_binary is False
|
|
|
|
:param failure_monitor: if True, will enable failure monitor for this request
|
|
|
|
:param use_tmr_limit: an API limit can be +ve before a fetch, but unwanted, set False to short should_skip
|
|
|
|
:param raise_skip_exception: if True, will raise ConnectionSkipException if this request should be skipped
|
|
|
|
:param exclude_client_http_codes: if True, exclude client http codes 4XX from failure monitor
|
|
|
|
:param exclude_http_codes: http codes to exclude from failure monitor, default: (404, 429)
|
|
|
|
:param exclude_no_data: exclude no data as failure
|
|
|
|
:param use_method: force any supported method by Session(): get, put, post, delete
|
|
|
|
:param return_response: return response object
|
|
|
|
:param kwargs: keyword params to passthru to Requests
|
|
|
|
:return: None or data fetched from address
|
|
|
|
"""
|
|
|
|
|
|
|
|
domain = None
|
|
|
|
if failure_monitor:
|
|
|
|
domain = DOMAIN_FAILURES.get_domain(url)
|
|
|
|
if domain not in DOMAIN_FAILURES.domain_list:
|
|
|
|
DOMAIN_FAILURES.domain_list[domain] = ConnectionFailList(domain)
|
|
|
|
|
|
|
|
if DOMAIN_FAILURES.should_skip(url, use_tmr_limit=use_tmr_limit):
|
|
|
|
if raise_skip_exception:
|
|
|
|
raise ConnectionSkipException
|
|
|
|
return
|
|
|
|
|
|
|
|
response_attr = ('text', 'content')[as_binary]
|
|
|
|
|
|
|
|
# selectively mute some errors
|
2023-02-11 18:02:58 +00:00
|
|
|
mute = list(filter(lambda x: kwargs.pop(x, False), [
|
|
|
|
'mute_connect_err', 'mute_read_timeout', 'mute_connect_timeout', 'mute_http_error']))
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
# reuse or instantiate request session
|
|
|
|
resp_sess = kwargs.pop('resp_sess', None)
|
|
|
|
if None is session:
|
|
|
|
session = CloudflareScraper.create_scraper()
|
|
|
|
session.headers.update({'User-Agent': USER_AGENT})
|
|
|
|
|
|
|
|
proxy_browser = kwargs.get('proxy_browser')
|
|
|
|
if isinstance(memcache_cookies, dict):
|
|
|
|
parsed_url = urlparse(url)
|
|
|
|
domain = parsed_url.netloc
|
|
|
|
if domain in memcache_cookies:
|
|
|
|
session.cookies.update(memcache_cookies[domain])
|
|
|
|
|
|
|
|
# download and save file or simply fetch url
|
|
|
|
savename = kwargs.pop('savename', None)
|
|
|
|
if savename:
|
|
|
|
# session streaming
|
|
|
|
session.stream = True
|
|
|
|
|
|
|
|
if not kwargs.pop('nocache', False):
|
|
|
|
cache_dir = CACHE_DIR or get_system_temp_dir()
|
2023-02-10 14:15:50 +00:00
|
|
|
session = CacheControl(sess=session, cache=caches.FileCache(os.path.join(cache_dir, 'sessions')))
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
provider = kwargs.pop('provider', None)
|
|
|
|
|
|
|
|
# handle legacy uses of `json` param
|
|
|
|
if kwargs.get('json'):
|
|
|
|
parse_json = kwargs.pop('json')
|
|
|
|
post_json = kwargs.pop('post_json', None)
|
|
|
|
|
|
|
|
# session main headers
|
|
|
|
req_headers = {'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',
|
2023-09-08 12:07:06 +00:00
|
|
|
'Accept-Encoding': ACCEPT_ENCODING}
|
2023-01-12 01:04:47 +00:00
|
|
|
if headers:
|
|
|
|
req_headers.update(headers)
|
|
|
|
if hasattr(session, 'reserved') and 'headers' in session.reserved:
|
|
|
|
req_headers.update(session.reserved['headers'] or {})
|
|
|
|
session.headers.update(req_headers)
|
|
|
|
|
|
|
|
# session parameters
|
|
|
|
session.params = params
|
|
|
|
|
|
|
|
# session ssl verify
|
|
|
|
session.verify = False
|
|
|
|
|
|
|
|
# don't trust os environments (auth, proxies, ...)
|
|
|
|
session.trust_env = False
|
|
|
|
|
|
|
|
result = response = raised = connection_fail_params = log_failure_url = None
|
|
|
|
try:
|
|
|
|
# sanitise url
|
|
|
|
parsed = list(urlparse(url))
|
|
|
|
parsed[2] = re.sub('/{2,}', '/', parsed[2]) # replace two or more / with one
|
|
|
|
url = urlunparse(parsed)
|
|
|
|
|
|
|
|
# session proxies
|
|
|
|
if PROXY_SETTING:
|
|
|
|
(proxy_address, pac_found) = proxy_setting(PROXY_SETTING, url)
|
|
|
|
msg = '%sproxy for url: %s' % (('', 'PAC parsed ')[pac_found], url)
|
|
|
|
if None is proxy_address:
|
|
|
|
logger.debug('Proxy error, aborted the request using %s' % msg)
|
|
|
|
return
|
|
|
|
elif proxy_address:
|
|
|
|
logger.debug('Using %s' % msg)
|
|
|
|
session.proxies = {'http': proxy_address, 'https': proxy_address}
|
|
|
|
|
|
|
|
if None is not use_method:
|
|
|
|
|
|
|
|
method = getattr(session, use_method.strip().lower())
|
|
|
|
|
|
|
|
elif post_data or post_json: # decide if to post data or send a get request to server
|
|
|
|
|
|
|
|
if True is post_data:
|
|
|
|
post_data = None
|
|
|
|
|
|
|
|
if post_data:
|
|
|
|
kwargs.setdefault('data', post_data)
|
|
|
|
|
|
|
|
if post_json:
|
|
|
|
kwargs.setdefault('json', post_json)
|
|
|
|
|
|
|
|
method = session.post
|
|
|
|
else:
|
|
|
|
method = session.get
|
|
|
|
|
|
|
|
for r in range(0, 5):
|
|
|
|
response = method(url, timeout=timeout, **kwargs)
|
|
|
|
if not savename and response.ok and not response.content:
|
|
|
|
if 'url=' in response.headers.get('Refresh', '').lower():
|
|
|
|
url = response.headers.get('Refresh').lower().split('url=')[1].strip('/')
|
|
|
|
if not url.startswith('http'):
|
|
|
|
parsed[2] = '/%s' % url
|
|
|
|
url = urlunparse(parsed)
|
|
|
|
response = session.get(url, timeout=timeout, **kwargs)
|
|
|
|
elif 'github' in url:
|
|
|
|
time.sleep(2)
|
|
|
|
continue
|
|
|
|
break
|
|
|
|
|
|
|
|
# if encoding is not in header try to use best guess
|
|
|
|
# ignore downloads with savename
|
|
|
|
if not savename and not as_binary:
|
|
|
|
if encoding:
|
|
|
|
response.encoding = encoding
|
|
|
|
elif not response.encoding or 'charset' not in response.headers.get('Content-Type', ''):
|
|
|
|
response.encoding = response.apparent_encoding
|
|
|
|
|
|
|
|
# noinspection PyProtectedMember
|
|
|
|
if provider and provider._has_signature(response.text):
|
|
|
|
result = getattr(response, response_attr)
|
|
|
|
else:
|
|
|
|
if raise_status_code:
|
|
|
|
response.raise_for_status()
|
|
|
|
|
|
|
|
if not response.ok:
|
|
|
|
http_err_text = 'CloudFlare Ray ID' in response.text and \
|
|
|
|
'CloudFlare reports, "Website is offline"; ' or ''
|
|
|
|
if response.status_code in http_error_code:
|
|
|
|
http_err_text += http_error_code[response.status_code]
|
|
|
|
elif response.status_code in range(520, 527):
|
|
|
|
http_err_text += 'Origin server connection failure'
|
|
|
|
else:
|
|
|
|
http_err_text = 'Custom HTTP error code'
|
|
|
|
if 'mute_http_error' not in mute:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f'Response not ok. {response.status_code}: {http_err_text} from requested url'
|
|
|
|
f' {url}')
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
except requests.exceptions.HTTPError as e:
|
|
|
|
raised = e
|
|
|
|
is_client_error = 400 <= e.response.status_code < 500
|
|
|
|
if failure_monitor and e.response.status_code not in exclude_http_codes and \
|
|
|
|
not (exclude_client_http_codes and is_client_error):
|
|
|
|
connection_fail_params = dict(fail_type=ConnectionFailTypes.http, code=e.response.status_code)
|
|
|
|
if not raise_status_code:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.warning(f'HTTP error {e.errno} while loading URL{_maybe_request_url(e)}')
|
2023-01-12 01:04:47 +00:00
|
|
|
except requests.exceptions.ConnectionError as e:
|
|
|
|
raised = e
|
|
|
|
if 'mute_connect_err' not in mute:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.warning(f"Connection error msg:{ex(e)} while loading URL{_maybe_request_url(e)}")
|
2023-01-12 01:04:47 +00:00
|
|
|
if failure_monitor:
|
|
|
|
connection_fail_params = dict(fail_type=ConnectionFailTypes.connection)
|
|
|
|
except requests.exceptions.ReadTimeout as e:
|
|
|
|
raised = e
|
|
|
|
if 'mute_read_timeout' not in mute:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.warning(f'Read timed out msg:{ex(e)} while loading URL{_maybe_request_url(e)}')
|
2023-01-12 01:04:47 +00:00
|
|
|
if failure_monitor:
|
|
|
|
connection_fail_params = dict(fail_type=ConnectionFailTypes.timeout)
|
|
|
|
except (requests.exceptions.Timeout, socket.timeout) as e:
|
|
|
|
raised = e
|
|
|
|
if 'mute_connect_timeout' not in mute:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.warning(f'Connection timed out msg:{ex(e)} while loading URL {_maybe_request_url(e, url)}')
|
2023-01-12 01:04:47 +00:00
|
|
|
if failure_monitor:
|
|
|
|
connection_fail_params = dict(fail_type=ConnectionFailTypes.connection_timeout)
|
|
|
|
except (BaseException, Exception) as e:
|
|
|
|
raised = e
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.warning(('Exception caught while loading URL {0}\r\nDetail... %s\r\n{1}' % ex(e),
|
|
|
|
'Unknown exception while loading URL {0}\r\nDetail... {1}')[not ex(e)]
|
2023-01-12 01:04:47 +00:00
|
|
|
.format(url, traceback.format_exc()))
|
|
|
|
if failure_monitor:
|
|
|
|
connection_fail_params = dict(fail_type=ConnectionFailTypes.other)
|
|
|
|
log_failure_url = True
|
|
|
|
finally:
|
|
|
|
if None is not connection_fail_params:
|
|
|
|
DOMAIN_FAILURES.inc_failure_count(url, ConnectionFail(**connection_fail_params))
|
|
|
|
save_failure(url, domain, log_failure_url, post_data, post_json)
|
|
|
|
|
|
|
|
if isinstance(raised, Exception):
|
|
|
|
if raise_exceptions or raise_status_code:
|
|
|
|
try:
|
|
|
|
if not hasattr(raised, 'text') and hasattr(response, 'text'):
|
|
|
|
raised.text = response.text
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
raise raised
|
|
|
|
return
|
|
|
|
|
|
|
|
if return_response:
|
|
|
|
result = response
|
|
|
|
elif None is result and None is not response and response.ok:
|
|
|
|
if isinstance(memcache_cookies, dict):
|
|
|
|
parsed_url = urlparse(url)
|
|
|
|
domain = parsed_url.netloc
|
|
|
|
memcache_cookies[domain] = session.cookies.copy()
|
|
|
|
|
|
|
|
if parse_json or proxy_browser:
|
|
|
|
try:
|
|
|
|
data_json = json_loads(requests_response=response)
|
|
|
|
if proxy_browser:
|
|
|
|
result = ({}, data_json.get('solution', {}).get('response', {}))[isinstance(data_json, dict)]
|
|
|
|
else:
|
|
|
|
result = ({}, data_json)[isinstance(data_json, (dict, list))]
|
|
|
|
if resp_sess:
|
|
|
|
result = result, session
|
|
|
|
except (TypeError, Exception) as e:
|
|
|
|
raised = e
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.warning(f'{("Proxy browser", "JSON")[parse_json]} data issue from URL {url}\r\n'
|
|
|
|
f'Detail... {ex(e)}')
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
elif savename:
|
|
|
|
try:
|
|
|
|
write_file(savename, response, raw=True, raise_exceptions=raise_exceptions)
|
|
|
|
result = True
|
|
|
|
except (BaseException, Exception) as e:
|
|
|
|
raised = e
|
|
|
|
|
|
|
|
else:
|
|
|
|
result = getattr(response, response_attr)
|
|
|
|
if resp_sess:
|
|
|
|
result = result, session
|
|
|
|
|
|
|
|
if raise_exceptions and isinstance(raised, Exception):
|
|
|
|
raise raised
|
|
|
|
|
|
|
|
if failure_monitor:
|
|
|
|
if return_response or (result and not isinstance(result, tuple)
|
|
|
|
or isinstance(result, tuple) and result[0]):
|
|
|
|
domain = DOMAIN_FAILURES.get_domain(url)
|
|
|
|
if 0 != DOMAIN_FAILURES.domain_list[domain].failure_count:
|
|
|
|
logger.info('Unblocking: %s' % domain)
|
|
|
|
DOMAIN_FAILURES.domain_list[domain].failure_count = 0
|
|
|
|
DOMAIN_FAILURES.domain_list[domain].failure_time = None
|
|
|
|
save_failure(url, domain, False, post_data, post_json)
|
|
|
|
elif not exclude_no_data:
|
|
|
|
DOMAIN_FAILURES.inc_failure_count(url, ConnectionFail(fail_type=ConnectionFailTypes.nodata))
|
|
|
|
save_failure(url, domain, True, post_data, post_json)
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
|
|
|
def save_failure(url, domain, log_failure_url, post_data, post_json):
|
|
|
|
DOMAIN_FAILURES.domain_list[domain].save_list()
|
|
|
|
if log_failure_url:
|
|
|
|
_log_failure_url(url, post_data, post_json)
|
|
|
|
|
|
|
|
|
|
|
|
def scantree(path, # type: AnyStr
|
|
|
|
exclude=None, # type: Optional[AnyStr, List[AnyStr]]
|
2023-10-28 00:31:34 +00:00
|
|
|
exclude_dirs=None, # type: Optional[AnyStr, List[AnyStr]]
|
2023-01-12 01:04:47 +00:00
|
|
|
include=None, # type: Optional[AnyStr, List[AnyStr]]
|
|
|
|
follow_symlinks=False, # type: bool
|
|
|
|
filter_kind=None, # type: Optional[bool]
|
2023-06-08 11:01:38 +00:00
|
|
|
recurse=True, # type: bool
|
2023-10-28 07:49:16 +00:00
|
|
|
exclude_folders_with_files=None, # type: Optional[List[AnyStr]]
|
|
|
|
internal_call=False, # type: bool
|
|
|
|
rc_exc=None, # type: List
|
|
|
|
rc_exc_dir=None, # type: List
|
2023-10-28 12:31:38 +00:00
|
|
|
rc_inc=None, # type: List
|
|
|
|
has_exclude=False, # type: bool
|
|
|
|
has_exclude_dirs=False, # type: bool
|
|
|
|
has_include=False # type: bool
|
2023-01-12 01:04:47 +00:00
|
|
|
):
|
|
|
|
# type: (...) -> Generator[DirEntry, None, None]
|
|
|
|
"""Yield DirEntry objects for given path. Returns without yield if path fails sanity check
|
|
|
|
|
|
|
|
:param path: Path to scan, sanity check is_dir and exists
|
2023-10-28 00:31:34 +00:00
|
|
|
:param exclude: Escaped regex string(s) to exclude (files and directories)
|
|
|
|
:param exclude_dirs: Escaped regex string(s) to exclude (directories only)
|
2023-01-12 01:04:47 +00:00
|
|
|
:param include: Escaped regex string(s) to include
|
|
|
|
:param follow_symlinks: Follow symlinks
|
|
|
|
:param filter_kind: None to yield everything, True yields directories, False yields files
|
|
|
|
:param recurse: Recursively scan the tree
|
2023-06-08 11:01:38 +00:00
|
|
|
:param exclude_folders_with_files: exclude folder that contain the listed file(s)
|
2023-10-28 07:49:16 +00:00
|
|
|
:param internal_call: internal use
|
|
|
|
:param rc_exc: internal use
|
|
|
|
:param rc_exc_dir: internal use
|
|
|
|
:param rc_inc: internal use
|
2023-10-28 12:31:38 +00:00
|
|
|
:param has_exclude: internal use
|
|
|
|
:param has_exclude_dirs: internal use
|
|
|
|
:param has_include: internal_use
|
2023-01-12 01:04:47 +00:00
|
|
|
"""
|
2023-02-10 14:15:50 +00:00
|
|
|
if isinstance(path, string_types) and path and os.path.isdir(path):
|
2023-10-28 07:49:16 +00:00
|
|
|
if not internal_call:
|
|
|
|
rc_exc, rc_exc_dir, rc_inc = [re.compile(rx % '|'.join(
|
|
|
|
[x for x in (param, ([param], [])[None is param])[not isinstance(param, list)]]))
|
|
|
|
for rx, param in ((r'(?i)^(?:(?!%s).)*$', exclude), (r'(?i)^(?:(?!%s).)*$', exclude_dirs),
|
|
|
|
(r'(?i)%s', include))]
|
2023-10-28 12:31:38 +00:00
|
|
|
has_exclude, has_exclude_dirs, has_include = bool(exclude), bool(exclude_dirs), bool(include)
|
|
|
|
|
2023-10-31 16:05:51 +00:00
|
|
|
with scandir(path) as s_d:
|
|
|
|
for entry in s_d:
|
|
|
|
is_dir = entry.is_dir(follow_symlinks=follow_symlinks)
|
|
|
|
is_file = entry.is_file(follow_symlinks=follow_symlinks)
|
|
|
|
no_filter = any([None is filter_kind, filter_kind and is_dir, not filter_kind and is_file])
|
|
|
|
if ((not has_exclude or rc_exc.search(entry.name))
|
|
|
|
and (not has_exclude_dirs or not is_dir or rc_exc_dir.search(entry.name))
|
|
|
|
and (not has_include or rc_inc.search(entry.name))
|
|
|
|
and (no_filter or (not filter_kind and is_dir and recurse))):
|
|
|
|
if is_dir and exclude_folders_with_files and any(os.path.isfile(os.path.join(entry.path, e_f))
|
|
|
|
for e_f in exclude_folders_with_files):
|
|
|
|
logger.debug(f'Ignoring Folder: "{entry.path}", because it contains a exclude file'
|
|
|
|
f' "{", ".join(exclude_folders_with_files)}"')
|
|
|
|
continue
|
|
|
|
if recurse and is_dir:
|
|
|
|
for subentry in scantree(
|
|
|
|
path=entry.path, exclude=exclude, exclude_dirs=exclude_dirs, include=include,
|
|
|
|
follow_symlinks=follow_symlinks, filter_kind=filter_kind, recurse=recurse,
|
|
|
|
exclude_folders_with_files=exclude_folders_with_files, internal_call=True,
|
|
|
|
rc_exc=rc_exc, rc_exc_dir=rc_exc_dir, rc_inc=rc_inc, has_exclude=has_exclude,
|
|
|
|
has_exclude_dirs=has_exclude_dirs, has_include=has_include):
|
|
|
|
yield subentry
|
|
|
|
if no_filter:
|
|
|
|
yield entry
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
|
|
|
|
def copy_file(src_file, dest_file):
|
|
|
|
if os.name.startswith('posix'):
|
2023-02-10 14:15:50 +00:00
|
|
|
subprocess.call(['cp', src_file, dest_file])
|
2023-01-12 01:04:47 +00:00
|
|
|
else:
|
2023-02-10 14:15:50 +00:00
|
|
|
shutil.copyfile(src_file, dest_file)
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
try:
|
2023-02-10 14:15:50 +00:00
|
|
|
shutil.copymode(src_file, dest_file)
|
2023-01-12 01:04:47 +00:00
|
|
|
except OSError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def move_file(src_file, dest_file, raise_exceptions=False):
|
|
|
|
try:
|
2023-02-10 14:15:50 +00:00
|
|
|
shutil.move(src_file, dest_file)
|
2023-01-12 01:04:47 +00:00
|
|
|
fix_set_group_id(dest_file)
|
|
|
|
except OSError:
|
|
|
|
copy_file(src_file, dest_file)
|
2023-02-10 14:15:50 +00:00
|
|
|
if os.path.exists(dest_file):
|
2023-01-12 01:04:47 +00:00
|
|
|
fix_set_group_id(dest_file)
|
2023-02-10 14:15:50 +00:00
|
|
|
os.unlink(src_file)
|
2023-01-12 01:04:47 +00:00
|
|
|
elif raise_exceptions:
|
|
|
|
raise OSError('Destination file could not be created: %s' % dest_file)
|
|
|
|
|
|
|
|
|
|
|
|
def fix_set_group_id(child_path):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param child_path: path
|
|
|
|
:type child_path: AnyStr
|
|
|
|
:return:
|
|
|
|
:rtype: None
|
|
|
|
"""
|
|
|
|
if os.name in ('nt', 'ce'):
|
|
|
|
return
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
parent_path = os.path.dirname(child_path)
|
|
|
|
parent_stat = os.stat(parent_path)
|
2023-01-12 01:04:47 +00:00
|
|
|
parent_mode = stat.S_IMODE(parent_stat[stat.ST_MODE])
|
|
|
|
|
|
|
|
if parent_mode & stat.S_ISGID:
|
|
|
|
parent_gid = parent_stat[stat.ST_GID]
|
2023-02-10 14:15:50 +00:00
|
|
|
child_stat = os.stat(child_path)
|
2023-01-12 01:04:47 +00:00
|
|
|
child_gid = child_stat[stat.ST_GID]
|
|
|
|
|
|
|
|
if child_gid == parent_gid:
|
|
|
|
return
|
|
|
|
|
|
|
|
child_path_owner = child_stat.st_uid
|
|
|
|
user_id = os.geteuid() # only available on UNIX
|
|
|
|
|
|
|
|
if 0 != user_id and user_id != child_path_owner:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f'Not running as root or owner of {child_path}, not trying to set the set-group-id')
|
2023-01-12 01:04:47 +00:00
|
|
|
return
|
|
|
|
|
|
|
|
try:
|
2023-02-10 14:15:50 +00:00
|
|
|
os.chown(child_path, -1, parent_gid) # only available on UNIX
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f'Respecting the set-group-ID bit on the parent directory for {child_path}')
|
2023-01-12 01:04:47 +00:00
|
|
|
except OSError:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.error(f'Failed to respect the set-group-id bit on the parent directory for {child_path}'
|
|
|
|
f' (setting group id {parent_gid:d})')
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
|
|
|
|
def remove_file_perm(filepath, log_err=True):
|
|
|
|
# type: (AnyStr, Optional[bool]) -> Optional[bool]
|
|
|
|
"""
|
|
|
|
Remove file
|
|
|
|
|
|
|
|
:param filepath: Path and file name
|
|
|
|
:param log_err: False to suppress log msgs
|
|
|
|
:return True if filepath does not exist else None if no removal
|
|
|
|
"""
|
2023-02-10 14:15:50 +00:00
|
|
|
if not os.path.exists(filepath):
|
2023-01-12 01:04:47 +00:00
|
|
|
return True
|
|
|
|
for t in list_range(10): # total seconds to wait 0 - 9 = 45s over 10 iterations
|
|
|
|
try:
|
2023-02-10 14:15:50 +00:00
|
|
|
os.remove(filepath)
|
2023-01-12 01:04:47 +00:00
|
|
|
except OSError as e:
|
|
|
|
if getattr(e, 'winerror', 0) not in (5, 32): # 5=access denied (e.g. av), 32=another process has lock
|
|
|
|
if log_err:
|
|
|
|
logger.warning('Unable to delete %s: %r / %s' % (filepath, e, ex(e)))
|
|
|
|
return
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
time.sleep(t)
|
2023-02-10 14:15:50 +00:00
|
|
|
if not os.path.exists(filepath):
|
2023-01-12 01:04:47 +00:00
|
|
|
return True
|
|
|
|
if log_err:
|
|
|
|
logger.warning('Unable to delete %s' % filepath)
|
|
|
|
|
|
|
|
|
|
|
|
def remove_file(filepath, tree=False, prefix_failure='', log_level=logging.INFO):
|
|
|
|
"""
|
|
|
|
Remove file based on setting for trash v permanent delete
|
|
|
|
|
|
|
|
:param filepath: Path and file name
|
|
|
|
:type filepath: String
|
|
|
|
:param tree: Remove file tree
|
|
|
|
:type tree: Bool
|
|
|
|
:param prefix_failure: Text to prepend to error log, e.g. show id
|
|
|
|
:type prefix_failure: String
|
|
|
|
:param log_level: Log level to use for error
|
|
|
|
:type log_level: Int
|
|
|
|
:return: Type of removal ('Deleted' or 'Trashed') if filepath does not exist or None if no removal occurred
|
|
|
|
:rtype: String or None
|
|
|
|
"""
|
|
|
|
result = None
|
|
|
|
if filepath:
|
|
|
|
for t in list_range(10): # total seconds to wait 0 - 9 = 45s over 10 iterations
|
|
|
|
try:
|
|
|
|
result = 'Deleted'
|
|
|
|
if TRASH_REMOVE_SHOW:
|
|
|
|
result = 'Trashed'
|
2023-02-10 14:15:50 +00:00
|
|
|
send2trash(filepath)
|
2023-01-12 01:04:47 +00:00
|
|
|
elif tree:
|
2023-02-10 14:15:50 +00:00
|
|
|
shutil.rmtree(filepath)
|
2023-01-12 01:04:47 +00:00
|
|
|
else:
|
2023-02-10 14:15:50 +00:00
|
|
|
os.remove(filepath)
|
2023-01-12 01:04:47 +00:00
|
|
|
except OSError as e:
|
|
|
|
if getattr(e, 'winerror', 0) not in (5, 32): # 5=access denied (e.g. av), 32=another process has lock
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.log(level=log_level,
|
|
|
|
msg=f'{prefix_failure}Unable to {("delete", "trash")[TRASH_REMOVE_SHOW]}'
|
|
|
|
f' {("file", "dir")[tree]} {filepath}: {ex(e)}')
|
2023-01-12 01:04:47 +00:00
|
|
|
break
|
|
|
|
time.sleep(t)
|
2023-02-10 14:15:50 +00:00
|
|
|
if not os.path.exists(filepath):
|
2023-01-12 01:04:47 +00:00
|
|
|
break
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
return (None, result)[filepath and not os.path.exists(filepath)]
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
|
|
|
|
def touch_file(name, atime=None, dir_name=None):
|
|
|
|
# type: (AnyStr, int, AnyStr) -> bool
|
|
|
|
"""
|
|
|
|
create an empty named file and set access time of file
|
|
|
|
|
|
|
|
:param name: filename to touch with time
|
|
|
|
:param atime: access time as epoch
|
|
|
|
:param dir_name: create empty file and directory if file doesn't exist
|
|
|
|
:return: success
|
|
|
|
"""
|
|
|
|
if None is not dir_name:
|
2023-02-10 14:15:50 +00:00
|
|
|
name = os.path.join(dir_name, name)
|
2023-01-12 01:04:47 +00:00
|
|
|
if make_path(dir_name):
|
2023-02-10 14:15:50 +00:00
|
|
|
if not os.path.exists(name):
|
2023-01-12 01:04:47 +00:00
|
|
|
with io.open(name, 'w') as fh:
|
|
|
|
fh.flush()
|
|
|
|
if None is atime:
|
|
|
|
return True
|
|
|
|
|
|
|
|
if None is not atime:
|
|
|
|
try:
|
|
|
|
with open(name, 'a'):
|
2023-02-10 14:15:50 +00:00
|
|
|
os.utime(name, (atime, atime))
|
2023-01-12 01:04:47 +00:00
|
|
|
return True
|
|
|
|
except (BaseException, Exception):
|
|
|
|
logger.debug('File air date stamping not available on your OS')
|
|
|
|
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
|
|
|
def make_path(name, syno=False):
|
|
|
|
# type: (AnyStr, bool) -> bool
|
|
|
|
"""
|
|
|
|
Create an entire directory path and
|
|
|
|
assign each subdir the parent permission
|
|
|
|
|
|
|
|
:param name: directory to create
|
|
|
|
:param syno: whether to trigger a syno library update for path
|
|
|
|
:return: success or dir exists
|
|
|
|
"""
|
2023-02-10 14:15:50 +00:00
|
|
|
if not os.path.isdir(name):
|
2023-01-12 01:04:47 +00:00
|
|
|
# Windows, create all missing folders
|
|
|
|
if os.name in ('nt', 'ce'):
|
|
|
|
try:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f"Path {name} doesn't exist, creating it")
|
2023-02-10 14:15:50 +00:00
|
|
|
os.makedirs(name)
|
2023-01-12 01:04:47 +00:00
|
|
|
except (OSError, IOError) as e:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.error(f'Failed creating {name} : {ex(e)}')
|
2023-01-12 01:04:47 +00:00
|
|
|
return False
|
|
|
|
|
|
|
|
# not Windows, create all missing folders and set permissions
|
|
|
|
else:
|
|
|
|
sofar = ''
|
|
|
|
folder_list = name.split(os.path.sep)
|
|
|
|
|
|
|
|
# look through each sub folder and make sure they all exist
|
|
|
|
for cur_folder in folder_list:
|
|
|
|
sofar += cur_folder + os.path.sep
|
|
|
|
|
|
|
|
# if it exists then just keep walking down the line
|
2023-02-10 14:15:50 +00:00
|
|
|
if os.path.isdir(sofar):
|
2023-01-12 01:04:47 +00:00
|
|
|
continue
|
|
|
|
|
|
|
|
try:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f"Path {sofar} doesn't exist, creating it")
|
2023-02-10 14:15:50 +00:00
|
|
|
os.mkdir(sofar)
|
2023-01-12 01:04:47 +00:00
|
|
|
# use normpath to remove end separator, otherwise checks permissions against itself
|
2023-02-10 14:15:50 +00:00
|
|
|
chmod_as_parent(os.path.normpath(sofar))
|
2023-01-12 01:04:47 +00:00
|
|
|
if syno:
|
|
|
|
# do the library update for synoindex
|
|
|
|
NOTIFIERS.NotifierFactory().get('SYNOINDEX').addFolder(sofar)
|
|
|
|
except (OSError, IOError) as e:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.error(f'Failed creating {sofar} : {ex(e)}')
|
2023-01-12 01:04:47 +00:00
|
|
|
return False
|
|
|
|
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
|
|
|
def chmod_as_parent(child_path):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param child_path: path
|
|
|
|
:type child_path: AnyStr
|
|
|
|
:return:
|
|
|
|
:rtype: None
|
|
|
|
"""
|
|
|
|
if os.name in ('nt', 'ce'):
|
|
|
|
return
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
parent_path = os.path.dirname(child_path)
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
if not parent_path:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f'No parent path provided in {child_path}, unable to get permissions from it')
|
2023-01-12 01:04:47 +00:00
|
|
|
return
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
parent_path_stat = os.stat(parent_path)
|
2023-01-12 01:04:47 +00:00
|
|
|
parent_mode = stat.S_IMODE(parent_path_stat[stat.ST_MODE])
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
child_path_stat = os.stat(child_path)
|
2023-01-12 01:04:47 +00:00
|
|
|
child_path_mode = stat.S_IMODE(child_path_stat[stat.ST_MODE])
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
if os.path.isfile(child_path):
|
2023-01-12 01:04:47 +00:00
|
|
|
child_mode = file_bit_filter(parent_mode)
|
|
|
|
else:
|
|
|
|
child_mode = parent_mode
|
|
|
|
|
|
|
|
if child_path_mode == child_mode:
|
|
|
|
return
|
|
|
|
|
|
|
|
child_path_owner = child_path_stat.st_uid
|
|
|
|
user_id = os.geteuid() # only available on UNIX
|
|
|
|
|
|
|
|
if 0 != user_id and user_id != child_path_owner:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f'Not running as root or owner of {child_path}, not trying to set permissions')
|
2023-01-12 01:04:47 +00:00
|
|
|
return
|
|
|
|
|
|
|
|
try:
|
2023-02-10 14:15:50 +00:00
|
|
|
os.chmod(child_path, child_mode)
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.debug(f'Setting permissions for {child_path} to {child_mode:o} as parent directory has {parent_mode:o}')
|
2023-01-12 01:04:47 +00:00
|
|
|
except OSError:
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.error(f'Failed to set permission for {child_path} to {child_mode:o}')
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
|
|
|
|
def file_bit_filter(mode):
|
|
|
|
for bit in [stat.S_IXUSR, stat.S_IXGRP, stat.S_IXOTH, stat.S_ISUID, stat.S_ISGID]:
|
|
|
|
if mode & bit:
|
|
|
|
mode -= bit
|
|
|
|
|
|
|
|
return mode
|
|
|
|
|
|
|
|
|
|
|
|
def write_file(filepath, # type: AnyStr
|
|
|
|
data, # type: Union[AnyStr, etree.Element, requests.Response]
|
|
|
|
raw=False, # type: bool
|
|
|
|
xmltree=False, # type: bool
|
|
|
|
xml_header=False, # type: bool
|
|
|
|
utf8=False, # type: bool
|
|
|
|
raise_exceptions=False # type: bool
|
|
|
|
): # type: (...) -> bool
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param filepath: filepath
|
|
|
|
:param data: data to write
|
|
|
|
:param raw: write binary or text
|
|
|
|
:param xmltree: use xmel tree
|
|
|
|
:param xml_header: prepend xmel header
|
|
|
|
:param utf8: use UTF8
|
|
|
|
:param raise_exceptions: raise excepitons
|
|
|
|
:return: succuess
|
|
|
|
"""
|
|
|
|
result = False
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
if make_path(os.path.dirname(filepath)):
|
2023-01-12 01:04:47 +00:00
|
|
|
try:
|
|
|
|
if raw:
|
|
|
|
empty_file = True
|
2023-02-10 14:15:50 +00:00
|
|
|
with io.FileIO(filepath, 'wb') as fh:
|
2023-01-12 01:04:47 +00:00
|
|
|
for chunk in data.iter_content(chunk_size=1024):
|
|
|
|
if chunk:
|
|
|
|
empty_file = False
|
|
|
|
fh.write(chunk)
|
|
|
|
fh.flush()
|
2023-02-10 14:15:50 +00:00
|
|
|
os.fsync(fh.fileno())
|
2023-01-12 01:04:47 +00:00
|
|
|
if empty_file:
|
|
|
|
remove_file_perm(filepath, log_err=False)
|
|
|
|
return result
|
|
|
|
else:
|
|
|
|
w_mode = 'w'
|
|
|
|
if utf8:
|
|
|
|
w_mode = 'a'
|
2023-02-10 14:15:50 +00:00
|
|
|
with io.FileIO(filepath, 'wb') as fh:
|
2023-01-12 01:04:47 +00:00
|
|
|
fh.write(codecs.BOM_UTF8)
|
|
|
|
|
|
|
|
if xmltree:
|
2023-02-10 14:15:50 +00:00
|
|
|
with io.FileIO(filepath, w_mode) as fh:
|
2023-01-12 01:04:47 +00:00
|
|
|
params = {}
|
|
|
|
if utf8:
|
|
|
|
params = dict(encoding='utf-8')
|
|
|
|
if xml_header:
|
|
|
|
params.update(dict(xml_declaration=True))
|
|
|
|
data.write(fh, **params)
|
|
|
|
else:
|
|
|
|
if isinstance(data, text_type):
|
2023-02-10 14:15:50 +00:00
|
|
|
with io.open(filepath, w_mode, encoding='utf-8') as fh:
|
2023-01-12 01:04:47 +00:00
|
|
|
fh.write(data)
|
|
|
|
else:
|
2023-02-10 14:15:50 +00:00
|
|
|
with io.FileIO(filepath, w_mode) as fh:
|
2023-01-12 01:04:47 +00:00
|
|
|
fh.write(data)
|
|
|
|
|
|
|
|
chmod_as_parent(filepath)
|
|
|
|
|
|
|
|
result = True
|
|
|
|
except (EnvironmentError, IOError) as e:
|
|
|
|
logger.error('Unable to write file %s : %s' % (filepath, ex(e)))
|
|
|
|
if raise_exceptions:
|
|
|
|
raise e
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
|
|
|
def sanitize_filename(name):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param name: filename
|
|
|
|
:type name: AnyStr
|
|
|
|
:return: sanitized filename
|
|
|
|
:rtype: AnyStr
|
|
|
|
"""
|
|
|
|
# remove bad chars from the filename
|
|
|
|
name = re.sub(r'[\\/*]', '-', name)
|
|
|
|
name = re.sub(r'[:"<>|?]', '', name)
|
|
|
|
|
|
|
|
# remove leading/trailing periods and spaces
|
|
|
|
name = name.strip(' .')
|
|
|
|
|
|
|
|
for char in REMOVE_FILENAME_CHARS or []:
|
|
|
|
name = name.replace(char, '')
|
|
|
|
|
|
|
|
return name
|
|
|
|
|
|
|
|
|
|
|
|
def replace_extension(filename, new_ext):
|
|
|
|
# type: (AnyStr, AnyStr) -> AnyStr
|
|
|
|
"""
|
|
|
|
:param filename: filename
|
|
|
|
:param new_ext: new extension
|
|
|
|
:return: filename with new extension
|
|
|
|
"""
|
|
|
|
sep_file = filename.rpartition('.')
|
|
|
|
if '' == sep_file[0]:
|
|
|
|
return filename
|
|
|
|
return '%s.%s' % (sep_file[0], new_ext)
|
|
|
|
|
|
|
|
|
|
|
|
def long_path(path):
|
|
|
|
# type: (AnyStr) -> AnyStr
|
|
|
|
"""add long path prefix for Windows"""
|
2023-02-10 14:15:50 +00:00
|
|
|
if 'nt' == os.name and 260 < len(path) and not path.startswith('\\\\?\\') and os.path.isabs(path):
|
2023-01-12 01:04:47 +00:00
|
|
|
return '\\\\?\\' + path
|
|
|
|
return path
|
|
|
|
|
|
|
|
|
|
|
|
def compress_file(target, filename, prefer_7z=True, remove_source=True):
|
|
|
|
# type: (AnyStr, AnyStr, bool, bool) -> bool
|
|
|
|
"""
|
|
|
|
compress given file to zip or 7z archive
|
|
|
|
|
|
|
|
:param target: file to compress with full path
|
|
|
|
:param filename: filename inside the archive
|
|
|
|
:param prefer_7z: prefer 7z over zip compression if available
|
|
|
|
:param remove_source: remove source file after successful creation of archive
|
|
|
|
:return: success of compression
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
if prefer_7z and None is not py7zr:
|
|
|
|
z_name = '%s.7z' % target.rpartition('.')[0]
|
|
|
|
# noinspection PyUnresolvedReferences
|
|
|
|
with py7zr.SevenZipFile(z_name, 'w') as z_file:
|
|
|
|
z_file.write(target, filename)
|
|
|
|
else:
|
|
|
|
zip_name = '%s.zip' % target.rpartition('.')[0]
|
|
|
|
with zipfile.ZipFile(zip_name, 'w', zipfile.ZIP_DEFLATED) as zip_fh:
|
|
|
|
zip_fh.write(target, filename)
|
|
|
|
except (BaseException, Exception) as e:
|
|
|
|
logger.error('error compressing %s' % target)
|
|
|
|
logger.debug('traceback: %s' % ex(e))
|
|
|
|
return False
|
|
|
|
if remove_source:
|
|
|
|
remove_file_perm(target)
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
|
|
|
def cmdline_runner(cmd, shell=False, suppress_stderr=False, env=None):
|
|
|
|
# type: (Union[AnyStr, List[AnyStr]], bool, bool, Dict) -> Tuple[AnyStr, Optional[AnyStr], int]
|
|
|
|
""" Execute a child program in a new process.
|
|
|
|
|
|
|
|
Can raise an exception to be caught in callee
|
|
|
|
|
|
|
|
:param cmd: A string, or a sequence of program arguments
|
|
|
|
:param shell: If true, the command will be executed through the shell.
|
|
|
|
:param suppress_stderr: Suppress stderr output if True
|
|
|
|
:param env: added env vars
|
|
|
|
"""
|
|
|
|
# noinspection PyUnresolvedReferences
|
|
|
|
kw = dict(cwd=PROG_DIR, shell=shell, stdin=subprocess.PIPE, stdout=subprocess.PIPE,
|
|
|
|
stderr=(open(os.devnull, 'w') if PY2 else subprocess.DEVNULL, subprocess.STDOUT)[not suppress_stderr])
|
|
|
|
|
|
|
|
if isinstance(env, dict):
|
|
|
|
kw.update(env=dict(os.environ, **env))
|
|
|
|
|
2023-02-10 14:15:50 +00:00
|
|
|
kw.update(dict(encoding=SYS_ENCODING, text=True, bufsize=0))
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
if 'win32' == sys.platform:
|
|
|
|
kw['creationflags'] = 0x08000000 # CREATE_NO_WINDOW (needed for py2exe)
|
|
|
|
|
|
|
|
with Popen(cmd, **kw) as p:
|
|
|
|
out, err = p.communicate()
|
|
|
|
if out:
|
|
|
|
out = out.strip()
|
|
|
|
|
|
|
|
return out, err, p.returncode
|
|
|
|
|
|
|
|
|
|
|
|
def md5_for_text(text):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param text: test
|
|
|
|
:type text: AnyStr
|
|
|
|
:return:
|
|
|
|
:rtype: AnyStr or None
|
|
|
|
"""
|
|
|
|
result = None
|
|
|
|
try:
|
|
|
|
md5 = hashlib.md5()
|
|
|
|
md5.update(decode_bytes(str(text)))
|
|
|
|
raw_md5 = md5.hexdigest()
|
|
|
|
result = raw_md5[17:] + raw_md5[9:17] + raw_md5[0:9]
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
|
|
|
def maybe_plural(subject=1):
|
|
|
|
"""
|
|
|
|
returns 's' or '' depending on numeric subject or length of subject
|
|
|
|
|
|
|
|
:param subject: number or list or dict
|
|
|
|
:type subject: int or list or dict
|
|
|
|
:return: returns s or ''
|
|
|
|
:rtype: AnyStr
|
|
|
|
"""
|
|
|
|
number = subject if not isinstance(subject, (list, dict)) else len(subject)
|
|
|
|
return ('s', '')[1 == number]
|
|
|
|
|
|
|
|
|
|
|
|
def time_to_int(dt):
|
|
|
|
# type: (Union[datetime.time, None]) -> Optional[integer_types]
|
|
|
|
"""
|
|
|
|
converts datetime.time to integer (hour + minute only)
|
|
|
|
|
|
|
|
:param dt: datetime.time obj
|
|
|
|
:return: integer of hour + min
|
|
|
|
"""
|
|
|
|
if None is dt:
|
|
|
|
return None
|
|
|
|
try:
|
|
|
|
return dt.hour * 100 + dt.minute
|
|
|
|
except (BaseException, Exception):
|
|
|
|
return 0
|
|
|
|
|
|
|
|
|
|
|
|
def int_to_time(d_int):
|
|
|
|
# type: (Union[integer_types, None]) -> Optional[datetime.time]
|
|
|
|
"""
|
|
|
|
convert integer from dt_to_int back to datetime.time
|
|
|
|
|
|
|
|
"""
|
|
|
|
if None is d_int:
|
|
|
|
return None
|
|
|
|
if isinstance(d_int, integer_types):
|
|
|
|
try:
|
|
|
|
return datetime.time(*divmod(d_int, 100))
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
return datetime.time(hour=0, minute=0)
|
|
|
|
|
|
|
|
|
|
|
|
def indent_xml(elem, level=0):
|
|
|
|
"""
|
|
|
|
Does our pretty printing, makes Matt very happy
|
|
|
|
"""
|
|
|
|
i = '\n' + level * ' '
|
|
|
|
if len(elem):
|
|
|
|
if not elem.text or not ('%s' % elem.text).strip():
|
|
|
|
elem.text = i + ' '
|
|
|
|
if not elem.tail or not elem.tail.strip():
|
|
|
|
elem.tail = i
|
|
|
|
for elem in elem:
|
|
|
|
indent_xml(elem, level + 1)
|
|
|
|
if not elem.tail or not elem.tail.strip():
|
|
|
|
elem.tail = i
|
|
|
|
else:
|
|
|
|
# Strip out the newlines from text
|
|
|
|
if elem.text:
|
|
|
|
elem.text = ('%s' % elem.text).replace('\n', ' ')
|
|
|
|
if level and (not elem.tail or not elem.tail.strip()):
|
|
|
|
elem.tail = i
|
|
|
|
|
|
|
|
|
|
|
|
def ast_eval(value, default=None):
|
|
|
|
# type: (AnyStr, Any) -> Any
|
|
|
|
"""Convert string typed value into actual Python type and value
|
|
|
|
|
|
|
|
:param value: string value to convert
|
2023-02-13 21:00:11 +00:00
|
|
|
:param default: value to return if it cannot convert
|
2023-01-12 01:04:47 +00:00
|
|
|
:return: converted type and value or default
|
|
|
|
"""
|
|
|
|
if not isinstance(value, string_types):
|
|
|
|
return default
|
|
|
|
|
|
|
|
if 'OrderedDict()' == value:
|
2023-02-11 18:02:58 +00:00
|
|
|
value = dict()
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
elif 'OrderedDict([(' == value[0:14]:
|
|
|
|
try:
|
|
|
|
list_of_tuples = ast.literal_eval(value[12:-1])
|
2023-02-11 18:02:58 +00:00
|
|
|
value = dict()
|
2023-01-12 01:04:47 +00:00
|
|
|
for cur_tuple in list_of_tuples:
|
|
|
|
value[cur_tuple[0]] = cur_tuple[1]
|
|
|
|
except (BaseException, Exception):
|
|
|
|
value = default
|
|
|
|
|
|
|
|
elif '{' == value[0:1] and '}' == value[-1]: # this way avoids index out of range with (value = '' and [-1])
|
|
|
|
try:
|
|
|
|
value = ast.literal_eval(value)
|
|
|
|
except (BaseException, Exception):
|
|
|
|
value = default
|
|
|
|
|
|
|
|
else:
|
|
|
|
value = default
|
|
|
|
|
|
|
|
return value
|
|
|
|
|
|
|
|
|
|
|
|
def download_file(url, filename, session=None, **kwargs):
|
|
|
|
"""
|
|
|
|
download given url to given filename
|
|
|
|
|
|
|
|
:param url: url to download
|
|
|
|
:type url: AnyStr
|
|
|
|
:param filename: filename to save the data to
|
|
|
|
:type filename: AnyStr
|
|
|
|
:param session: optional requests session object
|
|
|
|
:type session: requests.Session or None
|
|
|
|
:param kwargs:
|
|
|
|
:return: success of download
|
|
|
|
:rtype: bool
|
|
|
|
"""
|
|
|
|
MEMCACHE.setdefault('cookies', {})
|
|
|
|
if None is get_url(url, session=session, savename=filename,
|
|
|
|
url_solver=FLARESOLVERR_HOST, memcache_cookies=MEMCACHE['cookies'],
|
|
|
|
**kwargs):
|
|
|
|
remove_file_perm(filename)
|
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
|
|
|
def calc_age(birthday, deathday=None, date=None):
|
|
|
|
# type: (datetime.date, datetime.date, Optional[datetime.date]) -> Optional[int]
|
|
|
|
"""
|
|
|
|
returns age based on current date or given date
|
2023-02-13 21:00:11 +00:00
|
|
|
:param birthday: birthdate
|
|
|
|
:param deathday: deathdate
|
2023-01-12 01:04:47 +00:00
|
|
|
:param date:
|
|
|
|
"""
|
|
|
|
if isinstance(birthday, datetime.date):
|
|
|
|
today = (datetime.date.today(), date)[isinstance(date, datetime.date)]
|
|
|
|
today = (today, deathday)[isinstance(deathday, datetime.date) and today > deathday]
|
|
|
|
try:
|
|
|
|
b_d = birthday.replace(year=today.year)
|
|
|
|
|
2023-02-13 21:00:11 +00:00
|
|
|
# raised when birthdate is February 29
|
2023-01-12 01:04:47 +00:00
|
|
|
# and the current year is not a leap year
|
|
|
|
except ValueError:
|
|
|
|
b_d = birthday.replace(year=today.year, month=birthday.month + 1, day=1)
|
|
|
|
|
|
|
|
if b_d > today:
|
|
|
|
return today.year - birthday.year - 1
|
|
|
|
else:
|
|
|
|
return today.year - birthday.year
|
|
|
|
|
|
|
|
|
|
|
|
def convert_to_inch_faction_html(height):
|
|
|
|
# type: (float) -> AnyStr
|
|
|
|
"""
|
|
|
|
returns html string in foot and inches including fractions
|
|
|
|
:param height: height in cm
|
|
|
|
"""
|
|
|
|
total_inches = round(height / float(2.54), 2)
|
|
|
|
foot, inches = divmod(total_inches, 12)
|
|
|
|
_, fraction = '{0:.2f}'.format(total_inches).split('.')
|
|
|
|
fraction = int(fraction)
|
|
|
|
# fix rounding errors
|
|
|
|
fraction = next((html_convert_fractions.get(fraction + round_error)
|
|
|
|
or html_convert_fractions.get(fraction - round_error)
|
|
|
|
for round_error in moves.xrange(0, 25) if fraction + round_error in html_convert_fractions
|
|
|
|
or fraction - round_error in html_convert_fractions), '')
|
|
|
|
if 1 == fraction:
|
|
|
|
inches += 1
|
|
|
|
fraction = ''
|
|
|
|
if 12 <= inches:
|
|
|
|
foot += 1
|
|
|
|
inches = 0
|
|
|
|
inches = str(inches).split('.')[0]
|
|
|
|
return '%s\' %s%s%s' % (int(foot), (inches, '')['0' == inches], fraction,
|
|
|
|
('', '"')['0' != inches or '' != fraction])
|
|
|
|
|
|
|
|
|
|
|
|
def spoken_height(height):
|
|
|
|
# type: (float) -> AnyStr
|
|
|
|
"""
|
|
|
|
return text for spoken words of height
|
|
|
|
|
|
|
|
:param height: height in cm
|
|
|
|
"""
|
|
|
|
return convert_to_inch_faction_html(height).replace('\'', ' foot').replace('"', '')
|
|
|
|
|
|
|
|
|
|
|
|
def iterate_chunk(lst, n):
|
|
|
|
# type: (List, integer_types) -> Iterator
|
|
|
|
"""
|
|
|
|
Yield successive n-sized chunks from lst.
|
|
|
|
:param lst: list to split
|
|
|
|
:param n: length of chunks
|
|
|
|
"""
|
|
|
|
for i in moves.range(0, len(lst), n):
|
|
|
|
yield lst[i:i + n]
|
2023-02-15 13:37:20 +00:00
|
|
|
|
|
|
|
|
|
|
|
def is_virtualenv():
|
|
|
|
# type: (...) -> bool
|
|
|
|
"""
|
|
|
|
:return: True if virtualenv Python environment is detected
|
|
|
|
"""
|
|
|
|
"""Get base/real prefix, or `sys.prefix` if there is none."""
|
|
|
|
get_base_prefix_compat = getattr(sys, 'base_prefix', None) or getattr(sys, 'real_prefix', None) or sys.prefix
|
|
|
|
return get_base_prefix_compat != sys.prefix
|
Change improve tmdb_api, trakt_api, and TVInfoShow object.
Add `spoken_languages` to tmdb API and TVInfoShow object.
Add `trailers`, `homepage` to trakt API and TVInfoShow object.
Add trakt episode data if returned from api.
Add trakt API methods.
- get_most_played
- get_most_watched
- get_most_collected
- get_anticipated
- get_recommended
- get_trending
- get_popular
- get_recommended_for_account
- get_new_shows
- get_new_seasons
- get_watchlisted_for_account
- get_similar
- hide_recommended_for_account (to hide/remove recommended shows for account)
- unhide_recommended_for_account
- list_hidden_recommended_for_account
Fix caching tmdb language list over different runtime instances.
Add episode_count and fix ti_show in tmdb_api person object.
Change set additional properties in get_person trakt_api.
Add tmdb API methods and tvinfo_base.
- get_recommended_for_show
- get_similar
---
fix supported language caching
improve print output (source name) of tvinfo_api_tests
fix tvinfo_api_tests data creation
---
Add code so that it runs with all_test
use mock today() and now() dates
add option to only get new urls mock data
try also to make object creation only when needed
fix person parser in tmdb_api
add search_person test in tvinfo_api_tests
restore mocked methods at the end of the tvinfo_api_tests to prevent other tests to fail when called via all_tests
switch gzip with better lzma compression for mock files (default lib in py3)
move mock files in test unit sub folder
---
Fix trakt method `get_recommended`.
Fix browse trakt tests in tvinfo_api_tests.
Change set episode id in trakt api.
---
Add test_browse_endpoints to tvinfo_api_tests.
---
Add enforce_type to sg_helpers.
Change use enforce str for overviews.
Change remove `if PY2` code sections
Add support for datetime.time in _make_airtime in tv.py
Refactor tvmaze_api show data setter.
Change test to not allow None for seriesname.
Add additional missing showdata with caller load_data().
Add load_data() to TVInfoShow.
Add guestcast, guestcrew to episodes in pytvmaze lib.
---
Change make seriesid of TVInfoShow a alias property of id.
Add tvinfo tests.
Add search tests.
Add show, person tests.
Change add trakt tests.
Change add tmdb search tests.
tvmaze_api exclude rating from mapping.
Allow None for seriesname.
Fix origin_countries in trakt_api search.
Fix show_type in tvmaze_api.
Fix airtime for episodes in tvmaze_api.
---
Change switch to property instead of legacy dict-like use for trakt search results.
Change optimize speed of get() function.
Fix make BaseTVinfoSeasonnotfound and BaseTVinfoAttributenotfound also a subclass of AttributeError and KeyError.
Change mock get() to work with and without default args just like dict get().
Change add language to tmdb_api search results.
Change improve person search by remote id, by getting the complete persons data when there is only 1 result.
Change trakt API search results to tvinfoshow.
Change search results to TVInfoShow objs in tvmaze_api.
Change simplify poster URL generation for search results.
Change search results to TVInfoShow objs.
Change add tvdb genre links to displayShow.
Change workaround for missing data in person data (series set to None).
Fix add show to characters of person if there is no name on IMDb (set to 'unknown name').
Change add config and icons for linkedin, reddit, wikidata, youtube.
Add TVInfoIDs, TVInfoSocialIDs to Trakt.
Add TVInfoIDs to tmdb_api.
Add TVInfoIDs to tvmaze_api.
add TVInfoIDs to imdb_api.
Change make character name '' if None.
Fix for 'unknown name' persons and characters.
Add contentrating.
Change fill in new fields to get_person results.
----
Change set new in/active dates to network.
Change add active_date, inactive_date to TVInfoNetwork class.
Change add default kwargs to tmdb discover method if no kwargs are set.
Change default: English language shows with first air date greater then today.
Change add slug field to returned data from discover.
Change add 'score' mapped to rating to discover returned results.
Fix valid_data for discover method.
Change add result_count to discover.
Change add _sanitise_image_uri to discover method.
Fix convert_person.
Change add missing _sanitise_image_uri for images in some places.
Fix crew.
Change return type of tvinfo base: discover to list tvinfoshow.
Fix people remote id search.
Change add tmdb person id search.
Change fix people endpoint fieldname changes.
Change add biography to person object.
Change move 401 expired token handling into TvdbAuth class.
Change get new token if old token is expired.
Change add raise error if episodes fallback fails to load data.
Change add break if no valid_data to absolute and alternative numberings.
Change add filter only networks.
Change add new required parameter meta=translations to get translated (includes the original language) show overviews.
Change add check if show is set for person compare.
Fix person update properties with no show set.
Change add person image.
Change add alternative episode orders.
Change add alt_ep_numbering to TVINFO_Show.
Change add old interface for dvd order.
Change add trakt slug tvinfo search test cases.
Change add mock for old tvdb get new token.
Change old lib to newer tvinfo data.
Fix person id (not available on old api).
Change more places to new TVInfoAPI interface.
2021-09-16 20:03:19 +00:00
|
|
|
|
|
|
|
|
|
|
|
def enforce_type(value, allowed_types, default):
|
|
|
|
# type: (Any, Union[Type, Tuple[Type]], Any) -> Any
|
|
|
|
"""
|
|
|
|
enforces that value is given type(s)
|
|
|
|
:param value: value to check
|
|
|
|
:param allowed_types: type or tuple of types allowed
|
|
|
|
:param default: value to return if other type
|
|
|
|
"""
|
|
|
|
if not isinstance(value, allowed_types):
|
|
|
|
return default
|
|
|
|
return value
|