mirror of
https://github.com/SickGear/SickGear.git
synced 2024-12-11 21:53:37 +00:00
1721 lines
72 KiB
Python
1721 lines
72 KiB
Python
# coding=utf-8
|
|
# Author: Nic Wolfe <nic@wolfeden.ca>
|
|
# URL: http://code.google.com/p/sickbeard/
|
|
#
|
|
# This file is part of SickGear.
|
|
#
|
|
# SickGear is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# SickGear is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
from __future__ import with_statement
|
|
|
|
import datetime
|
|
import itertools
|
|
import math
|
|
import os
|
|
import re
|
|
import time
|
|
import urlparse
|
|
import threading
|
|
import socket
|
|
from urllib import quote_plus
|
|
import zlib
|
|
from base64 import b16encode, b32decode
|
|
|
|
import sickbeard
|
|
import requests
|
|
import requests.cookies
|
|
from cfscrape import CloudflareScraper
|
|
from hachoir.parser import guessParser
|
|
from hachoir.stream import FileInputStream
|
|
|
|
from sickbeard import helpers, classes, logger, db, tvcache, encodingKludge as ek
|
|
from sickbeard.common import Quality, MULTI_EP_RESULT, SEASON_RESULT, USER_AGENT
|
|
from sickbeard.exceptions import SickBeardException, AuthException, ex
|
|
from sickbeard.helpers import maybe_plural, remove_file_failed
|
|
from sickbeard.name_parser.parser import NameParser, InvalidNameException, InvalidShowException
|
|
from sickbeard.show_name_helpers import get_show_names_all_possible
|
|
from sickbeard.sbdatetime import sbdatetime
|
|
|
|
|
|
class HaltParseException(SickBeardException):
|
|
"""Something requires the current processing to abort"""
|
|
|
|
|
|
class ProviderFailTypes:
|
|
http = 1
|
|
connection = 2
|
|
connection_timeout = 3
|
|
timeout = 4
|
|
other = 5
|
|
limit = 6
|
|
nodata = 7
|
|
|
|
names = {http: 'http', timeout: 'timeout',
|
|
connection: 'connection', connection_timeout: 'connection_timeout',
|
|
nodata: 'nodata', other: 'other', limit: 'limit'}
|
|
|
|
def __init__(self):
|
|
pass
|
|
|
|
|
|
class ProviderFail(object):
|
|
def __init__(self, fail_type=ProviderFailTypes.other, code=None, fail_time=None):
|
|
self.code = code
|
|
self.fail_type = fail_type
|
|
self.fail_time = (datetime.datetime.now(), fail_time)[isinstance(fail_time, datetime.datetime)]
|
|
|
|
|
|
class ProviderFailList(object):
|
|
def __init__(self, provider_name):
|
|
self.provider_name = provider_name
|
|
self._fails = []
|
|
self.lock = threading.Lock()
|
|
self.clear_old()
|
|
self.load_list()
|
|
self.last_save = datetime.datetime.now()
|
|
self.dirty = False
|
|
|
|
@property
|
|
def fails(self):
|
|
return self._fails
|
|
|
|
@property
|
|
def fails_sorted(self):
|
|
fail_dict = {}
|
|
b_d = {'count': 0}
|
|
for e in self._fails:
|
|
fail_date = e.fail_time.date()
|
|
fail_hour = e.fail_time.time().hour
|
|
date_time = datetime.datetime.combine(fail_date, datetime.time(hour=fail_hour))
|
|
if ProviderFailTypes.names[e.fail_type] not in fail_dict.get(date_time, {}):
|
|
default = {'date': str(fail_date), 'date_time': date_time,
|
|
'timestamp': helpers.tryInt(sbdatetime.totimestamp(e.fail_time)), 'multirow': False}
|
|
for et in ProviderFailTypes.names.itervalues():
|
|
default[et] = b_d.copy()
|
|
fail_dict.setdefault(date_time, default)[ProviderFailTypes.names[e.fail_type]]['count'] = 1
|
|
else:
|
|
fail_dict[date_time][ProviderFailTypes.names[e.fail_type]]['count'] += 1
|
|
if ProviderFailTypes.http == e.fail_type:
|
|
if e.code in fail_dict[date_time].get(ProviderFailTypes.names[e.fail_type],
|
|
{'code': {}}).get('code', {}):
|
|
fail_dict[date_time][ProviderFailTypes.names[e.fail_type]]['code'][e.code] += 1
|
|
else:
|
|
fail_dict[date_time][ProviderFailTypes.names[e.fail_type]].setdefault('code', {})[e.code] = 1
|
|
|
|
row_count = {}
|
|
for (k, v) in fail_dict.iteritems():
|
|
row_count.setdefault(v.get('date'), 0)
|
|
if v.get('date') in row_count:
|
|
row_count[v.get('date')] += 1
|
|
for (k, v) in fail_dict.iteritems():
|
|
if 1 < row_count.get(v.get('date')):
|
|
fail_dict[k]['multirow'] = True
|
|
|
|
fail_list = sorted([fail_dict[k] for k in fail_dict.iterkeys()], key=lambda y: y.get('date_time'), reverse=True)
|
|
|
|
totals = {}
|
|
for fail_date in set([fail.get('date') for fail in fail_list]):
|
|
daytotals = {}
|
|
for et in ProviderFailTypes.names.itervalues():
|
|
daytotals.update({et: sum([x.get(et).get('count') for x in fail_list if fail_date == x.get('date')])})
|
|
totals.update({fail_date: daytotals})
|
|
for (fail_date, total) in totals.iteritems():
|
|
for i, item in enumerate(fail_list):
|
|
if fail_date == item.get('date'):
|
|
if item.get('multirow'):
|
|
fail_list[i:i] = [item.copy()]
|
|
for et in ProviderFailTypes.names.itervalues():
|
|
fail_list[i][et] = {'count': total[et]}
|
|
if et == ProviderFailTypes.names[ProviderFailTypes.http]:
|
|
fail_list[i][et]['code'] = {}
|
|
break
|
|
|
|
return fail_list
|
|
|
|
def add_fail(self, fail):
|
|
if isinstance(fail, ProviderFail):
|
|
with self.lock:
|
|
self.dirty = True
|
|
self._fails.append(fail)
|
|
logger.log('Adding fail.%s for %s' % (ProviderFailTypes.names.get(
|
|
fail.fail_type, ProviderFailTypes.names[ProviderFailTypes.other]), self.provider_name()),
|
|
logger.DEBUG)
|
|
self.save_list()
|
|
|
|
def save_list(self):
|
|
if self.dirty:
|
|
self.clear_old()
|
|
with self.lock:
|
|
my_db = db.DBConnection('cache.db')
|
|
cl = []
|
|
for f in self._fails:
|
|
cl.append(['INSERT OR IGNORE INTO provider_fails (prov_name, fail_type, fail_code, fail_time) '
|
|
'VALUES (?,?,?,?)', [self.provider_name(), f.fail_type, f.code,
|
|
sbdatetime.totimestamp(f.fail_time)]])
|
|
self.dirty = False
|
|
if cl:
|
|
my_db.mass_action(cl)
|
|
self.last_save = datetime.datetime.now()
|
|
|
|
def load_list(self):
|
|
with self.lock:
|
|
try:
|
|
my_db = db.DBConnection('cache.db')
|
|
if my_db.hasTable('provider_fails'):
|
|
results = my_db.select('SELECT * FROM provider_fails WHERE prov_name = ?', [self.provider_name()])
|
|
self._fails = []
|
|
for r in results:
|
|
try:
|
|
self._fails.append(ProviderFail(
|
|
fail_type=helpers.tryInt(r['fail_type']), code=helpers.tryInt(r['fail_code']),
|
|
fail_time=datetime.datetime.fromtimestamp(helpers.tryInt(r['fail_time']))))
|
|
except (StandardError, Exception):
|
|
continue
|
|
except (StandardError, Exception):
|
|
pass
|
|
|
|
def clear_old(self):
|
|
with self.lock:
|
|
try:
|
|
my_db = db.DBConnection('cache.db')
|
|
if my_db.hasTable('provider_fails'):
|
|
time_limit = sbdatetime.totimestamp(datetime.datetime.now() - datetime.timedelta(days=28))
|
|
my_db.action('DELETE FROM provider_fails WHERE fail_time < ?', [time_limit])
|
|
except (StandardError, Exception):
|
|
pass
|
|
|
|
|
|
class GenericProvider(object):
|
|
NZB = 'nzb'
|
|
TORRENT = 'torrent'
|
|
|
|
def __init__(self, name, supports_backlog=False, anime_only=False):
|
|
# these need to be set in the subclass
|
|
self.providerType = None
|
|
self.name = name
|
|
self.supports_backlog = supports_backlog
|
|
self.anime_only = anime_only
|
|
if anime_only:
|
|
self.proper_search_terms = 'v1|v2|v3|v4|v5'
|
|
self.url = ''
|
|
|
|
self.show = None
|
|
|
|
self.search_mode = None
|
|
self.search_fallback = False
|
|
self.enabled = False
|
|
self.enable_recentsearch = False
|
|
self.enable_backlog = False
|
|
self.enable_scheduled_backlog = True
|
|
self.categories = None
|
|
|
|
self.cache = tvcache.TVCache(self)
|
|
|
|
self.session = CloudflareScraper.create_scraper()
|
|
|
|
self.headers = {
|
|
# Using USER_AGENT instead of Mozilla to keep same user agent along authentication and download phases,
|
|
# otherwise session might be broken and download fail, asking again for authentication
|
|
# 'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) ' +
|
|
# 'Chrome/32.0.1700.107 Safari/537.36'}
|
|
'User-Agent': USER_AGENT}
|
|
|
|
self._failure_count = 0
|
|
self._failure_time = None
|
|
self.fails = ProviderFailList(self.get_id)
|
|
self._tmr_limit_count = 0
|
|
self._tmr_limit_time = None
|
|
self._tmr_limit_wait = None
|
|
self._last_fail_type = None
|
|
self.has_limit = False
|
|
self.fail_times = {1: (0, 15), 2: (0, 30), 3: (1, 0), 4: (2, 0), 5: (3, 0), 6: (6, 0), 7: (12, 0), 8: (24, 0)}
|
|
self._load_fail_values()
|
|
|
|
def _load_fail_values(self):
|
|
if hasattr(sickbeard, 'DATA_DIR'):
|
|
my_db = db.DBConnection('cache.db')
|
|
if my_db.hasTable('provider_fails_count'):
|
|
r = my_db.select('SELECT * FROM provider_fails_count WHERE prov_name = ?', [self.get_id()])
|
|
if r:
|
|
self._failure_count = helpers.tryInt(r[0]['failure_count'], 0)
|
|
if r[0]['failure_time']:
|
|
self._failure_time = datetime.datetime.fromtimestamp(r[0]['failure_time'])
|
|
else:
|
|
self._failure_time = None
|
|
self._tmr_limit_count = helpers.tryInt(r[0]['tmr_limit_count'], 0)
|
|
if r[0]['tmr_limit_time']:
|
|
self._tmr_limit_time = datetime.datetime.fromtimestamp(r[0]['tmr_limit_time'])
|
|
else:
|
|
self._tmr_limit_time = None
|
|
if r[0]['tmr_limit_wait']:
|
|
self._tmr_limit_wait = datetime.timedelta(seconds=helpers.tryInt(r[0]['tmr_limit_wait'], 0))
|
|
else:
|
|
self._tmr_limit_wait = None
|
|
self._last_fail_type = self.last_fail
|
|
|
|
def _save_fail_value(self, field, value):
|
|
my_db = db.DBConnection('cache.db')
|
|
if my_db.hasTable('provider_fails_count'):
|
|
r = my_db.action('UPDATE provider_fails_count SET %s = ? WHERE prov_name = ?' % field,
|
|
[value, self.get_id()])
|
|
if 0 == r.rowcount:
|
|
my_db.action('REPLACE INTO provider_fails_count (prov_name, %s) VALUES (?,?)' % field,
|
|
[self.get_id(), value])
|
|
|
|
@property
|
|
def last_fail(self):
|
|
try:
|
|
return sorted(self.fails.fails, key=lambda x: x.fail_time, reverse=True)[0].fail_type
|
|
except (StandardError, Exception):
|
|
return None
|
|
|
|
@property
|
|
def failure_count(self):
|
|
return self._failure_count
|
|
|
|
@failure_count.setter
|
|
def failure_count(self, value):
|
|
changed_val = self._failure_count != value
|
|
self._failure_count = value
|
|
if changed_val:
|
|
self._save_fail_value('failure_count', value)
|
|
|
|
@property
|
|
def failure_time(self):
|
|
return self._failure_time
|
|
|
|
@failure_time.setter
|
|
def failure_time(self, value):
|
|
if None is value or isinstance(value, datetime.datetime):
|
|
changed_val = self._failure_time != value
|
|
self._failure_time = value
|
|
if changed_val:
|
|
self._save_fail_value('failure_time', (sbdatetime.totimestamp(value), value)[None is value])
|
|
|
|
@property
|
|
def tmr_limit_count(self):
|
|
return self._tmr_limit_count
|
|
|
|
@tmr_limit_count.setter
|
|
def tmr_limit_count(self, value):
|
|
changed_val = self._tmr_limit_count != value
|
|
self._tmr_limit_count = value
|
|
if changed_val:
|
|
self._save_fail_value('tmr_limit_count', value)
|
|
|
|
@property
|
|
def tmr_limit_time(self):
|
|
return self._tmr_limit_time
|
|
|
|
@tmr_limit_time.setter
|
|
def tmr_limit_time(self, value):
|
|
if None is value or isinstance(value, datetime.datetime):
|
|
changed_val = self._tmr_limit_time != value
|
|
self._tmr_limit_time = value
|
|
if changed_val:
|
|
self._save_fail_value('tmr_limit_time', (sbdatetime.totimestamp(value), value)[None is value])
|
|
|
|
@property
|
|
def max_index(self):
|
|
return len(self.fail_times)
|
|
|
|
@property
|
|
def tmr_limit_wait(self):
|
|
return self._tmr_limit_wait
|
|
|
|
@tmr_limit_wait.setter
|
|
def tmr_limit_wait(self, value):
|
|
if isinstance(getattr(self, 'fails', None), ProviderFailList) and isinstance(value, datetime.timedelta):
|
|
self.fails.add_fail(ProviderFail(fail_type=ProviderFailTypes.limit))
|
|
changed_val = self._tmr_limit_wait != value
|
|
self._tmr_limit_wait = value
|
|
if changed_val:
|
|
if None is value:
|
|
self._save_fail_value('tmr_limit_wait', value)
|
|
elif isinstance(value, datetime.timedelta):
|
|
self._save_fail_value('tmr_limit_wait', value.total_seconds())
|
|
|
|
def fail_time_index(self, base_limit=2):
|
|
i = self.failure_count - base_limit
|
|
return (i, self.max_index)[i >= self.max_index]
|
|
|
|
def tmr_limit_update(self, period, unit, desc):
|
|
self.tmr_limit_time = datetime.datetime.now()
|
|
self.tmr_limit_count += 1
|
|
limit_set = False
|
|
if None not in (period, unit):
|
|
limit_set = True
|
|
if unit in ('s', 'sec', 'secs', 'seconds', 'second'):
|
|
self.tmr_limit_wait = datetime.timedelta(seconds=helpers.tryInt(period))
|
|
elif unit in ('m', 'min', 'mins', 'minutes', 'minute'):
|
|
self.tmr_limit_wait = datetime.timedelta(minutes=helpers.tryInt(period))
|
|
elif unit in ('h', 'hr', 'hrs', 'hours', 'hour'):
|
|
self.tmr_limit_wait = datetime.timedelta(hours=helpers.tryInt(period))
|
|
elif unit in ('d', 'days', 'day'):
|
|
self.tmr_limit_wait = datetime.timedelta(days=helpers.tryInt(period))
|
|
else:
|
|
limit_set = False
|
|
if not limit_set:
|
|
time_index = self.fail_time_index(base_limit=0)
|
|
self.tmr_limit_wait = self.wait_time(time_index)
|
|
logger.log('Request limit reached. Waiting for %s until next retry. Message: %s' %
|
|
(self.tmr_limit_wait, desc or 'none found'), logger.WARNING)
|
|
|
|
def wait_time(self, time_index=None):
|
|
"""
|
|
Return a suitable wait time, selected by parameter, or based on the current failure count
|
|
|
|
:param time_index: A key value index into the fail_times dict, or selects using failure count if None
|
|
:type time_index: Integer
|
|
:return: Time
|
|
:rtype: Timedelta
|
|
"""
|
|
if None is time_index:
|
|
time_index = self.fail_time_index()
|
|
return datetime.timedelta(hours=self.fail_times[time_index][0], minutes=self.fail_times[time_index][1])
|
|
|
|
def fail_newest_delta(self):
|
|
"""
|
|
Return how long since most recent failure
|
|
:return: Period since most recent failure on record
|
|
:rtype: timedelta
|
|
"""
|
|
return datetime.datetime.now() - self.failure_time
|
|
|
|
def is_waiting(self):
|
|
return self.fail_newest_delta() < self.wait_time()
|
|
|
|
def valid_tmr_time(self):
|
|
return isinstance(self.tmr_limit_wait, datetime.timedelta) and \
|
|
isinstance(self.tmr_limit_time, datetime.datetime)
|
|
|
|
@property
|
|
def get_next_try_time(self):
|
|
n = None
|
|
h = datetime.timedelta(seconds=0)
|
|
f = datetime.timedelta(seconds=0)
|
|
if self.valid_tmr_time():
|
|
h = self.tmr_limit_time + self.tmr_limit_wait - datetime.datetime.now()
|
|
if 3 <= self.failure_count and isinstance(self.failure_time, datetime.datetime) and self.is_waiting():
|
|
h = self.failure_time + self.wait_time() - datetime.datetime.now()
|
|
if datetime.timedelta(seconds=0) < max((h, f)):
|
|
n = max((h, f))
|
|
return n
|
|
|
|
def retry_next(self):
|
|
if self.valid_tmr_time():
|
|
self.tmr_limit_time = datetime.datetime.now() - self.tmr_limit_wait
|
|
if 3 <= self.failure_count and isinstance(self.failure_time, datetime.datetime) and self.is_waiting():
|
|
self.failure_time = datetime.datetime.now() - self.wait_time()
|
|
|
|
@staticmethod
|
|
def fmt_delta(delta):
|
|
return str(delta).rsplit('.')[0]
|
|
|
|
def should_skip(self, log_warning=True, use_tmr_limit=True):
|
|
"""
|
|
Determine if a subsequent server request should be skipped. The result of this logic is based on most recent
|
|
server connection activity including, exhausted request limits, and counting connect failures to determine a
|
|
"cool down" period before recommending reconnection attempts; by returning False.
|
|
:param log_warning: Output to log if True (default) otherwise set False for no output.
|
|
:type log_warning: Boolean
|
|
:param use_tmr_limit: Setting this to False will ignore a tmr limit being reached and will instead return False.
|
|
:type use_tmr_limit: Boolean
|
|
:return: True for any known issue that would prevent a subsequent server connection, otherwise False.
|
|
:rtype: Boolean
|
|
"""
|
|
if self.valid_tmr_time():
|
|
time_left = self.tmr_limit_time + self.tmr_limit_wait - datetime.datetime.now()
|
|
if time_left > datetime.timedelta(seconds=0):
|
|
if log_warning:
|
|
# Ensure provider name output (e.g. when displaying config/provs) instead of e.g. thread "Tornado"
|
|
prepend = ('[%s] :: ' % self.name, '')[any([x.name in threading.currentThread().getName()
|
|
for x in sickbeard.providers.sortedProviderList()])]
|
|
logger.log('%sToo many requests reached at %s, waiting for %s' % (
|
|
prepend, self.fmt_delta(self.tmr_limit_time), self.fmt_delta(time_left)), logger.WARNING)
|
|
return use_tmr_limit
|
|
else:
|
|
self.tmr_limit_time = None
|
|
self.tmr_limit_wait = None
|
|
if 3 <= self.failure_count:
|
|
if None is self.failure_time:
|
|
self.failure_time = datetime.datetime.now()
|
|
if self.is_waiting():
|
|
if log_warning:
|
|
time_left = self.wait_time() - self.fail_newest_delta()
|
|
logger.log('Failed %s times, skipping provider for %s, last failure at %s with fail type: %s' % (
|
|
self.failure_count, self.fmt_delta(time_left), self.fmt_delta(self.failure_time),
|
|
ProviderFailTypes.names.get(
|
|
self.last_fail, ProviderFailTypes.names[ProviderFailTypes.other])), logger.WARNING)
|
|
return True
|
|
return False
|
|
|
|
def inc_failure_count(self, *args, **kwargs):
|
|
fail_type = ('fail_type' in kwargs and kwargs['fail_type'].fail_type) or \
|
|
(isinstance(args, tuple) and isinstance(args[0], ProviderFail) and args[0].fail_type)
|
|
if not isinstance(self.failure_time, datetime.datetime) or \
|
|
fail_type != self._last_fail_type or \
|
|
self.fail_newest_delta() > datetime.timedelta(seconds=3):
|
|
self.failure_count += 1
|
|
self.failure_time = datetime.datetime.now()
|
|
self._last_fail_type = fail_type
|
|
self.fails.add_fail(*args, **kwargs)
|
|
else:
|
|
logger.log('%s: Not logging same failure within 3 seconds' % self.name, logger.DEBUG)
|
|
|
|
def get_url(self, url, skip_auth=False, use_tmr_limit=True, *args, **kwargs):
|
|
"""
|
|
Return data from a URI with a possible check for authentication prior to the data fetch.
|
|
Raised errors and no data in responses are tracked for making future logic decisions.
|
|
|
|
:param url: Address where to fetch data from
|
|
:type url: String
|
|
:param skip_auth: Skip authentication check of provider if True
|
|
:type skip_auth: Boolean
|
|
:param use_tmr_limit: An API limit can be +ve before a fetch, but unwanted, set False to short should_skip
|
|
:type use_tmr_limit: Boolean
|
|
:param args: params to pass-through to getURL
|
|
:type args:
|
|
:param kwargs: keyword params to pass-through to getURL
|
|
:type kwargs:
|
|
:return: None or data fetched from URL
|
|
:rtype: String or Nonetype
|
|
"""
|
|
data = None
|
|
|
|
# check for auth
|
|
if (not skip_auth and not (self.is_public_access()
|
|
and type(self).__name__ not in ['TorrentRssProvider']) and not self._authorised()) \
|
|
or self.should_skip(use_tmr_limit=use_tmr_limit):
|
|
return
|
|
|
|
kwargs['raise_exceptions'] = True
|
|
kwargs['raise_status_code'] = True
|
|
for k, v in dict(headers=self.headers, hooks=dict(response=self.cb_response)).items():
|
|
kwargs.setdefault(k, v)
|
|
if 'nzbs.in' not in url: # this provider returns 503's 3 out of 4 requests with the persistent session system
|
|
kwargs.setdefault('session', self.session)
|
|
|
|
post_data = kwargs.get('post_data')
|
|
post_json = kwargs.get('post_json')
|
|
|
|
# noinspection PyUnusedLocal
|
|
log_failure_url = False
|
|
try:
|
|
data = helpers.getURL(url, *args, **kwargs)
|
|
if data and not isinstance(data, tuple) \
|
|
or isinstance(data, tuple) and data[0]:
|
|
if 0 != self.failure_count:
|
|
logger.log('Unblocking provider: %s' % self.get_id(), logger.DEBUG)
|
|
self.failure_count = 0
|
|
self.failure_time = None
|
|
else:
|
|
self.inc_failure_count(ProviderFail(fail_type=ProviderFailTypes.nodata))
|
|
log_failure_url = True
|
|
except requests.exceptions.HTTPError as e:
|
|
self.inc_failure_count(ProviderFail(fail_type=ProviderFailTypes.http, code=e.response.status_code))
|
|
except requests.exceptions.ConnectionError:
|
|
self.inc_failure_count(ProviderFail(fail_type=ProviderFailTypes.connection))
|
|
except requests.exceptions.ReadTimeout:
|
|
self.inc_failure_count(ProviderFail(fail_type=ProviderFailTypes.timeout))
|
|
except (requests.exceptions.Timeout, socket.timeout):
|
|
self.inc_failure_count(ProviderFail(fail_type=ProviderFailTypes.connection_timeout))
|
|
except (StandardError, Exception) as e:
|
|
log_failure_url = True
|
|
self.inc_failure_count(ProviderFail(fail_type=ProviderFailTypes.other))
|
|
|
|
self.fails.save_list()
|
|
if log_failure_url:
|
|
self.log_failure_url(url, post_data, post_json)
|
|
return data
|
|
|
|
def log_failure_url(self, url, post_data=None, post_json=None):
|
|
if self.should_skip(log_warning=False):
|
|
post = []
|
|
if post_data:
|
|
post += [' .. Post params: [%s]' % '&'.join([post_data])]
|
|
if post_json:
|
|
post += [' .. Json params: [%s]' % '&'.join([post_json])]
|
|
logger.log('Failure URL: %s%s' % (url, ''.join(post)), logger.WARNING)
|
|
|
|
def get_id(self):
|
|
return GenericProvider.make_id(self.name)
|
|
|
|
@staticmethod
|
|
def make_id(name):
|
|
return re.sub('[^\w\d_]', '_', name.strip().lower())
|
|
|
|
def image_name(self, *default_name):
|
|
|
|
for name in ['%s.%s' % (self.get_id(), image_ext) for image_ext in ['png', 'gif', 'jpg']]:
|
|
if ek.ek(os.path.isfile,
|
|
ek.ek(os.path.join, sickbeard.PROG_DIR, 'gui', sickbeard.GUI_NAME, 'images', 'providers', name)):
|
|
return name
|
|
|
|
return '%s.png' % ('newznab', default_name[0])[any(default_name)]
|
|
|
|
def _authorised(self):
|
|
return True
|
|
|
|
def _check_auth(self, is_required=None):
|
|
return True
|
|
|
|
def is_public_access(self):
|
|
try:
|
|
return bool(re.search('(?i)rarbg|sick|anizb', self.name)) \
|
|
or False is bool(('_authorised' in self.__class__.__dict__ or hasattr(self, 'digest')
|
|
or self._check_auth(is_required=True)))
|
|
except AuthException:
|
|
return False
|
|
|
|
def is_active(self):
|
|
if GenericProvider.NZB == self.providerType and sickbeard.USE_NZBS:
|
|
return self.is_enabled()
|
|
elif GenericProvider.TORRENT == self.providerType and sickbeard.USE_TORRENTS:
|
|
return self.is_enabled()
|
|
else:
|
|
return False
|
|
|
|
def is_enabled(self):
|
|
"""
|
|
This should be overridden and should return the config setting eg. sickbeard.MYPROVIDER
|
|
"""
|
|
return self.enabled
|
|
|
|
def get_result(self, episodes, url):
|
|
"""
|
|
Returns a result of the correct type for this provider
|
|
"""
|
|
|
|
if GenericProvider.NZB == self.providerType:
|
|
result = classes.NZBSearchResult(episodes)
|
|
elif GenericProvider.TORRENT == self.providerType:
|
|
result = classes.TorrentSearchResult(episodes)
|
|
else:
|
|
result = classes.SearchResult(episodes)
|
|
|
|
result.provider = self
|
|
result.url = url
|
|
|
|
return result
|
|
|
|
# noinspection PyUnusedLocal
|
|
def cb_response(self, r, *args, **kwargs):
|
|
self.session.response = dict(url=r.url, status_code=r.status_code, elapsed=r.elapsed, from_cache=r.from_cache)
|
|
return r
|
|
|
|
def download_result(self, result):
|
|
"""
|
|
Save the result to disk.
|
|
"""
|
|
|
|
# check for auth
|
|
if not self._authorised():
|
|
return False
|
|
|
|
if GenericProvider.TORRENT == self.providerType:
|
|
final_dir = sickbeard.TORRENT_DIR
|
|
link_type = 'magnet'
|
|
try:
|
|
btih = None
|
|
try:
|
|
btih = re.findall('urn:btih:([\w]{32,40})', result.url)[0]
|
|
if 32 == len(btih):
|
|
from base64 import b16encode, b32decode
|
|
btih = b16encode(b32decode(btih))
|
|
except (StandardError, Exception):
|
|
pass
|
|
|
|
if not btih or not re.search('(?i)[0-9a-f]{32,40}', btih):
|
|
assert not result.url.startswith('http')
|
|
logger.log('Unable to extract torrent hash from link: ' + ex(result.url), logger.ERROR)
|
|
return False
|
|
|
|
urls = ['http%s://%s/torrent/%s.torrent' % (u + (btih.upper(),))
|
|
for u in (('s', 'itorrents.org'), ('s', 'torrage.info'), ('', 'reflektor.karmorra.info'),
|
|
('', 'thetorrent.org'))]
|
|
except (StandardError, Exception):
|
|
link_type = 'torrent'
|
|
urls = [result.url]
|
|
|
|
elif GenericProvider.NZB == self.providerType:
|
|
final_dir = sickbeard.NZB_DIR
|
|
link_type = 'nzb'
|
|
urls = [result.url]
|
|
|
|
else:
|
|
return
|
|
|
|
ref_state = 'Referer' in self.session.headers and self.session.headers['Referer']
|
|
saved = False
|
|
for url in urls:
|
|
cache_dir = sickbeard.CACHE_DIR or helpers._getTempDir()
|
|
base_name = '%s.%s' % (helpers.sanitizeFileName(result.name), self.providerType)
|
|
final_file = ek.ek(os.path.join, final_dir, base_name)
|
|
cached = getattr(result, 'cache_file', None)
|
|
if cached and ek.ek(os.path.isfile, cached):
|
|
base_name = ek.ek(os.path.basename, cached)
|
|
cache_file = ek.ek(os.path.join, cache_dir, base_name)
|
|
|
|
self.session.headers['Referer'] = url
|
|
if cached or helpers.download_file(url, cache_file, session=self.session, allow_redirects='/it' not in url):
|
|
|
|
if self._verify_download(cache_file):
|
|
logger.log(u'Downloaded %s result from %s' % (self.name, url))
|
|
try:
|
|
helpers.moveFile(cache_file, final_file)
|
|
msg = 'moved'
|
|
except (OSError, Exception):
|
|
msg = 'copied cached file'
|
|
logger.log(u'Saved .%s data and %s to %s' % (
|
|
(link_type, 'torrent cache')['magnet' == link_type], msg, final_file))
|
|
saved = True
|
|
break
|
|
|
|
remove_file_failed(cache_file)
|
|
|
|
if 'Referer' in self.session.headers:
|
|
if ref_state:
|
|
self.session.headers['Referer'] = ref_state
|
|
else:
|
|
del(self.session.headers['Referer'])
|
|
|
|
if not saved and 'magnet' == link_type:
|
|
logger.log(u'All torrent cache servers failed to return a downloadable result', logger.DEBUG)
|
|
final_file = ek.ek(os.path.join, final_dir, '%s.%s' % (helpers.sanitizeFileName(result.name), link_type))
|
|
try:
|
|
with open(final_file, 'wb') as fp:
|
|
fp.write(result.url)
|
|
fp.flush()
|
|
os.fsync(fp.fileno())
|
|
saved = True
|
|
logger.log(u'Saved magnet link to file as some clients (or plugins) support this, %s' % final_file)
|
|
if 'blackhole' == sickbeard.TORRENT_METHOD:
|
|
logger.log('Tip: If your client fails to load magnet in files, ' +
|
|
'change blackhole to a client connection method in search settings')
|
|
except (StandardError, Exception):
|
|
logger.log(u'Failed to save magnet link to file, %s' % final_file)
|
|
elif not saved:
|
|
logger.log(u'Server failed to return anything useful', logger.ERROR)
|
|
|
|
return saved
|
|
|
|
def _verify_download(self, file_name=None):
|
|
"""
|
|
Checks the saved file to see if it was actually valid, if not then consider the download a failure.
|
|
"""
|
|
result = True
|
|
# primitive verification of torrents, just make sure we didn't get a text file or something
|
|
if GenericProvider.TORRENT == self.providerType:
|
|
parser = stream = None
|
|
try:
|
|
stream = FileInputStream(file_name)
|
|
parser = guessParser(stream)
|
|
except (StandardError, Exception):
|
|
pass
|
|
result = parser and 'application/x-bittorrent' == parser.mime_type
|
|
|
|
try:
|
|
stream._input.close()
|
|
except (StandardError, Exception):
|
|
pass
|
|
|
|
return result
|
|
|
|
def search_rss(self, episodes):
|
|
return self.cache.findNeededEpisodes(episodes)
|
|
|
|
def get_quality(self, item, anime=False):
|
|
"""
|
|
Figures out the quality of the given RSS item node
|
|
|
|
item: An elementtree.ElementTree element representing the <item> tag of the RSS feed
|
|
|
|
Returns a Quality value obtained from the node's data
|
|
"""
|
|
(title, url) = self._title_and_url(item) # @UnusedVariable
|
|
quality = Quality.sceneQuality(title, anime)
|
|
return quality
|
|
|
|
def _search_provider(self, search_params, search_mode='eponly', epcount=0, age=0, **kwargs):
|
|
return []
|
|
|
|
def _season_strings(self, episode):
|
|
return []
|
|
|
|
def _episode_strings(self, *args, **kwargs):
|
|
return []
|
|
|
|
def _title_and_url(self, item):
|
|
"""
|
|
Retrieves the title and URL data from the item
|
|
|
|
item: An elementtree.ElementTree element representing the <item> tag of the RSS feed, or a two part tup
|
|
|
|
Returns: A tuple containing two strings representing title and URL respectively
|
|
"""
|
|
|
|
title, url = None, None
|
|
try:
|
|
title, url = isinstance(item, tuple) and (item[0], item[1]) or \
|
|
(item.get('title', None), item.get('link', None))
|
|
except (StandardError, Exception):
|
|
pass
|
|
|
|
title = title and re.sub(r'\s+', '.', u'%s' % title)
|
|
url = url and str(url).replace('&', '&')
|
|
|
|
return title, url
|
|
|
|
def _link(self, url, url_tmpl=None):
|
|
|
|
url = url and str(url).strip().replace('&', '&') or ''
|
|
return url if re.match('(?i)(https?://|magnet:)', url) \
|
|
else (url_tmpl or self.urls.get('get', (getattr(self, 'url', '') or
|
|
getattr(self, 'url_base')) + '%s')) % url.lstrip('/')
|
|
|
|
@staticmethod
|
|
def _header_row(table_row, custom_match=None, custom_tags=None, header_strip=''):
|
|
"""
|
|
:param header_row: Soup resultset of table header row
|
|
:param custom_match: Dict key/values to override one or more default regexes
|
|
:param custom_tags: List of tuples with tag and attribute
|
|
:param header_strip: String regex of ambiguities to remove from headers
|
|
:return: dict column indices or None for leech, seeds, and size
|
|
"""
|
|
results = {}
|
|
rc = dict((k, re.compile('(?i)' + r)) for (k, r) in dict(
|
|
{'seed': r'(?:seed|s/l)', 'leech': r'(?:leech|peers)', 'size': r'(?:size)'}.items()
|
|
+ ({}, custom_match)[any([custom_match])].items()).items())
|
|
table = table_row.find_parent('table')
|
|
header_row = table.tr or table.thead.tr or table.tbody.tr
|
|
for y in [x for x in header_row(True) if x.attrs.get('class')]:
|
|
y['class'] = '..'.join(y['class'])
|
|
all_cells = header_row.find_all('th')
|
|
all_cells = all_cells if any(all_cells) else header_row.find_all('td')
|
|
|
|
headers = [re.sub(
|
|
r'[\s]+', '',
|
|
((any([cell.get_text()]) and any([rc[x].search(cell.get_text()) for x in rc.keys()]) and cell.get_text())
|
|
or (cell.attrs.get('id') and any([rc[x].search(cell['id']) for x in rc.keys()]) and cell['id'])
|
|
or (cell.attrs.get('title') and any([rc[x].search(cell['title']) for x in rc.keys()]) and cell['title'])
|
|
or next(iter(set(filter(lambda z: any([z]), [
|
|
next(iter(set(filter(lambda y: any([y]), [
|
|
cell.find(tag, **p) for p in [{attr: rc[x]} for x in rc.keys()]]))), {}).get(attr)
|
|
for (tag, attr) in [
|
|
('img', 'title'), ('img', 'src'), ('i', 'title'), ('i', 'class'),
|
|
('abbr', 'title'), ('a', 'title'), ('a', 'href')] + (custom_tags or [])]))), '')
|
|
or cell.get_text()
|
|
)).strip() for cell in all_cells]
|
|
headers = [re.sub(header_strip, '', x) for x in headers]
|
|
all_headers = headers
|
|
colspans = [int(cell.attrs.get('colspan', 0)) for cell in all_cells]
|
|
if any(colspans):
|
|
all_headers = []
|
|
for i, width in enumerate(colspans):
|
|
all_headers += [headers[i]] + ([''] * (width - 1))
|
|
|
|
for k, r in rc.iteritems():
|
|
if k not in results:
|
|
for name in filter(lambda v: any([v]) and r.search(v), all_headers[::-1]):
|
|
results[k] = all_headers.index(name) - len(all_headers)
|
|
break
|
|
|
|
for missing in set(rc.keys()) - set(results.keys()):
|
|
results[missing] = None
|
|
|
|
return results
|
|
|
|
@staticmethod
|
|
def _dhtless_magnet(btih, name=None):
|
|
"""
|
|
:param btih: torrent hash
|
|
:param name: torrent name
|
|
:return: a magnet loaded with default trackers for clients without enabled DHT or None if bad hash
|
|
"""
|
|
try:
|
|
btih = btih.lstrip('/').upper()
|
|
if 32 == len(btih):
|
|
btih = b16encode(b32decode(btih)).lower()
|
|
btih = re.search('(?i)[0-9a-f]{32,40}', btih) and btih or None
|
|
except (StandardError, Exception):
|
|
btih = None
|
|
return (btih and 'magnet:?xt=urn:btih:%s&dn=%s&tr=%s' % (btih, quote_plus(name or btih), '&tr='.join(
|
|
[quote_plus(tr) for tr in
|
|
'http://atrack.pow7.com/announce', 'http://mgtracker.org:2710/announce',
|
|
'http://pow7.com/announce', 'http://t1.pow7.com/announce',
|
|
'http://tracker.tfile.me/announce', 'udp://9.rarbg.com:2710/announce',
|
|
'udp://9.rarbg.me:2710/announce', 'udp://9.rarbg.to:2710/announce',
|
|
'udp://eddie4.nl:6969/announce', 'udp://explodie.org:6969/announce',
|
|
'udp://inferno.demonoid.pw:3395/announce', 'udp://inferno.subdemon.com:3395/announce',
|
|
'udp://ipv4.tracker.harry.lu:80/announce', 'udp://p4p.arenabg.ch:1337/announce',
|
|
'udp://shadowshq.yi.org:6969/announce', 'udp://tracker.aletorrenty.pl:2710/announce',
|
|
'udp://tracker.coppersurfer.tk:6969', 'udp://tracker.coppersurfer.tk:6969/announce',
|
|
'udp://tracker.internetwarriors.net:1337', 'udp://tracker.internetwarriors.net:1337/announce',
|
|
'udp://tracker.leechers-paradise.org:6969', 'udp://tracker.leechers-paradise.org:6969/announce',
|
|
'udp://tracker.opentrackr.org:1337/announce', 'udp://tracker.torrent.eu.org:451/announce',
|
|
'udp://tracker.trackerfix.com:80/announce', 'udp://tracker.zer0day.to:1337/announce'])) or None)
|
|
|
|
def get_show(self, item, **kwargs):
|
|
return None
|
|
|
|
def get_size_uid(self, item, **kwargs):
|
|
return -1, None
|
|
|
|
def find_search_results(self, show, episodes, search_mode, manual_search=False, **kwargs):
|
|
|
|
self._check_auth()
|
|
self.show = show
|
|
|
|
results = {}
|
|
item_list = []
|
|
if self.should_skip():
|
|
return results
|
|
|
|
searched_scene_season = None
|
|
search_list = []
|
|
for ep_obj in episodes:
|
|
# search cache for episode result
|
|
cache_result = self.cache.searchCache(ep_obj, manual_search)
|
|
if cache_result:
|
|
if ep_obj.episode not in results:
|
|
results[ep_obj.episode] = cache_result
|
|
else:
|
|
results[ep_obj.episode].extend(cache_result)
|
|
|
|
# found result, search next episode
|
|
continue
|
|
|
|
if 'sponly' == search_mode:
|
|
# skip if season already searched
|
|
if 1 < len(episodes) and searched_scene_season == ep_obj.scene_season:
|
|
continue
|
|
|
|
searched_scene_season = ep_obj.scene_season
|
|
|
|
# get season search params
|
|
search_params = self._season_strings(ep_obj)
|
|
else:
|
|
# get single episode search params
|
|
search_params = self._episode_strings(ep_obj)
|
|
|
|
search_list += [search_params]
|
|
|
|
search_done = []
|
|
for search_params in search_list:
|
|
if self.should_skip(log_warning=False):
|
|
break
|
|
for cur_param in search_params:
|
|
if cur_param in search_done:
|
|
continue
|
|
search_done += [cur_param]
|
|
item_list += self._search_provider(cur_param, search_mode=search_mode, epcount=len(episodes))
|
|
if self.should_skip():
|
|
break
|
|
|
|
return self.finish_find_search_results(show, episodes, search_mode, manual_search, results, item_list)
|
|
|
|
def finish_find_search_results(self, show, episodes, search_mode, manual_search, results, item_list, **kwargs):
|
|
|
|
# if we found what we needed already from cache then return results and exit
|
|
if len(results) == len(episodes):
|
|
return results
|
|
|
|
# sort list by quality
|
|
if len(item_list):
|
|
items = {}
|
|
items_unknown = []
|
|
for item in item_list:
|
|
quality = self.get_quality(item, anime=show.is_anime)
|
|
if Quality.UNKNOWN == quality:
|
|
items_unknown += [item]
|
|
else:
|
|
if quality not in items:
|
|
items[quality] = [item]
|
|
else:
|
|
items[quality].append(item)
|
|
|
|
item_list = list(itertools.chain(*[v for (k, v) in sorted(items.items(), reverse=True)]))
|
|
item_list += items_unknown if items_unknown else []
|
|
|
|
# filter results
|
|
cl = []
|
|
for item in item_list:
|
|
(title, url) = self._title_and_url(item)
|
|
|
|
parser = NameParser(False, showObj=self.get_show(item, **kwargs), convert=True, indexer_lookup=False)
|
|
# parse the file name
|
|
try:
|
|
parse_result = parser.parse(title)
|
|
except InvalidNameException:
|
|
logger.log(u'Unable to parse the filename %s into a valid episode' % title, logger.DEBUG)
|
|
continue
|
|
except InvalidShowException:
|
|
logger.log(u'No match for search criteria in the parsed filename ' + title, logger.DEBUG)
|
|
continue
|
|
|
|
show_obj = parse_result.show
|
|
quality = parse_result.quality
|
|
release_group = parse_result.release_group
|
|
version = parse_result.version
|
|
|
|
add_cache_entry = False
|
|
if not (show_obj.air_by_date or show_obj.is_sports):
|
|
if 'sponly' == search_mode:
|
|
if len(parse_result.episode_numbers):
|
|
logger.log(u'This is supposed to be a season pack search but the result ' + title +
|
|
u' is not a valid season pack, skipping it', logger.DEBUG)
|
|
add_cache_entry = True
|
|
if len(parse_result.episode_numbers)\
|
|
and (parse_result.season_number not in set([ep.season for ep in episodes]) or not [
|
|
ep for ep in episodes if ep.scene_episode in parse_result.episode_numbers]):
|
|
logger.log(u'The result ' + title + u' doesn\'t seem to be a valid episode that we are trying' +
|
|
u' to snatch, ignoring', logger.DEBUG)
|
|
add_cache_entry = True
|
|
else:
|
|
if not len(parse_result.episode_numbers)\
|
|
and parse_result.season_number\
|
|
and not [ep for ep in episodes
|
|
if ep.season == parse_result.season_number and
|
|
ep.episode in parse_result.episode_numbers]:
|
|
logger.log(u'The result ' + title + u' doesn\'t seem to be a valid season that we are trying' +
|
|
u' to snatch, ignoring', logger.DEBUG)
|
|
add_cache_entry = True
|
|
elif len(parse_result.episode_numbers) and not [
|
|
ep for ep in episodes if ep.season == parse_result.season_number and
|
|
ep.episode in parse_result.episode_numbers]:
|
|
logger.log(u'The result ' + title + ' doesn\'t seem to be a valid episode that we are trying' +
|
|
u' to snatch, ignoring', logger.DEBUG)
|
|
add_cache_entry = True
|
|
|
|
if not add_cache_entry:
|
|
# we just use the existing info for normal searches
|
|
actual_season = parse_result.season_number
|
|
actual_episodes = parse_result.episode_numbers
|
|
else:
|
|
if not parse_result.is_air_by_date:
|
|
logger.log(u'This is supposed to be a date search but the result ' + title +
|
|
u' didn\'t parse as one, skipping it', logger.DEBUG)
|
|
add_cache_entry = True
|
|
else:
|
|
actual_season = parse_result.season_number
|
|
actual_episodes = parse_result.episode_numbers
|
|
|
|
if not actual_episodes or \
|
|
not [ep for ep in episodes if ep.season == actual_season and ep.episode in actual_episodes]:
|
|
logger.log(u'The result ' + title + ' doesn\'t seem to be a valid episode that we are trying' +
|
|
u' to snatch, ignoring', logger.DEBUG)
|
|
add_cache_entry = True
|
|
|
|
# add parsed result to cache for usage later on
|
|
if add_cache_entry:
|
|
logger.log(u'Adding item from search to cache: ' + title, logger.DEBUG)
|
|
ci = self.cache.add_cache_entry(title, url, parse_result=parse_result)
|
|
if None is not ci:
|
|
cl.append(ci)
|
|
continue
|
|
|
|
# make sure we want the episode
|
|
want_ep = True
|
|
multi_ep = False
|
|
for epNo in actual_episodes:
|
|
want_ep = show_obj.wantEpisode(actual_season, epNo, quality, manual_search, multi_ep)
|
|
if not want_ep:
|
|
break
|
|
# after initial single ep perspective, prepare multi ep for subsequent iterations
|
|
multi_ep = 1 < actual_episodes
|
|
|
|
if not want_ep:
|
|
logger.log(u'Ignoring result %s because we don\'t want an episode that is %s'
|
|
% (title, Quality.qualityStrings[quality]), logger.DEBUG)
|
|
continue
|
|
|
|
logger.log(u'Found result %s at %s' % (title, url), logger.DEBUG)
|
|
|
|
# make a result object
|
|
ep_obj = []
|
|
for curEp in actual_episodes:
|
|
ep_obj.append(show_obj.getEpisode(actual_season, curEp))
|
|
|
|
result = self.get_result(ep_obj, url)
|
|
if None is result:
|
|
continue
|
|
result.show = show_obj
|
|
result.name = title
|
|
result.quality = quality
|
|
result.release_group = release_group
|
|
result.content = None
|
|
result.version = version
|
|
result.size, result.puid = self.get_size_uid(item, **kwargs)
|
|
result.is_repack, result.properlevel = Quality.get_proper_level(parse_result.extra_info_no_name(),
|
|
parse_result.version, show_obj.is_anime,
|
|
check_is_repack=True)
|
|
|
|
if 1 == len(ep_obj):
|
|
ep_num = ep_obj[0].episode
|
|
logger.log(u'Single episode result.', logger.DEBUG)
|
|
elif 1 < len(ep_obj):
|
|
ep_num = MULTI_EP_RESULT
|
|
logger.log(u'Separating multi-episode result to check for later - result contains episodes: ' +
|
|
str(parse_result.episode_numbers), logger.DEBUG)
|
|
elif 0 == len(ep_obj):
|
|
ep_num = SEASON_RESULT
|
|
logger.log(u'Separating full season result to check for later', logger.DEBUG)
|
|
|
|
if ep_num not in results:
|
|
results[ep_num] = [result]
|
|
else:
|
|
results[ep_num].append(result)
|
|
|
|
# check if we have items to add to cache
|
|
if 0 < len(cl):
|
|
my_db = self.cache.get_db()
|
|
my_db.mass_action(cl)
|
|
|
|
return results
|
|
|
|
def find_propers(self, search_date=None, **kwargs):
|
|
|
|
results = self.cache.listPropers(search_date)
|
|
|
|
return [classes.Proper(x['name'], x['url'], datetime.datetime.fromtimestamp(x['time']), self.show) for x in
|
|
results]
|
|
|
|
def seed_ratio(self):
|
|
"""
|
|
Provider should override this value if custom seed ratio enabled
|
|
It should return the value of the provider seed ratio
|
|
"""
|
|
return ''
|
|
|
|
def _log_search(self, mode='Cache', count=0, url='url missing'):
|
|
"""
|
|
Simple function to log the result of a search types except propers
|
|
:param count: count of successfully processed items
|
|
:param url: source url of item(s)
|
|
"""
|
|
if 'Propers' != mode:
|
|
self.log_result(mode, count, url)
|
|
|
|
def log_result(self, mode='Cache', count=0, url='url missing'):
|
|
"""
|
|
Simple function to log the result of any search
|
|
:param mode: string that this log relates to
|
|
:param count: count of successfully processed items
|
|
:param url: source url of item(s)
|
|
"""
|
|
if not self.should_skip():
|
|
str1, thing, str3 = (('', '%s item' % mode.lower(), ''), (' usable', 'proper', ' found'))['Propers' == mode]
|
|
logger.log(u'%s %s in response from %s' % (('No' + str1, count)[0 < count], (
|
|
'%s%s%s%s' % (('', 'freeleech ')[getattr(self, 'freeleech', False)], thing, maybe_plural(count), str3)),
|
|
re.sub('(\s)\s+', r'\1', url)))
|
|
|
|
def check_auth_cookie(self):
|
|
|
|
if hasattr(self, 'cookies'):
|
|
cookies = self.cookies
|
|
|
|
if not (cookies and re.match('^(?:\w+=[^;\s]+[;\s]*)+$', cookies)):
|
|
return False
|
|
|
|
cj = requests.utils.add_dict_to_cookiejar(self.session.cookies,
|
|
dict([x.strip().split('=') for x in cookies.split(';')
|
|
if x != ''])),
|
|
for item in cj:
|
|
if not isinstance(item, requests.cookies.RequestsCookieJar):
|
|
return False
|
|
|
|
return True
|
|
|
|
def _check_cookie(self):
|
|
|
|
if self.check_auth_cookie():
|
|
return True, None
|
|
|
|
return False, 'Cookies not correctly formatted key=value pairs e.g. uid=xx;pass=yy)'
|
|
|
|
def has_all_cookies(self, cookies=None, pre=''):
|
|
|
|
cookies = cookies and ([cookies], cookies)[isinstance(cookies, list)] or ['uid', 'pass']
|
|
return all(['%s%s' % (pre, item) in self.session.cookies for item in cookies])
|
|
|
|
def _categories_string(self, mode='Cache', template='c%s=1', delimiter='&'):
|
|
|
|
return delimiter.join([('%s', template)[any(template)] % c for c in sorted(
|
|
'shows' in self.categories and (isinstance(self.categories['shows'], type([])) and
|
|
self.categories['shows'] or [self.categories['shows']]) or
|
|
self.categories[(mode, 'Episode')['Propers' == mode]] +
|
|
([], self.categories.get('anime') or [])[
|
|
(mode in ['Cache', 'Propers'] and helpers.has_anime()) or
|
|
((mode in ['Season', 'Episode']) and self.show and self.show.is_anime)])])
|
|
|
|
@staticmethod
|
|
def _bytesizer(size_dim=''):
|
|
|
|
try:
|
|
value = float('.'.join(re.findall('(?i)(\d+)(?:[.,](\d+))?', size_dim)[0]))
|
|
except TypeError:
|
|
return size_dim
|
|
except IndexError:
|
|
return None
|
|
try:
|
|
value *= 1024 ** ['b', 'k', 'm', 'g', 't'].index(re.findall('(t|g|m|k)[i]?b', size_dim.lower())[0])
|
|
except IndexError:
|
|
pass
|
|
return long(math.ceil(value))
|
|
|
|
@staticmethod
|
|
def _should_stop():
|
|
if getattr(threading.currentThread(), 'stop', False):
|
|
return True
|
|
return False
|
|
|
|
def _sleep_with_stop(self, t):
|
|
t_l = t
|
|
while t_l > 0:
|
|
time.sleep(3)
|
|
t_l -= 3
|
|
if self._should_stop():
|
|
return
|
|
|
|
|
|
class NZBProvider(GenericProvider):
|
|
|
|
def __init__(self, name, supports_backlog=True, anime_only=False):
|
|
GenericProvider.__init__(self, name, supports_backlog, anime_only)
|
|
|
|
self.providerType = GenericProvider.NZB
|
|
self.has_limit = True
|
|
|
|
def image_name(self):
|
|
|
|
return GenericProvider.image_name(self, 'newznab')
|
|
|
|
def maybe_apikey(self):
|
|
|
|
if getattr(self, 'needs_auth', None):
|
|
return (getattr(self, 'key', '') and self.key) or (getattr(self, 'api_key', '') and self.api_key) or None
|
|
return False
|
|
|
|
def _check_auth(self, is_required=None):
|
|
|
|
has_key = self.maybe_apikey()
|
|
if has_key:
|
|
return has_key
|
|
if None is has_key:
|
|
raise AuthException('%s for %s is empty in Media Providers/Options'
|
|
% ('API key' + ('', ' and/or Username')[hasattr(self, 'username')], self.name))
|
|
|
|
return GenericProvider._check_auth(self)
|
|
|
|
def find_propers(self, search_date=None, shows=None, anime=None, **kwargs):
|
|
|
|
cache_results = self.cache.listPropers(search_date)
|
|
results = [classes.Proper(x['name'], x['url'], datetime.datetime.fromtimestamp(x['time']), self.show) for x in
|
|
cache_results]
|
|
|
|
if self.should_skip():
|
|
return results
|
|
|
|
index = 0
|
|
alt_search = ('nzbs_org' == self.get_id())
|
|
do_search_alt = False
|
|
|
|
search_terms = []
|
|
regex = []
|
|
if shows:
|
|
search_terms += ['.proper.', '.repack.', '.real.']
|
|
regex += ['proper|repack', Quality.real_check]
|
|
proper_check = re.compile(r'(?i)(\b%s\b)' % '|'.join(regex))
|
|
if anime:
|
|
terms = 'v2|v3|v4|v5|v6|v7|v8|v9'
|
|
search_terms += [terms]
|
|
regex += [terms]
|
|
proper_check = re.compile(r'(?i)(%s)' % '|'.join(regex))
|
|
|
|
urls = []
|
|
while index < len(search_terms):
|
|
if self.should_skip(log_warning=False):
|
|
break
|
|
|
|
search_params = {'q': search_terms[index], 'maxage': sickbeard.BACKLOG_DAYS + 2}
|
|
if alt_search:
|
|
|
|
if do_search_alt:
|
|
search_params['t'] = 'search'
|
|
index += 1
|
|
|
|
do_search_alt = not do_search_alt
|
|
|
|
else:
|
|
index += 1
|
|
|
|
for item in self._search_provider({'Propers': [search_params]}):
|
|
|
|
(title, url) = self._title_and_url(item)
|
|
|
|
if not proper_check.search(title) or url in urls:
|
|
continue
|
|
urls.append(url)
|
|
|
|
if 'published_parsed' in item and item['published_parsed']:
|
|
result_date = item.published_parsed
|
|
if result_date:
|
|
result_date = datetime.datetime(*result_date[0:6])
|
|
else:
|
|
logger.log(u'Unable to figure out the date for entry %s, skipping it' % title)
|
|
continue
|
|
|
|
if not search_date or search_date < result_date:
|
|
search_result = classes.Proper(title, url, result_date, self.show)
|
|
results.append(search_result)
|
|
|
|
time.sleep(0.5)
|
|
|
|
return results
|
|
|
|
def cache_data(self, *args, **kwargs):
|
|
|
|
search_params = {'Cache': [{}]}
|
|
return self._search_provider(search_params=search_params, **kwargs)
|
|
|
|
|
|
class TorrentProvider(GenericProvider):
|
|
|
|
def __init__(self, name, supports_backlog=True, anime_only=False, cache_update_freq=None, update_freq=None):
|
|
GenericProvider.__init__(self, name, supports_backlog, anime_only)
|
|
|
|
self.providerType = GenericProvider.TORRENT
|
|
|
|
self._seed_ratio = None
|
|
self.seed_time = None
|
|
self._url = None
|
|
self.urls = {}
|
|
self.cache._cache_data = self._cache_data
|
|
if cache_update_freq:
|
|
self.cache.update_freq = cache_update_freq
|
|
self.ping_freq = update_freq
|
|
self.ping_skip = None
|
|
|
|
@property
|
|
def url(self):
|
|
if None is self._url or (hasattr(self, 'url_tmpl') and not self.urls):
|
|
self._url = self._valid_home(False)
|
|
self._valid_url()
|
|
return self._url
|
|
|
|
@url.setter
|
|
def url(self, value=None):
|
|
self._url = value
|
|
|
|
def _valid_url(self):
|
|
return True
|
|
|
|
def image_name(self):
|
|
|
|
return GenericProvider.image_name(self, 'torrent')
|
|
|
|
def seed_ratio(self):
|
|
|
|
return self._seed_ratio
|
|
|
|
@staticmethod
|
|
def _sort_seeders(mode, items):
|
|
""" legacy function used by a custom provider, do not remove """
|
|
mode in ['Season', 'Episode'] and items[mode].sort(key=lambda tup: tup[2], reverse=True)
|
|
|
|
@staticmethod
|
|
def _sort_seeding(mode, items):
|
|
|
|
if mode in ['Season', 'Episode']:
|
|
return sorted(set(items), key=lambda tup: tup[2], reverse=True)
|
|
return items
|
|
|
|
def _peers_fail(self, mode, seeders=0, leechers=0):
|
|
|
|
return 'Cache' != mode and (seeders < getattr(self, 'minseed', 0) or leechers < getattr(self, 'minleech', 0))
|
|
|
|
def get_quality(self, item, anime=False):
|
|
|
|
if isinstance(item, tuple):
|
|
name = item[0]
|
|
elif isinstance(item, dict):
|
|
name, url = self._title_and_url(item)
|
|
else:
|
|
name = item.title
|
|
return Quality.sceneQuality(name, anime)
|
|
|
|
@staticmethod
|
|
def _reverse_quality(quality):
|
|
|
|
return {
|
|
Quality.SDTV: 'HDTV x264',
|
|
Quality.SDDVD: 'DVDRIP',
|
|
Quality.HDTV: '720p HDTV x264',
|
|
Quality.FULLHDTV: '1080p HDTV x264',
|
|
Quality.RAWHDTV: '1080i HDTV mpeg2',
|
|
Quality.HDWEBDL: '720p WEB-DL h264',
|
|
Quality.FULLHDWEBDL: '1080p WEB-DL h264',
|
|
Quality.HDBLURAY: '720p Bluray x264',
|
|
Quality.FULLHDBLURAY: '1080p Bluray x264'
|
|
}.get(quality, '')
|
|
|
|
def _season_strings(self, ep_obj, detail_only=False, scene=True, prefix='', **kwargs):
|
|
|
|
if not ep_obj:
|
|
return []
|
|
|
|
show = ep_obj.show
|
|
ep_dict = self._ep_dict(ep_obj)
|
|
sp_detail = (show.air_by_date or show.is_sports) and str(ep_obj.airdate).split('-')[0] or \
|
|
(show.is_anime and ep_obj.scene_absolute_number or
|
|
('sp_detail' in kwargs.keys() and kwargs['sp_detail'](ep_dict)) or 'S%(seasonnumber)02d' % ep_dict)
|
|
sp_detail = ([sp_detail], sp_detail)[isinstance(sp_detail, list)]
|
|
detail = ({}, {'Season_only': sp_detail})[detail_only and not self.show.is_sports and not self.show.is_anime]
|
|
return [dict({'Season': self._build_search_strings(sp_detail, scene, prefix)}.items() + detail.items())]
|
|
|
|
def _episode_strings(self, ep_obj, detail_only=False, scene=True, prefix='', sep_date=' ', date_or=False, **kwargs):
|
|
|
|
if not ep_obj:
|
|
return []
|
|
|
|
show = ep_obj.show
|
|
if show.air_by_date or show.is_sports:
|
|
ep_detail = [str(ep_obj.airdate).replace('-', sep_date)]\
|
|
if 'date_detail' not in kwargs.keys() else kwargs['date_detail'](ep_obj.airdate)
|
|
if show.is_sports:
|
|
month = ep_obj.airdate.strftime('%b')
|
|
ep_detail = (ep_detail + [month], ['%s|%s' % (x, month) for x in ep_detail])[date_or]
|
|
elif show.is_anime:
|
|
ep_detail = ep_obj.scene_absolute_number \
|
|
if 'ep_detail_anime' not in kwargs.keys() else kwargs['ep_detail_anime'](ep_obj.scene_absolute_number)
|
|
else:
|
|
ep_dict = self._ep_dict(ep_obj)
|
|
ep_detail = sickbeard.config.naming_ep_type[2] % ep_dict \
|
|
if 'ep_detail' not in kwargs.keys() else kwargs['ep_detail'](ep_dict)
|
|
if sickbeard.scene_exceptions.has_abs_episodes(ep_obj):
|
|
ep_detail = ([ep_detail], ep_detail)[isinstance(ep_detail, list)] + ['%d' % ep_dict['episodenumber']]
|
|
ep_detail = ([ep_detail], ep_detail)[isinstance(ep_detail, list)]
|
|
detail = ({}, {'Episode_only': ep_detail})[detail_only and not show.is_sports and not show.is_anime]
|
|
return [dict({'Episode': self._build_search_strings(ep_detail, scene, prefix)}.items() + detail.items())]
|
|
|
|
@staticmethod
|
|
def _ep_dict(ep_obj):
|
|
season, episode = ((ep_obj.season, ep_obj.episode),
|
|
(ep_obj.scene_season, ep_obj.scene_episode))[bool(ep_obj.show.is_scene)]
|
|
return {'seasonnumber': season, 'episodenumber': episode}
|
|
|
|
def _build_search_strings(self, ep_detail, process_name=True, prefix=''):
|
|
"""
|
|
Build a list of search strings for querying a provider
|
|
:param ep_detail: String of episode detail or List of episode details
|
|
:param process_name: Bool Whether to call sanitizeSceneName() on show name
|
|
:param prefix: String to insert to search strings
|
|
:return: List of search string parameters
|
|
"""
|
|
ep_detail = ([ep_detail], ep_detail)[isinstance(ep_detail, list)]
|
|
prefix = ([prefix], prefix)[isinstance(prefix, list)]
|
|
|
|
search_params = []
|
|
crop = re.compile(r'([.\s])(?:\1)+')
|
|
for name in get_show_names_all_possible(self.show, scenify=process_name and getattr(self, 'scene', True)):
|
|
for detail in ep_detail:
|
|
search_params += [crop.sub(r'\1', '%s %s%s' % (name, x, detail)) for x in prefix]
|
|
return search_params
|
|
|
|
@staticmethod
|
|
def _has_signature(data=None):
|
|
return data and re.search(r'(?sim)<input[^<]+?name=["\'\s]*?password', data) and \
|
|
re.search(r'(?sim)<input[^<]+?name=["\'\s]*?username', data)
|
|
|
|
def _valid_home(self, attempt_fetch=True):
|
|
"""
|
|
:return: signature verified home url else None if validation fail
|
|
"""
|
|
url_base = getattr(self, 'url_base', None)
|
|
if url_base:
|
|
return url_base
|
|
|
|
url_list = getattr(self, 'url_home', None)
|
|
if not url_list and getattr(self, 'url_edit', None) or 10 > max([len(x) for x in url_list]):
|
|
return None
|
|
|
|
url_list = ['%s/' % x.rstrip('/') for x in url_list]
|
|
last_url, expire = sickbeard.PROVIDER_HOMES.get(self.get_id(), ('', None))
|
|
url_drop = getattr(self, 'url_drop', [])
|
|
if url_drop and any([url in last_url for url in url_drop]): # deprecate url
|
|
last_url = ''
|
|
|
|
if 'site down' == last_url:
|
|
if expire and (expire > int(time.time())) or not self.enabled:
|
|
return None
|
|
elif last_url:
|
|
last_url = last_url.replace('getrss.php', '/') # correct develop typo after a network outage (0.11>0.12)
|
|
last_url in url_list and url_list.remove(last_url)
|
|
url_list.insert(0, last_url)
|
|
|
|
if not self.enabled:
|
|
return last_url
|
|
|
|
self.failure_count = failure_count = 0
|
|
for cur_url in url_list:
|
|
if not self.is_valid_mod(cur_url):
|
|
return None
|
|
failure_count += self.failure_count
|
|
self.failure_count = 0
|
|
cur_url = cur_url.replace('{ts}', '%s.' % str(time.time())[2:6])
|
|
if 10 < len(cur_url) and ((expire and (expire > int(time.time()))) or
|
|
self._has_signature(self.get_url(cur_url, skip_auth=True))):
|
|
for k, v in getattr(self, 'url_tmpl', {}).items():
|
|
self.urls[k] = v % {'home': cur_url, 'vars': getattr(self, 'url_vars', {}).get(k, '')}
|
|
|
|
if last_url != cur_url or (expire and not (expire > int(time.time()))):
|
|
sickbeard.PROVIDER_HOMES[self.get_id()] = (cur_url, int(time.time()) + (60*60))
|
|
sickbeard.save_config()
|
|
return cur_url
|
|
|
|
self.failure_count = 3 * bool(failure_count)
|
|
if self.should_skip():
|
|
return None
|
|
|
|
logger.log('Failed to identify a "%s" page with %s %s (local network issue, site down, or ISP blocked) ' %
|
|
(self.name, len(url_list), ('URL', 'different URLs')[1 < len(url_list)]) +
|
|
(attempt_fetch and ('Suggest; 1) Disable "%s" 2) Use a proxy/VPN' % self.get_id()) or ''),
|
|
(logger.WARNING, logger.ERROR)[self.enabled])
|
|
self.urls = {}
|
|
sickbeard.PROVIDER_HOMES[self.get_id()] = ('site down', int(time.time()) + (5 * 60))
|
|
sickbeard.save_config()
|
|
return None
|
|
|
|
def is_valid_mod(self, url):
|
|
parsed, s, is_valid = urlparse.urlparse(url), 70000700, True
|
|
if 2012691328 == s + zlib.crc32(('.%s' % (parsed.netloc or parsed.path)).split('.')[-2]):
|
|
is_valid = False
|
|
file_name = '%s.py' % os.path.join(sickbeard.PROG_DIR, *self.__module__.split('.'))
|
|
if ek.ek(os.path.isfile, file_name):
|
|
with open(file_name, 'rb') as file_hd:
|
|
is_valid = s + zlib.crc32(file_hd.read()) in (1661931498, 472149389)
|
|
return is_valid
|
|
|
|
def _authorised(self, logged_in=None, post_params=None, failed_msg=None, url=None, timeout=30, **kwargs):
|
|
|
|
maxed_out = (lambda y: re.search(r'(?i)[1-3]((<[^>]+>)|\W)*' +
|
|
'(attempts|tries|remain)[\W\w]{,40}?(remain|left|attempt)', y))
|
|
logged_in, failed_msg = [None is not a and a or b for (a, b) in (
|
|
(logged_in, (lambda y=None: self.has_all_cookies())),
|
|
(failed_msg, (lambda y='': maxed_out(y) and u'Urgent abort, running low on login attempts. ' +
|
|
u'Password flushed to prevent service disruption to %s.' or
|
|
(re.search(r'(?i)(username|password)((<[^>]+>)|\W)*' +
|
|
'(or|and|/|\s)((<[^>]+>)|\W)*(password|incorrect)', y) and
|
|
u'Invalid username or password for %s. Check settings' or
|
|
u'Failed to authenticate or parse a response from %s, abort provider')))
|
|
)]
|
|
|
|
if logged_in() and (not hasattr(self, 'urls') or bool(len(getattr(self, 'urls')))):
|
|
return True
|
|
|
|
if not self._valid_home():
|
|
return False
|
|
|
|
if getattr(self, 'digest', None):
|
|
self.cookies = re.sub(r'(?i)([\s\']+|cookie\s*:)', '', self.digest)
|
|
success, msg = self._check_cookie()
|
|
if not success:
|
|
self.cookies = None
|
|
logger.log(u'%s: [%s]' % (msg, self.cookies), logger.WARNING)
|
|
return False
|
|
else:
|
|
try:
|
|
if not self._check_auth():
|
|
return False
|
|
except AuthException as e:
|
|
logger.log('%s' % ex(e), logger.ERROR)
|
|
return False
|
|
|
|
if isinstance(url, type([])):
|
|
for i in range(0, len(url)):
|
|
self.get_url(url.pop(), skip_auth=True, **kwargs)
|
|
if self.should_skip():
|
|
return False
|
|
|
|
passfield, userfield = None, None
|
|
post_params = isinstance(post_params, type({})) and post_params or {}
|
|
if not url:
|
|
if hasattr(self, 'urls'):
|
|
url = self.urls.get('login_action')
|
|
if url:
|
|
response = self.get_url(url, skip_auth=True, **kwargs)
|
|
if isinstance(response, tuple):
|
|
response = response[0]
|
|
if self.should_skip() or None is response:
|
|
return False
|
|
try:
|
|
form = 'form_tmpl' in post_params and post_params.pop('form_tmpl')
|
|
if form:
|
|
form = re.findall(
|
|
'(?is)(<form[^>]+%s.*?</form>)' % (True is form and 'login' or form), response)
|
|
response = form and form[0] or response
|
|
|
|
action = re.findall('<form[^>]+action=[\'"]([^\'"]*)', response)[0]
|
|
url = action if action.startswith('http') else \
|
|
url if not action else \
|
|
(url + action) if action.startswith('?') else \
|
|
(self.urls.get('login_base') or self.urls['config_provider_home_uri']) + action.lstrip('/')
|
|
|
|
tags = re.findall(r'(?is)(<input[^>]*?name=[\'"][^\'"]+[^>]*)', response)
|
|
attrs = [[(re.findall(r'(?is)%s=[\'"]([^\'"]+)' % attr, x) or [''])[0]
|
|
for attr in ['type', 'name', 'value']] for x in tags]
|
|
for itype, name, value in attrs:
|
|
if 'password' in [itype, name]:
|
|
passfield = name
|
|
if name not in ('username', 'password') and 'password' != itype:
|
|
post_params.setdefault(name, value)
|
|
except IndexError:
|
|
return False
|
|
except KeyError:
|
|
return super(TorrentProvider, self)._authorised()
|
|
else:
|
|
url = self.urls.get('login')
|
|
if not url:
|
|
return super(TorrentProvider, self)._authorised()
|
|
|
|
if getattr(self, 'username', None) and getattr(self, 'password', None) and post_params.pop('login', True):
|
|
if not post_params:
|
|
post_params = dict(username=self.username, password=self.password)
|
|
elif isinstance(post_params, type({})):
|
|
if self.username not in post_params.values():
|
|
post_params['username'] = self.username
|
|
if self.password not in post_params.values():
|
|
post_params[(passfield, 'password')[not passfield]] = self.password
|
|
|
|
response = self.get_url(url, skip_auth=True, post_data=post_params, timeout=timeout, **kwargs)
|
|
session = True
|
|
if isinstance(response, tuple):
|
|
session = response[1]
|
|
response = response[0]
|
|
if not self.should_skip() and response:
|
|
if logged_in(response):
|
|
return session
|
|
|
|
if maxed_out(response) and hasattr(self, 'password'):
|
|
self.password = None
|
|
sickbeard.save_config()
|
|
msg = failed_msg(response)
|
|
if msg:
|
|
logger.log(msg % self.name, logger.ERROR)
|
|
|
|
return False
|
|
|
|
def _check_auth(self, is_required=False):
|
|
|
|
if hasattr(self, 'username') and hasattr(self, 'password'):
|
|
if self.username and self.password:
|
|
return True
|
|
setting = 'Password or Username'
|
|
elif hasattr(self, 'username') and hasattr(self, 'api_key'):
|
|
if self.username and self.api_key:
|
|
return True
|
|
setting = 'Api key or Username'
|
|
elif hasattr(self, 'username') and hasattr(self, 'passkey'):
|
|
if self.username and self.passkey:
|
|
return True
|
|
setting = 'Passkey or Username'
|
|
elif hasattr(self, 'uid') and hasattr(self, 'passkey'):
|
|
if self.uid and self.passkey:
|
|
return True
|
|
setting = 'Passkey or uid'
|
|
elif hasattr(self, 'api_key'):
|
|
if self.api_key:
|
|
return True
|
|
setting = 'Api key'
|
|
elif hasattr(self, 'passkey'):
|
|
if self.passkey:
|
|
return True
|
|
setting = 'Passkey'
|
|
else:
|
|
return not is_required and GenericProvider._check_auth(self)
|
|
|
|
raise AuthException('%s for %s is empty in Media Providers/Options' % (setting, self.name))
|
|
|
|
def find_propers(self, anime=False, **kwargs):
|
|
"""
|
|
Search for releases of type PROPER
|
|
:return: list of Proper objects
|
|
"""
|
|
results = []
|
|
if self.should_skip():
|
|
return results
|
|
|
|
# chance of a v6-v9 is so rare that to do every bl search with each in turn is too aggressive
|
|
search_terms = getattr(self, 'proper_search_terms', ['proper', 'repack', 'real'] +
|
|
([], ['v2', 'v3', 'v4', 'v5'])[True is anime])
|
|
if not isinstance(search_terms, list):
|
|
if None is search_terms:
|
|
search_terms = ['proper|repack|real']
|
|
if anime:
|
|
search_terms += ['v2|v3|v4|v5']
|
|
else:
|
|
search_terms = [search_terms]
|
|
|
|
items = self._search_provider({'Propers': search_terms})
|
|
|
|
clean_term = re.compile(r'(?i)[^a-z1-9|.]+')
|
|
for proper_term in search_terms:
|
|
if self.should_skip(log_warning=False):
|
|
break
|
|
|
|
proper_check = re.compile(r'(?i)(?:%s)' % clean_term.sub('', proper_term))
|
|
for item in items:
|
|
if self.should_skip(log_warning=False):
|
|
break
|
|
|
|
title, url = self._title_and_url(item)
|
|
if proper_check.search(title):
|
|
results.append(classes.Proper(title, url, datetime.datetime.today(),
|
|
helpers.findCertainShow(sickbeard.showList, None)))
|
|
return results
|
|
|
|
@staticmethod
|
|
def _has_no_results(html):
|
|
return re.search(r'(?i)<(?:b|div|h\d|p|span|strong|td)[^>]*>\s*(?:' +
|
|
'your\ssearch.*?did\snot\smatch|' +
|
|
'(?:nothing|0</b>\s+torrents)\sfound|' +
|
|
'(?:sorry,\s)?no\s(?:results|torrents)\s(found|here|match)|' +
|
|
'no\s(?:match|results|torrents)!*|'
|
|
'[^<]*?there\sare\sno\sresults|' +
|
|
'[^<]*?no\shits\.\sTry\sadding' +
|
|
')', html)
|
|
|
|
def _cache_data(self, **kwargs):
|
|
|
|
return self._search_provider({'Cache': ['']})
|
|
|
|
def _ping(self):
|
|
while not self._should_stop():
|
|
if self.ping_skip:
|
|
self.ping_skip -= 1
|
|
else:
|
|
self.ping_skip = ((60*60)/self.ping_freq, None)[self._authorised()]
|
|
|
|
self._sleep_with_stop(self.ping_freq)
|
|
|
|
def get_result(self, episodes, url):
|
|
result = None
|
|
|
|
if url:
|
|
result = super(TorrentProvider, self).get_result(episodes, url)
|
|
if hasattr(self, 'get_data'):
|
|
result.get_data_func = self.get_data
|
|
|
|
return result
|