From 983280b48bf7ad6685740d9543f9b7986a35d7f5 Mon Sep 17 00:00:00 2001 From: JackDandy Date: Thu, 16 Jul 2015 13:09:13 +0100 Subject: [PATCH] Add GTI torrent provider. --- CHANGES.md | 1 + gui/slick/images/providers/grabtheinfo.png | Bin 0 -> 1011 bytes sickbeard/__init__.py | 2 +- sickbeard/providers/__init__.py | 1 + sickbeard/providers/grabtheinfo.py | 160 +++++++++++++++++++++ 5 files changed, 163 insertions(+), 1 deletion(-) create mode 100644 gui/slick/images/providers/grabtheinfo.png create mode 100644 sickbeard/providers/grabtheinfo.py diff --git a/CHANGES.md b/CHANGES.md index 34468678..9102c5d4 100644 --- a/CHANGES.md +++ b/CHANGES.md @@ -63,6 +63,7 @@ * Add BeyondHD torrent provider * Add GFTracker torrent provider * Add TtN torrent provider +* Add GTI torrent provider * Fix getManualSearchStatus: object has no attribute 'segment' * Change handling of general HTTP error response codes to prevent issues * Add handling for CloudFlare custom HTTP response codes diff --git a/gui/slick/images/providers/grabtheinfo.png b/gui/slick/images/providers/grabtheinfo.png new file mode 100644 index 0000000000000000000000000000000000000000..4a11f06eddfa12bfd8705a7a2fd49077ccedbfe7 GIT binary patch literal 1011 zcmV3^ zX>Mj|ZER|7Zfb6AYHn_BZF6mNW^Qt9ZgF*Qb98oMWprg_d}e8WZf|pVZ*q5Yb9Z)g zcXx1eczknmd476&e|ma;eR_U=dwzg>eu8^_hkSu}e1U*`fP#O6fqsF6r2qiE004|* z2#jk8lyeB2a|oP!5T$zvt$PTAa(0P#eC7ZE{}~zoLPCjofTe^GrHl}*j1aAq5WS2D zwTuwGlnBn82-chk#ibd}tq|6w2-c+#)~yiMtr^~}5Z1LB-n9_sy%6TT8Ro?q@5LGK z&KdXCA@|-P|K1`0<{^QChJc2Ehl7BDgp7oVgoKEOgoufRh>nMgjEIefj*EhbjERVh zj){$qijIwljF63yjg66%j**m(l9!T+ft-(+kd%^$p`)g!rKhK(rmCc=rl+W;qo}B*si>u?s->u|s;j1{ ztg@@Eu&b@HtFN@HueGbLxT&zUuduMLv#_nOx399dt+BbRvbe6XxUaIfud}(Ww7jsX zsIaNBv9q~XW z&BMjg$H~*j%G1Tj)ymA)%g)t1bZ7hk008qzL_t&-(_>)30H>~8xqkb>&6{^GpFDhE zJ5b>CxwFSlU%heS$ljeB)~^E!9P2JE>D}L7S>$Au*DMYcDATdkQ0s7%l@OED3}OZf zs5|dyk+0p{C#l@p+0Q)r&`u5T%gYh7O*pGJos#` zZo}&N=G;JmdxthH-*f)j#?FcJrW8m61#aGac<0We`. + +import re +import datetime +import traceback + +from . import generic +from sickbeard import logger, tvcache, helpers +from sickbeard.bs4_parser import BS4Parser +from lib.unidecode import unidecode + + +class GrabTheInfoProvider(generic.TorrentProvider): + + def __init__(self): + generic.TorrentProvider.__init__(self, 'GrabTheInfo') + + self.url_base = 'http://grabthe.info/' + self.urls = {'config_provider_home_uri': self.url_base, + 'login': self.url_base + 'takelogin.php', + 'cache': self.url_base + 'browse.php?%s', + 'search': '&search=%s', + 'get': self.url_base + '%s'} + + self.categories = 'c56=1&c8=1&c61=1&c10=1&incldead=0&blah=0' + + self.url = self.urls['config_provider_home_uri'] + + self.username, self.password, self.minseed, self.minleech = 4 * [None] + self.cache = GrabTheInfoCache(self) + + def _do_login(self): + + logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies + if logged_in(): + return True + + if self._check_auth(): + login_params = {'username': self.username, 'password': self.password} + response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session) + if response and logged_in(): + return True + + msg = u'Failed to authenticate with %s, abort provider' + if response and 'Username or password incorrect' in response: + msg = u'Invalid username or password for %s. Check settings' + logger.log(msg % self.name, logger.ERROR) + + return False + + def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0): + + results = [] + if not self._do_login(): + return results + + items = {'Season': [], 'Episode': [], 'Cache': []} + + rc = dict((k, re.compile('(?i)' + v)) for (k, v) in {'info': 'detail', 'get': 'download'}.items()) + for mode in search_params.keys(): + for search_string in search_params[mode]: + + if isinstance(search_string, unicode): + search_string = unidecode(search_string) + + search_url = self.urls['cache'] % self.categories + if 'cache' != mode.lower(): + search_url += self.urls['search'] % search_string + html = self.get_url(search_url) + + cnt = len(items[mode]) + try: + if not html or self._has_no_results(html): + raise generic.HaltParseException + + html = html.replace('', '') + html = re.sub(r'()[^<]*', r'\1', html) + html = re.sub(r'( (len(torrent_rows) - shows_found): + raise generic.HaltParseException + + for tr in torrent_rows[1 + shows_found:]: + try: + info = tr.find('a', href=rc['info']) + if None is info: + continue + title = (('title' in info.attrs.keys() and info['title']) or info.get_text()).strip() + + download_url = tr.find('a', href=rc['get']) + if None is download_url: + continue + + seeders, leechers = [int(tr.find_all('td')[x].get_text().strip()) for x in (-2, -1)] + if 'Cache' != mode and (seeders < self.minseed or leechers < self.minleech): + continue + except (AttributeError, TypeError, KeyError): + continue + + if title: + items[mode].append((title, self.urls['get'] + % str(download_url['href'].lstrip('/')), seeders)) + + except generic.HaltParseException: + pass + except Exception: + logger.log(u'Failed to parse. Traceback: %s' % traceback.format_exc(), logger.ERROR) + self._log_result(mode, len(items[mode]) - cnt, search_url) + + # for each search mode sort all the items by seeders + 'Cache' != mode and items[mode].sort(key=lambda tup: tup[2], reverse=True) + + results += items[mode] + + return results + + def find_propers(self, search_date=datetime.datetime.today()): + + return self._find_propers(search_date) + + def _get_episode_search_strings(self, ep_obj, add_string='', **kwargs): + + return generic.TorrentProvider._get_episode_search_strings(self, ep_obj, add_string, sep_date='|', use_or=False) + + +class GrabTheInfoCache(tvcache.TVCache): + + def __init__(self, this_provider): + tvcache.TVCache.__init__(self, this_provider) + + self.minTime = 20 # cache update frequency + + def _getRSSData(self): + + return self.provider.get_cache_data() + + +provider = GrabTheInfoProvider()