diff --git a/CHANGES.md b/CHANGES.md index f1d9a469..8b7ce90f 100644 --- a/CHANGES.md +++ b/CHANGES.md @@ -60,6 +60,7 @@ * Add Strike torrent provider * Add TorrentShack torrent provider * Add BeyondHD torrent provider +* Add GFTracker torrent provider * Fix getManualSearchStatus: object has no attribute 'segment' * Change handling of general HTTP error response codes to prevent issues * Add handling for CloudFlare custom HTTP response codes diff --git a/gui/slick/images/providers/GFTracker.png b/gui/slick/images/providers/GFTracker.png new file mode 100644 index 00000000..a9f2d2ff Binary files /dev/null and b/gui/slick/images/providers/GFTracker.png differ diff --git a/sickbeard/__init__.py b/sickbeard/__init__.py index c3295503..96ae22f9 100755 --- a/sickbeard/__init__.py +++ b/sickbeard/__init__.py @@ -36,7 +36,7 @@ sys.path.insert(1, os.path.abspath('../lib')) from sickbeard import providers, metadata, config, webserveInit from sickbeard.providers.generic import GenericProvider from providers import btn, newznab, womble, thepiratebay, torrentleech, kat, iptorrents, \ - omgwtfnzbs, scc, torrentday, hdbits, speedcd, nyaatorrents, torrentbytes, beyondhd, \ + omgwtfnzbs, scc, torrentday, hdbits, speedcd, nyaatorrents, torrentbytes, beyondhd, gftracker, \ freshontv, bitsoup, tokyotoshokan, animenzb, totv, rarbg, morethan, alpharatio, pisexy, strike, torrentshack from sickbeard.config import CheckSection, check_setting_int, check_setting_str, check_setting_float, ConfigMigrator, \ naming_ep_type, minimax diff --git a/sickbeard/providers/__init__.py b/sickbeard/providers/__init__.py index 5b04392e..7f8586d8 100755 --- a/sickbeard/providers/__init__.py +++ b/sickbeard/providers/__init__.py @@ -40,7 +40,8 @@ __all__ = ['womble', 'pisexy', 'strike', 'torrentshack', - 'beyondhd' + 'beyondhd', + 'gftracker', ] from os import sys diff --git a/sickbeard/providers/gftracker.py b/sickbeard/providers/gftracker.py new file mode 100644 index 00000000..53e32324 --- /dev/null +++ b/sickbeard/providers/gftracker.py @@ -0,0 +1,153 @@ +# coding=utf-8 +# +# This file is part of SickGear. +# +# SickGear is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# SickGear is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with SickGear. If not, see . + +import re +import datetime +import time +import traceback + +from . import generic +from sickbeard import logger, tvcache, helpers +from sickbeard.bs4_parser import BS4Parser +from lib.unidecode import unidecode + + +class GFTrackerProvider(generic.TorrentProvider): + + def __init__(self): + generic.TorrentProvider.__init__(self, 'GFTracker') + + self.url_base = 'https://thegft.org/' + self.urls = {'config_provider_home_uri': self.url_base, + 'login_get': self.url_base + 'login.php', + 'login_post': self.url_base + 'loginsite.php', + 'cache': self.url_base + 'browse.php?view=0&c26=1&c37=1&c19=1&c47=1&c17=1&c4=1&searchtype=1', + 'search': '&search=%s', + 'get': self.url_base + '%s'} + + self.url = self.urls['config_provider_home_uri'] + + self.username, self.password, self.minseed, self.minleech = 4 * [None] + self.cache = GFTrackerCache(self) + + def _doLogin(self): + + logged_in = lambda: 'gft_uid' in self.session.cookies and 'gft_pass' in self.session.cookies + if logged_in(): + return True + + if self._checkAuth(): + helpers.getURL(self.urls['login_get'], session=self.session) + login_params = {'username': self.username, 'password': self.password} + response = helpers.getURL(self.urls['login_post'], post_data=login_params, session=self.session) + if response and logged_in(): + return True + + logger.log(u'Failed to authenticate with %s, abort provider.' % self.name, logger.ERROR) + + return False + + def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0): + + results = [] + if not self._doLogin(): + return results + + items = {'Season': [], 'Episode': [], 'Cache': []} + + rc = dict((k, re.compile('(?i)' + v)) for (k, v) in {'info': 'details', 'get': 'download', + 'seeders': '(^\d+)', 'leechers': '(\d+)$'}.items()) + for mode in search_params.keys(): + for search_string in search_params[mode]: + + if isinstance(search_string, unicode): + search_string = unidecode(search_string) + + search_url = self.urls['cache'] + if 'Cache' != mode: + search_url += self.urls['search'] % search_string + + html = self.getURL(search_url) + + cnt = len(items[mode]) + try: + if not html or self._has_no_results(html): + time.sleep(1.1) + raise generic.HaltParseException + + with BS4Parser(html, features=['html5lib', 'permissive']) as soup: + torrent_table = soup.find('div', id='torrentBrowse').find('table') + torrent_rows = [] if not torrent_table else torrent_table.find_all('tr') + + if 2 > len(torrent_rows): + raise generic.HaltParseException + + for tr in torrent_rows[1:]: + try: + seeders, leechers = 2 * [tr.find_all('td')[-1].get_text().strip()] + seeders = int(rc['seeders'].findall(seeders)[0]) + leechers = int(rc['leechers'].findall(leechers)[0]) + if mode != 'Cache' and (seeders < self.minseed or leechers < self.minleech): + continue + + info = tr.find('a', href=rc['info']) + title = ('title' in info.attrs and info['title']) or info.get_text().strip() + + download_url = self.urls['get'] % str(tr.find('a', href=rc['get'])['href']).lstrip('/') + except (AttributeError, TypeError): + continue + + if title and download_url: + items[mode].append((title, download_url, seeders)) + + except generic.HaltParseException: + pass + except Exception: + logger.log(u'Failed to parse. Traceback: %s' % traceback.format_exc(), logger.ERROR) + self._log_result(mode, len(items[mode]) - cnt, search_url) + + 'Cache' != mode and items[mode].sort(key=lambda tup: tup[2], reverse=True) + + results += items[mode] + + return results + + def findPropers(self, search_date=datetime.datetime.today()): + + return self._find_propers(search_date) + + def _get_season_search_strings(self, ep_obj, **kwargs): + + return generic.TorrentProvider._get_season_search_strings(self, ep_obj, scene=False) + + def _get_episode_search_strings(self, ep_obj, add_string='', **kwargs): + + return generic.TorrentProvider._get_episode_search_strings(self, ep_obj, add_string, scene=False, use_or=False) + + +class GFTrackerCache(tvcache.TVCache): + + def __init__(self, this_provider): + tvcache.TVCache.__init__(self, this_provider) + + self.minTime = 17 # cache update frequency + + def _getRSSData(self): + + return self.provider.get_cache_data() + +provider = GFTrackerProvider()