# coding=utf-8 # Author: Daniel Heimans # URL: http://code.google.com/p/sickbeard # # This file is part of SickGear. # # SickGear is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # SickGear is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with SickGear. If not, see . import time import socket import math import sickbeard import generic from sickbeard import classes from sickbeard import scene_exceptions from sickbeard import logger from sickbeard import tvcache from sickbeard.helpers import sanitizeSceneName from sickbeard.exceptions import ex, AuthException from lib import jsonrpclib from datetime import datetime class BTNProvider(generic.TorrentProvider): def __init__(self): generic.TorrentProvider.__init__(self, "BTN") self.supportsBacklog = True self.supportsAbsoluteNumbering = True self.enabled = False self.api_key = None self.ratio = None self.cache = BTNCache(self) self.url = "http://api.btnapps.net" def isEnabled(self): return self.enabled def imageName(self): return 'btn.png' def _checkAuth(self): if not self.api_key: raise AuthException("Your authentication credentials for " + self.name + " are missing, check your config.") return True def _checkAuthFromData(self, parsedJSON): if parsedJSON is None: return self._checkAuth() if 'api-error' in parsedJSON: logger.log(u"Incorrect authentication credentials for " + self.name + " : " + parsedJSON['api-error'], logger.DEBUG) raise AuthException( "Your authentication credentials for " + self.name + " are incorrect, check your config.") return True def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0): self._checkAuth() params = {} apikey = self.api_key # age in seconds if age: params['age'] = "<=" + str(int(age)) if search_params: params.update(search_params) parsedJSON = self._api_call(apikey, params) if not parsedJSON: logger.log(u"No data returned from " + self.name, logger.ERROR) return [] if self._checkAuthFromData(parsedJSON): if 'torrents' in parsedJSON: found_torrents = parsedJSON['torrents'] else: found_torrents = {} # We got something, we know the API sends max 1000 results at a time. # See if there are more than 1000 results for our query, if not we # keep requesting until we've got everything. # max 150 requests per hour so limit at that. Scan every 15 minutes. 60 / 15 = 4. max_pages = 150 results_per_page = 1000 if 'results' in parsedJSON and int(parsedJSON['results']) >= results_per_page: pages_needed = int(math.ceil(int(parsedJSON['results']) / results_per_page)) if pages_needed > max_pages: pages_needed = max_pages # +1 because range(1,4) = 1, 2, 3 for page in range(1, pages_needed + 1): parsedJSON = self._api_call(apikey, params, results_per_page, page * results_per_page) # Note that this these are individual requests and might time out individually. This would result in 'gaps' # in the results. There is no way to fix this though. if 'torrents' in parsedJSON: found_torrents.update(parsedJSON['torrents']) results = [] for torrentid, torrent_info in found_torrents.iteritems(): (title, url) = self._get_title_and_url(torrent_info) if title and url: results.append(torrent_info) return results return [] def _api_call(self, apikey, params={}, results_per_page=1000, offset=0): server = jsonrpclib.Server(self.url) parsedJSON = {} logger.log(u'Searching with parameters: %s' % params, logger.DEBUG) try: parsedJSON = server.getTorrents(apikey, params, int(results_per_page), int(offset)) except jsonrpclib.jsonrpc.ProtocolError, error: logger.log(u"JSON-RPC protocol error while accessing " + self.name + ": " + ex(error), logger.ERROR) parsedJSON = {'api-error': ex(error)} return parsedJSON except socket.timeout: logger.log(u"Timeout while accessing " + self.name, logger.WARNING) except socket.error, error: # Note that sometimes timeouts are thrown as socket errors logger.log(u"Socket error while accessing " + self.name + ": " + error[1], logger.ERROR) except Exception, error: errorstring = str(error) if (errorstring.startswith('<') and errorstring.endswith('>')): errorstring = errorstring[1:-1] logger.log(u"Unknown error while accessing " + self.name + ": " + errorstring, logger.ERROR) return parsedJSON def _get_title_and_url(self, parsedJSON): # The BTN API gives a lot of information in response, # however SickGear is built mostly around Scene or # release names, which is why we are using them here. if 'ReleaseName' in parsedJSON and parsedJSON['ReleaseName']: title = parsedJSON['ReleaseName'] else: # If we don't have a release name we need to get creative title = u'' if 'Series' in parsedJSON: title += parsedJSON['Series'] if 'GroupName' in parsedJSON: title += '.' + parsedJSON['GroupName'] if title else parsedJSON['GroupName'] if 'Resolution' in parsedJSON: title += '.' + parsedJSON['Resolution'] if title else parsedJSON['Resolution'] if 'Source' in parsedJSON: title += '.' + parsedJSON['Source'] if title else parsedJSON['Source'] if 'Codec' in parsedJSON: title += '.' + parsedJSON['Codec'] if title else parsedJSON['Codec'] if title: title = title.replace(' ', '.') url = None if 'DownloadURL' in parsedJSON: url = parsedJSON['DownloadURL'] if url: # unescaped / is valid in JSON, but it can be escaped url = url.replace("\\/", "/") return (title, url) def _get_season_search_strings(self, ep_obj): search_params = [] current_params = {'category': 'Season'} # Search for entire seasons: no need to do special things for air by date or sports shows if ep_obj.show.air_by_date or ep_obj.show.sports: # Search for the year of the air by date show current_params['name'] = str(ep_obj.airdate).split('-')[0] else: current_params['name'] = 'Season ' + str(ep_obj.scene_season) # search if ep_obj.show.indexer == 1: current_params['tvdb'] = ep_obj.show.indexerid search_params.append(current_params) elif ep_obj.show.indexer == 2: current_params['tvrage'] = ep_obj.show.indexerid search_params.append(current_params) else: name_exceptions = list( set(scene_exceptions.get_scene_exceptions(ep_obj.show.indexerid) + [ep_obj.show.name])) for name in name_exceptions: # Search by name if we don't have tvdb or tvrage id current_params['series'] = sanitizeSceneName(name) search_params.append(current_params) return search_params def _get_episode_search_strings(self, ep_obj, add_string=''): if not ep_obj: return [{}] to_return = [] search_params = {'category': 'Episode'} # episode if ep_obj.show.air_by_date or ep_obj.show.sports: date_str = str(ep_obj.airdate) # BTN uses dots in dates, we just search for the date since that # combined with the series identifier should result in just one episode search_params['name'] = date_str.replace('-', '.') else: # Do a general name search for the episode, formatted like SXXEYY search_params['name'] = "S%02dE%02d" % (ep_obj.scene_season, ep_obj.scene_episode) # search if ep_obj.show.indexer == 1: search_params['tvdb'] = ep_obj.show.indexerid to_return.append(search_params) elif ep_obj.show.indexer == 2: search_params['tvrage'] = ep_obj.show.indexerid to_return.append(search_params) else: # add new query string for every exception name_exceptions = list( set(scene_exceptions.get_scene_exceptions(ep_obj.show.indexerid) + [ep_obj.show.name])) for cur_exception in name_exceptions: search_params['series'] = sanitizeSceneName(cur_exception) to_return.append(search_params) return to_return def _doGeneralSearch(self, search_string): # 'search' looks as broad is it can find. Can contain episode overview and title for example, # use with caution! return self._doSearch({'search': search_string}) def findPropers(self, search_date=None): results = [] search_terms = ['%.proper.%', '%.repack.%'] for term in search_terms: for item in self._doSearch({'release': term}, age=4 * 24 * 60 * 60): if item['Time']: try: result_date = datetime.fromtimestamp(float(item['Time'])) except TypeError: result_date = None if result_date: if not search_date or result_date > search_date: title, url = self._get_title_and_url(item) results.append(classes.Proper(title, url, result_date, self.show)) return results def seedRatio(self): return self.ratio class BTNCache(tvcache.TVCache): def __init__(self, provider): tvcache.TVCache.__init__(self, provider) # At least 15 minutes between queries self.minTime = 15 def _getRSSData(self): # Get the torrents uploaded since last check. seconds_since_last_update = math.ceil(time.time() - time.mktime(self._getLastUpdate().timetuple())) # default to 15 minutes seconds_minTime = self.minTime * 60 if seconds_since_last_update < seconds_minTime: seconds_since_last_update = seconds_minTime # Set maximum to 24 hours (24 * 60 * 60 = 86400 seconds) of "RSS" data search, older things will need to be done through backlog if seconds_since_last_update > 86400: logger.log( u"The last known successful update on " + self.provider.name + " was more than 24 hours ago, only trying to fetch the last 24 hours!", logger.WARNING) seconds_since_last_update = 86400 return self.provider._doSearch(search_params=None, age=seconds_since_last_update) provider = BTNProvider()