# Author: Nic Wolfe # URL: http://code.google.com/p/sickbeard/ # # This file is part of SickGear. # # SickGear is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # SickGear is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with SickGear. If not, see . import urllib import time import os try: import xml.etree.cElementTree as etree except ImportError: import elementtree.ElementTree as etree import sickbeard import generic from sickbeard import classes,helpers,scene_exceptions,logger,tvcache from sickbeard import encodingKludge as ek from sickbeard.exceptions import AuthException class NewznabProvider(generic.NZBProvider): def __init__(self, name, url, key='', catIDs='5030,5040', search_mode='eponly', search_fallback=False, enable_recentsearch=False, enable_backlog=False): generic.NZBProvider.__init__(self, name, True, False) self.cache = NewznabCache(self) self.url = url self.key = key self.search_mode = search_mode self.search_fallback = search_fallback self.enable_recentsearch = enable_recentsearch self.enable_backlog = enable_backlog # a 0 in the key spot indicates that no key is needed if self.key == '0': self.needs_auth = False else: self.needs_auth = True if catIDs: self.catIDs = catIDs else: self.catIDs = '5030,5040' self.default = False def configStr(self): return self.name + '|' + self.url + '|' + self.key + '|' + self.catIDs + '|' + str( int(self.enabled)) + '|' + self.search_mode + '|' + str(int(self.search_fallback)) + '|' + str( int(self.enable_recentsearch)) + '|' + str(int(self.enable_backlog)) def imageName(self): if ek.ek(os.path.isfile, ek.ek(os.path.join, sickbeard.PROG_DIR, 'gui', sickbeard.GUI_NAME, 'images', 'providers', self.getID() + '.png')): return self.getID() + '.png' return 'newznab.png' def _getURL(self, url, post_data=None, params=None, timeout=30, json=False): """ By default this is just a simple urlopen call but this method should be overridden for providers with special URL requirements (like cookies) Not really changed much from the superclass, can be used in future. """ # check for auth if not self._doLogin(): return return helpers.getURL(url, post_data=post_data, params=params, headers=self.headers, timeout=timeout, session=self.session, json=json) def get_newznab_categories(self): """ Uses the newznab provider url and apikey to get the capabilities. Makes use of the default newznab caps param. e.a. http://yournewznab/api?t=caps&apikey=skdfiw7823sdkdsfjsfk Returns a tuple with (succes or not, array with dicts [{"id": "5070", "name": "Anime"}, {"id": "5080", "name": "Documentary"}, {"id": "5020", "name": "Foreign"}...etc}], error message) """ return_categories = [] self._checkAuth() params = {"t": "caps"} if self.needs_auth and self.key: params['apikey'] = self.key try: categories = self.getURL("%s/api" % (self.url), params=params, timeout=10) except: logger.log(u"Error getting html for [%s]" % ("%s/api?%s" % (self.url, '&'.join("%s=%s" % (x,y) for x,y in params.items())) ), logger.DEBUG) return (False, return_categories, "Error getting html for [%s]" % ("%s/api?%s" % (self.url, '&'.join("%s=%s" % (x,y) for x,y in params.items()) ))) xml_categories = helpers.parse_xml(categories) if not xml_categories: logger.log(u"Error parsing xml for [%s]" % (self.name), logger.DEBUG) return (False, return_categories, "Error parsing xml for [%s]" % (self.name)) try: for category in xml_categories.iter('category'): if category.get('name') == 'TV': for subcat in category.findall('subcat'): return_categories.append(subcat.attrib) except: logger.log(u"Error parsing result for [%s]" % (self.name), logger.DEBUG) return (False, return_categories, "Error parsing result for [%s]" % (self.name)) return (True, return_categories, "") def _get_season_search_strings(self, ep_obj): to_return = [] cur_params = {} # season if ep_obj.show.air_by_date or ep_obj.show.sports: date_str = str(ep_obj.airdate).split('-')[0] cur_params['season'] = date_str cur_params['q'] = date_str.replace('-', '.') elif ep_obj.show.is_anime: cur_params['season'] = "%d" % ep_obj.scene_absolute_number else: cur_params['season'] = str(ep_obj.scene_season) # search rid = helpers.mapIndexersToShow(ep_obj.show)[2] if rid: cur_return = cur_params.copy() cur_return['rid'] = rid to_return.append(cur_return) # add new query strings for exceptions name_exceptions = list( set([helpers.sanitizeSceneName(a) for a in scene_exceptions.get_scene_exceptions(ep_obj.show.indexerid) + [ep_obj.show.name]])) for cur_exception in name_exceptions: cur_return = cur_params.copy() if 'q' in cur_return: cur_return['q'] = cur_exception + '.' + cur_return['q'] to_return.append(cur_return) return to_return def _get_episode_search_strings(self, ep_obj, add_string=''): to_return = [] params = {} if not ep_obj: return [params] if ep_obj.show.air_by_date or ep_obj.show.sports: date_str = str(ep_obj.airdate) params['season'] = date_str.partition('-')[0] params['ep'] = date_str.partition('-')[2].replace('-', '/') elif ep_obj.show.anime: params['ep'] = "%i" % int(ep_obj.scene_absolute_number if int(ep_obj.scene_absolute_number) > 0 else ep_obj.scene_episode) else: params['season'] = ep_obj.scene_season params['ep'] = ep_obj.scene_episode # search rid = helpers.mapIndexersToShow(ep_obj.show)[2] if rid: cur_return = params.copy() cur_return['rid'] = rid to_return.append(cur_return) # add new query strings for exceptions name_exceptions = list( set([helpers.sanitizeSceneName(a) for a in scene_exceptions.get_scene_exceptions(ep_obj.show.indexerid) + [ep_obj.show.name]])) for cur_exception in name_exceptions: cur_return = params.copy() cur_return['q'] = cur_exception to_return.append(cur_return) if ep_obj.show.anime: # Experimental, add a searchstring without search explicitly for the episode! # Remove the ?ep=e46 paramater and use add the episode number to the query paramater. # Can be usefull for newznab indexers that do not have the episodes 100% parsed. # Start with only applying the searchstring to anime shows params['q'] = cur_exception paramsNoEp = params.copy() paramsNoEp['q'] = '%s.%02d' % (paramsNoEp['q'], int(paramsNoEp['ep'])) if "ep" in paramsNoEp: paramsNoEp.pop("ep") to_return.append(paramsNoEp) return to_return def _doGeneralSearch(self, search_string): return self._doSearch({'q': search_string}) def _checkAuth(self): if self.needs_auth and not self.key: logger.log(u"Incorrect authentication credentials for " + self.name + " : " + "API key is missing", logger.DEBUG) raise AuthException("Your authentication credentials for " + self.name + " are missing, check your config.") return True def _checkAuthFromData(self, data): if data is None: return self._checkAuth() if 'error' in data.feed: code = data.feed['error']['code'] if code == '100': raise AuthException("Your API key for " + self.name + " is incorrect, check your config.") elif code == '101': raise AuthException("Your account on " + self.name + " has been suspended, contact the administrator.") elif code == '102': raise AuthException( "Your account isn't allowed to use the API on " + self.name + ", contact the administrator") elif code == '910': logger.log(u"" + self.name + " currently has their API disabled, please check with provider.", logger.WARNING) return False else: logger.log(u"Unknown error given from " + self.name + ": " + data.feed['error']['description'], logger.ERROR) return False return True def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0): self._checkAuth() params = {"t": "tvsearch", "maxage": sickbeard.USENET_RETENTION, "limit": 100, "attrs": "rageid", "offset": 0} # category ids if self.show and self.show.is_sports: params['cat'] = self.catIDs + ',5060' elif self.show and self.show.is_anime: params['cat'] = self.catIDs + ',5070' else: params['cat'] = self.catIDs # if max_age is set, use it, don't allow it to be missing if age or not params['maxage']: params['maxage'] = age if search_params: params.update(search_params) if 'rid' not in search_params and 'q' not in search_params: logger.log("Error no rid or search term given. Report to forums with a full debug log") return [] if self.needs_auth and self.key: params['apikey'] = self.key results = [] offset = total = 0 while (total >= offset) and (offset < 1000): search_url = self.url + 'api?' + urllib.urlencode(params) logger.log(u"Search url: " + search_url, logger.DEBUG) data = self.cache.getRSSFeed(search_url) if not data or not self._checkAuthFromData(data): break for item in data.entries: (title, url) = self._get_title_and_url(item) if title and url: results.append(item) else: logger.log( u"The data returned from " + self.name + " is incomplete, this result is unusable", logger.DEBUG) # get total and offset attribs try: if total == 0: total = int(data.feed.newznab_response['total'] or 0) offset = int(data.feed.newznab_response['offset'] or 0) except AttributeError: break # No items found, prevent from doing another search if total == 0: break if offset != params['offset']: logger.log("Tell your newznab provider to fix their bloody newznab responses") break params['offset'] += params['limit'] if (total > int(params['offset'])): offset = int(params['offset']) # if there are more items available then the amount given in one call, grab some more logger.log(str( total - int(params['offset'])) + " more items to be fetched from provider. Fetching another " + str( params['limit']) + " items.", logger.DEBUG) else: logger.log(str( total - int(params['offset'])) + " No more searches needed, couldn't find anything I was looking for! " + str( params['limit']) + " items.", logger.DEBUG) break time.sleep(0.2) return results def findPropers(self, search_date=None): return self._find_propers(search_date) class NewznabCache(tvcache.TVCache): def __init__(self, provider): tvcache.TVCache.__init__(self, provider) # only poll newznab providers every 15 minutes max self.minTime = 15 def _getRSSData(self): params = {"t": "tvsearch", "cat": self.provider.catIDs + ',5060,5070', "attrs": "rageid"} if self.provider.needs_auth and self.provider.key: params['apikey'] = self.provider.key rss_url = self.provider.url + 'api?' + urllib.urlencode(params) logger.log(self.provider.name + " cache update URL: " + rss_url, logger.DEBUG) return self.getRSSFeed(rss_url) def _checkAuth(self, data): return self.provider._checkAuthFromData(data) def updateCache(self): if self.shouldUpdate() and self._checkAuth(None): data = self._getRSSData() # as long as the http request worked we count this as an update if not data: return [] # clear cache self._clearCache() self.setLastUpdate() if self._checkAuth(data): items = data.entries cl = [] for item in items: ci = self._parseItem(item) if ci is not None: cl.append(ci) if len(cl) > 0: myDB = self._getDB() myDB.mass_action(cl) else: raise AuthException( u"Your authentication credentials for " + self.provider.name + " are incorrect, check your config") return [] # overwrite method with that parses the rageid from the newznab feed def _parseItem(self, item): title = item.title url = item.link attrs = item.newznab_attr if not isinstance(attrs, list): attrs = [item.newznab_attr] tvrageid = 0 for attr in attrs: if attr['name'] == 'tvrageid': tvrageid = int(attr['value']) break self._checkItemAuth(title, url) if not title or not url: logger.log( u"The data returned from the " + self.provider.name + " feed is incomplete, this result is unusable", logger.DEBUG) return None url = self._translateLinkURL(url) logger.log(u"Attempting to add item from RSS to cache: " + title, logger.DEBUG) return self._addCacheEntry(title, url, indexer_id=tvrageid)