# coding=utf-8 # # This file is part of SickGear. # # SickGear is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # SickGear is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with SickGear. If not, see . import re import traceback from . import generic from .. import common, logger from ..helpers import try_int from bs4_parser import BS4Parser from json_helper import json_dumps from _23 import unquote_plus from six import iteritems class NebulanceProvider(generic.TorrentProvider): def __init__(self): generic.TorrentProvider.__init__(self, 'Nebulance', cache_update_iv=15) self.url_base = 'https://nebulance.io/' self.urls = {'config_provider_home_uri': self.url_base, 'login_action': self.url_base + 'login.php', 'user': self.url_base + 'ajax.php?action=index', 'api_key': self.url_base + 'user.php?action=edit&userid=%s', 'api': self.url_base + 'api.php', 'browse': self.url_base + 'ajax.php?action=browse&auth=%s&passkey=%s', 'search': '&searchstr=%s', 'get': self.url_base + 'torrents.php?action=download&authkey=%s&torrent_pass=%s&id=%s'} self.url = self.urls['config_provider_home_uri'] self.user_authkey, self.user_passkey, self.uid = 3 * [None] self.chk_td = True self.username, self.password, self.api_key, self.scene, self.minseed, self.minleech = 6 * [None] self.api_priority = False def _authorised(self, **kwargs): if not super(NebulanceProvider, self)._authorised( logged_in=(lambda y=None: self.has_all_cookies('session')), post_params={'keeplogged': '1', 'form_tmpl': True}): return False if not self.user_authkey: response = self.get_url(self.urls['user'], skip_auth=True, parse_json=True) if self.should_skip(): return False if 'response' in response: self.user_authkey, self.user_passkey, self.uid = [response['response'].get(v) for v in ('authkey', 'passkey', 'id')] return self.user_authkey def _search_provider(self, search_params, **kwargs): if self.api_priority: return self.api_key and self._search_rpc(search_params) or self._search(search_params) return self._search(search_params) or self.api_key and self._search_rpc(search_params) or [] def _search(self, search_params): results = [] if not self._authorised(): return results items = {'Cache': [], 'Season': [], 'Episode': [], 'Propers': []} rc = dict([(k, re.compile('(?i)' + v)) for (k, v) in iteritems({'nodots': r'[\.\s]+'})]) for mode in search_params: for search_string in search_params[mode]: search_url = self.urls['browse'] % (self.user_authkey, self.user_passkey) if 'Cache' != mode: search_url += self.urls['search'] % rc['nodots'].sub('+', search_string) data_json = self.get_url(search_url, parse_json=True) if self.should_skip(): return results cnt = len(items[mode]) try: for item in data_json.get('response', {}).get('results', []): seeders, leechers, group_name, torrent_id, size, title = [try_int(n, n) for n in [ item.get(x) for x in ['seeders', 'leechers', 'groupName', 'torrentId', 'size', 'rlsName']]] if self._reject_item(seeders, leechers): continue if None is title: try: title_parts = group_name.split('[') maybe_res = re.findall(r'((?:72|108|216)0\w)', title_parts[1]) maybe_ext = re.findall('(?i)(%s)' % '|'.join(common.mediaExtensions), title_parts[1]) detail = title_parts[1].split('/') detail[1] = detail[1].strip().lower().replace('mkv', 'x264') with BS4Parser(title_parts[0].strip()).soup as soup: title = '%s.%s' % (soup.string, '.'.join( (maybe_res and [maybe_res[0]] or []) + [detail[0].strip(), detail[1], maybe_ext and maybe_ext[0].lower() or 'mkv'])) except (IndexError, KeyError): title = self.regulate_title(item, group_name) download_url = self.urls['get'] % (self.user_authkey, self.user_passkey, torrent_id) if title and download_url: items[mode].append((title, download_url, seeders, self._bytesizer(size))) except (BaseException, Exception): logger.log(u'Failed to parse. Traceback: %s' % traceback.format_exc(), logger.ERROR) self._log_search(mode, len(items[mode]) - cnt, search_url) results = self._sort_seeding(mode, results + items[mode]) return results @staticmethod def regulate_title(item, t_param): if 'tags' not in item or not any(item['tags']): return t_param t = [''] bl = r'[*\[({]+\s*' br = r'\s*[})\]*]+' title = re.sub('(?i)(.*?)(%sproper%s)(.*)' % (bl, br), r'\1\3\2', item['groupName']) for r in (r'\s+-\s+', r'(?:19|20)\d\d(?:\-\d\d\-\d\d)?', r'S\d\d+(?:E\d\d+)?'): m = re.findall('(.*%s)(.*)' % r, title) if any(m) and len(m[0][0]) > len(t[0]): t = m[0] t = (tuple(title), t)[any(t)] tag_str = '_'.join(item['tags']) tags = [re.findall(x, tag_str, flags=re.X) for x in ('(?i)%sProper%s|\bProper\b$' % (bl, br), r'(?i)\d{3,4}(?:[pi]|hd)', ''' (?i)(hr.ws.pdtv|blu.?ray|hddvd| pdtv|hdtv|dsr|tvrip|web.?(?:dl|rip)|dvd.?rip|b[r|d]rip|mpeg-?2) ''', ''' (?i)([hx].?26[45]|divx|xvid) ''', ''' (?i)(avi|mkv|mp4|sub(?:b?ed|pack|s)) ''')] title = ('%s`%s' % ( re.sub('|'.join(['|'.join([re.escape(y) for y in x]) for x in tags if x]).strip('|'), '', t[-1]), re.sub(r'(?i)(\d{3,4})hd', r'\1p', '`'.join(['`'.join(x) for x in tags[:-1]]).rstrip('`')) + ('', '`hdtv')[not any(tags[2])] + ('', '`x264')[not any(tags[3])])) for r in [(r'(?i)(?:\W(?:Series|Season))?\W(Repack)\W', r'`\1`'), ('(?i)%s(Proper)%s' % (bl, br), r'`\1`'), (r'%s\s*%s' % (bl, br), '`')]: title = re.sub(r[0], r[1], title) grp = list(filter(lambda rn: '.release' in rn.lower(), item['tags'])) title = '%s%s-%s' % (('', t[0])[1 < len(t)], title, (any(grp) and grp[0] or 'nogrp').upper().replace('.RELEASE', '')) for r in [(r'\s+[-]?\s+|\s+`|`\s+', '`'), ('`+', '.')]: title = re.sub(r[0], r[1], title) title += + any(tags[4]) and ('.%s' % tags[4][0]) or '' return title def _search_rpc(self, search_params): results = [] items = {'Cache': [], 'Season': [], 'Episode': [], 'Propers': []} json_rpc = (lambda args: '{"jsonrpc": "2.0", "id": 1, "method": "getTorrents", "params": ["%s", %s, %s, %s, %s]}' % (self.api_key, json_dumps(args.get('params', '')), 100, args.get('page', 0), 0)) for mode in search_params: for search_string in search_params[mode]: search_string = unquote_plus(search_string) params = {'release': search_string} if 'Cache' == mode: params = {'age': '< %s' % (24 * 60 * 60)} elif 'Propers' == mode: params.update({'age': '< %s' % (4 * 24 * 60 * 60)}) response = self.get_url(self.urls['api'], post_data=json_rpc({'params': params}), parse_json=True) if self.should_skip(): return results data_json = response and response.get('result', {}).get('items') or [] cnt = len(items[mode]) for cur_item in data_json: seeders, leechers, size, download = [try_int(n, n) for n in [ cur_item.get(x) for x in ['seed', 'leech', 'size', 'download']]] if not self._reject_item(seeders, leechers): items[mode].append((cur_item.get('rls_name'), '%s%s' % (self.url_base.rstrip('/'), download), seeders, self._bytesizer(size))) self._log_search(mode, len(items[mode]) - cnt, ('search_string: ' + str(search_string), self.name)['Cache' == mode]) results = self._sort_seeding(mode, results + items[mode]) return results def ui_string(self, key): profile_page = 'profile page' try: if self._check_auth(True) and self._authorised(): profile_page = '%s' % (self.urls['api_key'] % self.uid, profile_page) except (BaseException, Exception): pass return ('%s_api_key' % self.get_id()) == key and 'API key' or \ ('%s_api_key_tip' % self.get_id()) == key and \ '\'API key\' is at %s %s with "Download" enabled
' \ '%s has no result' \ % (self.name, profile_page, ('and is used if Username/Password', 'Username/Password is used if API')[self.api_priority]) \ or '' provider = NebulanceProvider()