SickGear/sickgear/providers/nebulance.py
2023-02-09 13:41:15 +00:00

235 lines
10 KiB
Python

# coding=utf-8
#
# This file is part of SickGear.
#
# SickGear is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# SickGear is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
import re
import traceback
from . import generic
from .. import common, logger
from ..helpers import try_int
from bs4_parser import BS4Parser
from json_helper import json_dumps
from _23 import filter_list, unidecode, unquote_plus
from six import iteritems
class NebulanceProvider(generic.TorrentProvider):
def __init__(self):
generic.TorrentProvider.__init__(self, 'Nebulance', cache_update_iv=15)
self.url_base = 'https://nebulance.io/'
self.urls = {'config_provider_home_uri': self.url_base,
'login_action': self.url_base + 'login.php',
'user': self.url_base + 'ajax.php?action=index',
'api_key': self.url_base + 'user.php?action=edit&userid=%s',
'api': self.url_base + 'api.php',
'browse': self.url_base + 'ajax.php?action=browse&auth=%s&passkey=%s',
'search': '&searchstr=%s',
'get': self.url_base + 'torrents.php?action=download&authkey=%s&torrent_pass=%s&id=%s'}
self.url = self.urls['config_provider_home_uri']
self.user_authkey, self.user_passkey, self.uid = 3 * [None]
self.chk_td = True
self.username, self.password, self.api_key, self.scene, self.minseed, self.minleech = 6 * [None]
self.api_priority = False
def _authorised(self, **kwargs):
if not super(NebulanceProvider, self)._authorised(
logged_in=(lambda y=None: self.has_all_cookies('session')),
post_params={'keeplogged': '1', 'form_tmpl': True}):
return False
if not self.user_authkey:
response = self.get_url(self.urls['user'], skip_auth=True, parse_json=True)
if self.should_skip():
return False
if 'response' in response:
self.user_authkey, self.user_passkey, self.uid = [response['response'].get(v)
for v in ('authkey', 'passkey', 'id')]
return self.user_authkey
def _search_provider(self, search_params, **kwargs):
if self.api_priority:
return self.api_key and self._search_rpc(search_params) or self._search(search_params)
return self._search(search_params) or self.api_key and self._search_rpc(search_params) or []
def _search(self, search_params):
results = []
if not self._authorised():
return results
items = {'Cache': [], 'Season': [], 'Episode': [], 'Propers': []}
rc = dict([(k, re.compile('(?i)' + v)) for (k, v) in iteritems({'nodots': r'[\.\s]+'})])
for mode in search_params:
for search_string in search_params[mode]:
search_string = unidecode(search_string)
search_url = self.urls['browse'] % (self.user_authkey, self.user_passkey)
if 'Cache' != mode:
search_url += self.urls['search'] % rc['nodots'].sub('+', search_string)
data_json = self.get_url(search_url, parse_json=True)
if self.should_skip():
return results
cnt = len(items[mode])
try:
for item in data_json.get('response', {}).get('results', []):
seeders, leechers, group_name, torrent_id, size, title = [try_int(n, n) for n in [
item.get(x) for x in ['seeders', 'leechers', 'groupName', 'torrentId', 'size', 'rlsName']]]
if self._reject_item(seeders, leechers):
continue
if None is title:
try:
title_parts = group_name.split('[')
maybe_res = re.findall(r'((?:72|108|216)0\w)', title_parts[1])
maybe_ext = re.findall('(?i)(%s)' % '|'.join(common.mediaExtensions), title_parts[1])
detail = title_parts[1].split('/')
detail[1] = detail[1].strip().lower().replace('mkv', 'x264')
with BS4Parser(title_parts[0].strip()).soup as soup:
title = '%s.%s' % (soup.string, '.'.join(
(maybe_res and [maybe_res[0]] or []) +
[detail[0].strip(), detail[1], maybe_ext and maybe_ext[0].lower() or 'mkv']))
except (IndexError, KeyError):
title = self.regulate_title(item, group_name)
download_url = self.urls['get'] % (self.user_authkey, self.user_passkey, torrent_id)
if title and download_url:
items[mode].append((title, download_url, seeders, self._bytesizer(size)))
except (BaseException, Exception):
logger.log(u'Failed to parse. Traceback: %s' % traceback.format_exc(), logger.ERROR)
self._log_search(mode, len(items[mode]) - cnt, search_url)
results = self._sort_seeding(mode, results + items[mode])
return results
@staticmethod
def regulate_title(item, t_param):
if 'tags' not in item or not any(item['tags']):
return t_param
t = ['']
bl = r'[*\[({]+\s*'
br = r'\s*[})\]*]+'
title = re.sub('(?i)(.*?)(%sproper%s)(.*)' % (bl, br), r'\1\3\2', item['groupName'])
for r in (r'\s+-\s+', r'(?:19|20)\d\d(?:\-\d\d\-\d\d)?', r'S\d\d+(?:E\d\d+)?'):
m = re.findall('(.*%s)(.*)' % r, title)
if any(m) and len(m[0][0]) > len(t[0]):
t = m[0]
t = (tuple(title), t)[any(t)]
tag_str = '_'.join(item['tags'])
tags = [re.findall(x, tag_str, flags=re.X) for x in
('(?i)%sProper%s|\bProper\b$' % (bl, br),
r'(?i)\d{3,4}(?:[pi]|hd)',
'''
(?i)(hr.ws.pdtv|blu.?ray|hddvd|
pdtv|hdtv|dsr|tvrip|web.?(?:dl|rip)|dvd.?rip|b[r|d]rip|mpeg-?2)
''', '''
(?i)([hx].?26[45]|divx|xvid)
''', '''
(?i)(avi|mkv|mp4|sub(?:b?ed|pack|s))
''')]
title = ('%s`%s' % (
re.sub('|'.join(['|'.join([re.escape(y) for y in x]) for x in tags if x]).strip('|'), '', t[-1]),
re.sub(r'(?i)(\d{3,4})hd', r'\1p', '`'.join(['`'.join(x) for x in tags[:-1]]).rstrip('`')) +
('', '`hdtv')[not any(tags[2])] + ('', '`x264')[not any(tags[3])]))
for r in [(r'(?i)(?:\W(?:Series|Season))?\W(Repack)\W', r'`\1`'),
('(?i)%s(Proper)%s' % (bl, br), r'`\1`'), (r'%s\s*%s' % (bl, br), '`')]:
title = re.sub(r[0], r[1], title)
grp = filter_list(lambda rn: '.release' in rn.lower(), item['tags'])
title = '%s%s-%s' % (('', t[0])[1 < len(t)], title,
(any(grp) and grp[0] or 'nogrp').upper().replace('.RELEASE', ''))
for r in [(r'\s+[-]?\s+|\s+`|`\s+', '`'), ('`+', '.')]:
title = re.sub(r[0], r[1], title)
title += + any(tags[4]) and ('.%s' % tags[4][0]) or ''
return title
def _search_rpc(self, search_params):
results = []
items = {'Cache': [], 'Season': [], 'Episode': [], 'Propers': []}
json_rpc = (lambda args:
'{"jsonrpc": "2.0", "id": 1, "method": "getTorrents", "params": ["%s", %s, %s, %s, %s]}' %
(self.api_key, json_dumps(args.get('params', '')), 100, args.get('page', 0), 0))
for mode in search_params:
for search_string in search_params[mode]:
search_string = unquote_plus(unidecode(search_string))
params = {'release': search_string}
if 'Cache' == mode:
params = {'age': '< %s' % (24 * 60 * 60)}
elif 'Propers' == mode:
params.update({'age': '< %s' % (4 * 24 * 60 * 60)})
response = self.get_url(self.urls['api'], post_data=json_rpc({'params': params}), parse_json=True)
if self.should_skip():
return results
data_json = response and response.get('result', {}).get('items') or []
cnt = len(items[mode])
for cur_item in data_json:
seeders, leechers, size, download = [try_int(n, n) for n in [
cur_item.get(x) for x in ['seed', 'leech', 'size', 'download']]]
if not self._reject_item(seeders, leechers):
items[mode].append((cur_item.get('rls_name'), '%s%s' % (self.url_base.rstrip('/'), download),
seeders, self._bytesizer(size)))
self._log_search(mode, len(items[mode]) - cnt,
('search_string: ' + str(search_string), self.name)['Cache' == mode])
results = self._sort_seeding(mode, results + items[mode])
return results
def ui_string(self, key):
profile_page = 'profile page'
try:
if self._check_auth(True) and self._authorised():
profile_page = '<a href="%s">%s</a>' % (self.urls['api_key'] % self.uid, profile_page)
except (BaseException, Exception):
pass
return ('%s_api_key' % self.get_id()) == key and 'API key' or \
('%s_api_key_tip' % self.get_id()) == key and \
'\'API key\' is at %s %s with "Download" enabled<br>' \
'%s has no result' \
% (self.name, profile_page,
('and is used if Username/Password', 'Username/Password is used if API')[self.api_priority]) \
or ''
provider = NebulanceProvider()