mirror of
https://github.com/SickGear/SickGear.git
synced 2024-12-11 05:33:37 +00:00
Merge pull request #464 from JackDandy/feature/ChangeProvRefactorAndPEP8
Change refactor, PEP8, code convention cleanse for nzb/torrent code…
This commit is contained in:
commit
40bcc38c5b
51 changed files with 1083 additions and 1191 deletions
10
.gitignore
vendored
10
.gitignore
vendored
|
@ -1,5 +1,5 @@
|
|||
# SB User Related #
|
||||
######################
|
||||
# SB User Related #
|
||||
cache/*
|
||||
cache.db*
|
||||
config.ini*
|
||||
|
@ -11,18 +11,18 @@ server.crt
|
|||
server.key
|
||||
restore/
|
||||
|
||||
# SB Test Related #
|
||||
######################
|
||||
# SB Test Related #
|
||||
tests/Logs/*
|
||||
tests/sickbeard.*
|
||||
tests/cache.db
|
||||
|
||||
# Compiled source #
|
||||
######################
|
||||
# Compiled source #
|
||||
*.py[co]
|
||||
|
||||
# IDE specific #
|
||||
######################
|
||||
# IDE specific #
|
||||
*.bak
|
||||
*.tmp
|
||||
*.wpr
|
||||
|
@ -35,8 +35,8 @@ tests/cache.db
|
|||
Session.vim
|
||||
.ropeproject/*
|
||||
|
||||
# OS generated files #
|
||||
######################
|
||||
# OS generated files #
|
||||
.Spotlight-V100
|
||||
.Trashes
|
||||
.DS_Store
|
||||
|
|
|
@ -1,13 +1,17 @@
|
|||
from distutils.core import setup
|
||||
import py2exe, sys, shutil
|
||||
import sys
|
||||
import shutil
|
||||
try:
|
||||
import py2exe
|
||||
except:
|
||||
pass
|
||||
|
||||
sys.argv.append('py2exe')
|
||||
|
||||
setup(
|
||||
options = {'py2exe': {'bundle_files': 1}},
|
||||
# windows = [{'console': "sabToSickbeard.py"}],
|
||||
zipfile = None,
|
||||
console = ['sabToSickbeard.py'],
|
||||
)
|
||||
setup(options={'py2exe': {'bundle_files': 1}},
|
||||
# windows = [{'console': "sabToSickbeard.py"}],
|
||||
zipfile=None,
|
||||
console=['sabToSickbeard.py']
|
||||
)
|
||||
|
||||
shutil.copy('dist/sabToSickbeard.exe', '.')
|
||||
|
|
Before Width: | Height: | Size: 886 B After Width: | Height: | Size: 886 B |
|
@ -3,8 +3,8 @@
|
|||
#from sickbeard.providers import thepiratebay
|
||||
#from sickbeard.helpers import anon_url, starify
|
||||
##
|
||||
#set global $title="Config - Providers"
|
||||
#set global $header="Search Providers"
|
||||
#set global $title = 'Config - Providers'
|
||||
#set global $header = 'Search Providers'
|
||||
#set global $sbPath = '../..'
|
||||
#set global $topmenu = 'config'
|
||||
##
|
||||
|
@ -39,7 +39,7 @@
|
|||
|
||||
#for $curNewznabProvider in $sickbeard.newznabProviderList:
|
||||
|
||||
\$(this).addProvider('$curNewznabProvider.getID()', '$curNewznabProvider.name', '$curNewznabProvider.url', '<%= starify(curNewznabProvider.key) %>', '$curNewznabProvider.cat_ids', $int($curNewznabProvider.default), show_nzb_providers);
|
||||
\$(this).addProvider('$curNewznabProvider.get_id()', '$curNewznabProvider.name', '$curNewznabProvider.url', '<%= starify(curNewznabProvider.key) %>', '$curNewznabProvider.cat_ids', $int($curNewznabProvider.default), show_nzb_providers);
|
||||
|
||||
#end for
|
||||
|
||||
|
@ -49,7 +49,7 @@
|
|||
|
||||
#for $curTorrentRssProvider in $sickbeard.torrentRssProviderList:
|
||||
|
||||
\$(this).addTorrentRssProvider('$curTorrentRssProvider.getID()', '$curTorrentRssProvider.name', '$curTorrentRssProvider.url', '<%= starify(curTorrentRssProvider.cookies) %>');
|
||||
\$(this).addTorrentRssProvider('$curTorrentRssProvider.get_id()', '$curTorrentRssProvider.name', '$curTorrentRssProvider.url', '<%= starify(curTorrentRssProvider.cookies) %>');
|
||||
|
||||
#end for
|
||||
|
||||
|
@ -90,7 +90,7 @@
|
|||
<p>At least one provider is required but two are recommended.</p>
|
||||
|
||||
#if $methods_notused
|
||||
<blockquote style="margin: 20px 0"><%= '/'.join(x for x in methods_notused) %> providers can be enabled in <a href="$sbRoot/config/search/">Search Settings</a></blockquote>
|
||||
<blockquote style="margin:20px 0"><%= '/'.join(x for x in methods_notused) %> providers can be enabled in <a href="$sbRoot/config/search/">Search Settings</a></blockquote>
|
||||
#else
|
||||
<br/>
|
||||
#end if
|
||||
|
@ -104,11 +104,11 @@
|
|||
#elif $curProvider.providerType == $GenericProvider.TORRENT and not $sickbeard.USE_TORRENTS
|
||||
#continue
|
||||
#end if
|
||||
#set $curName = $curProvider.getID()
|
||||
#set $curName = $curProvider.get_id()
|
||||
<li class="ui-state-default" id="$curName">
|
||||
<input type="checkbox" id="enable_$curName" class="provider_enabler" <%= html_checked if curProvider.isEnabled() else '' %>/>
|
||||
<a href="<%= anon_url(curProvider.url) %>" class="imgLink" rel="noreferrer" onclick="window.open(this.href, '_blank'); return false;"><img src="$sbRoot/images/providers/$curProvider.imageName()" alt="$curProvider.name" title="$curProvider.name" width="16" height="16" style="vertical-align:middle;"/></a>
|
||||
<span style="vertical-align:middle;">$curProvider.name</span>
|
||||
<input type="checkbox" id="enable_$curName" class="provider_enabler" <%= html_checked if curProvider.is_enabled() else '' %>/>
|
||||
<a href="<%= anon_url(curProvider.url) %>" class="imgLink" rel="noreferrer" onclick="window.open(this.href, '_blank'); return false;"><img src="$sbRoot/images/providers/$curProvider.image_name()" alt="$curProvider.name" title="$curProvider.name" width="16" height="16" style="vertical-align:middle;"/></a>
|
||||
<span style="vertical-align:middle">$curProvider.name</span>
|
||||
<%= '*' if not curProvider.supportsBacklog else '' %>
|
||||
<span class="ui-icon ui-icon-arrowthick-2-n-s pull-right" style="margin-top:3px"></span>
|
||||
</li>
|
||||
|
@ -124,7 +124,7 @@
|
|||
##<h4 class="note">!</h4><p class="note">Provider is <b>NOT WORKING</b></p>
|
||||
</div>
|
||||
|
||||
<input type="hidden" name="provider_order" id="provider_order" value="<%=" ".join([x.getID()+':'+str(int(x.isEnabled())) for x in sickbeard.providers.sortedProviderList()])%>"/>
|
||||
<input type="hidden" name="provider_order" id="provider_order" value="<%=' '.join([x.get_id()+':'+str(int(x.is_enabled())) for x in sickbeard.providers.sortedProviderList()])%>"/>
|
||||
<div style="width: 300px; float: right">
|
||||
<div style="margin: 0px auto; width: 101px">
|
||||
<input type="submit" class="btn config_submitter" value="Save Changes" />
|
||||
|
@ -156,7 +156,7 @@
|
|||
#elif $curProvider.providerType == $GenericProvider.TORRENT and not $sickbeard.USE_TORRENTS
|
||||
#continue
|
||||
#end if
|
||||
#if $curProvider.isEnabled()
|
||||
#if $curProvider.is_enabled()
|
||||
$provider_config_list_enabled.append($curProvider)
|
||||
#else
|
||||
$provider_config_list.append($curProvider)
|
||||
|
@ -168,14 +168,14 @@
|
|||
#if $provider_config_list_enabled
|
||||
<optgroup label="Enabled...">
|
||||
#for $cur_provider in $provider_config_list_enabled:
|
||||
<option value="$cur_provider.getID()">$cur_provider.name</option>
|
||||
<option value="$cur_provider.get_id()">$cur_provider.name</option>
|
||||
#end for
|
||||
</optgroup>
|
||||
#end if
|
||||
#if $provider_config_list
|
||||
<optgroup label="Not Enabled...">
|
||||
#for $cur_provider in $provider_config_list
|
||||
<option value="$cur_provider.getID()">$cur_provider.name</option>
|
||||
<option value="$cur_provider.get_id()">$cur_provider.name</option>
|
||||
#end for
|
||||
</optgroup>
|
||||
#end if
|
||||
|
@ -187,76 +187,71 @@
|
|||
</label>
|
||||
</div>
|
||||
|
||||
|
||||
<!-- start div for editing providers //-->
|
||||
#for $curNewznabProvider in [$curProvider for $curProvider in $sickbeard.newznabProviderList]
|
||||
<div class="providerDiv" id="${curNewznabProvider.getID()}Div">
|
||||
<div class="providerDiv" id="${curNewznabProvider.get_id()}Div">
|
||||
#if $curNewznabProvider.default and $curNewznabProvider.needs_auth
|
||||
<div class="field-pair">
|
||||
<label for="${curNewznabProvider.getID()}_url">
|
||||
<label for="${curNewznabProvider.get_id()}_url">
|
||||
<span class="component-title">URL</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" id="${curNewznabProvider.getID()}_url" value="$curNewznabProvider.url" class="form-control input-sm input350" disabled/>
|
||||
<input type="text" id="${curNewznabProvider.get_id()}_url" value="$curNewznabProvider.url" class="form-control input-sm input350" disabled/>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
||||
<div class="field-pair">
|
||||
<label for="${curNewznabProvider.getID()}_hash">
|
||||
<label for="${curNewznabProvider.get_id()}_hash">
|
||||
<span class="component-title">API key</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" id="${curNewznabProvider.getID()}_hash" value="<%= starify(curNewznabProvider.key) %>" newznab_name="${curNewznabProvider.getID()}_hash" class="newznab_key form-control input-sm input350" />
|
||||
<input type="text" id="${curNewznabProvider.get_id()}_hash" value="<%= starify(curNewznabProvider.key) %>" newznab_name="${curNewznabProvider.get_id()}_hash" class="newznab_key form-control input-sm input350" />
|
||||
<div class="clear-left"><p>get API key from provider website</p></div>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNewznabProvider, 'enable_recentsearch'):
|
||||
#if $hasattr($curNewznabProvider, 'enable_recentsearch') and $curNewznabProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curNewznabProvider.getID()}_enable_recentsearch">
|
||||
<label for="${curNewznabProvider.get_id()}_enable_recentsearch">
|
||||
<span class="component-title">Enable recent searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNewznabProvider.getID()}_enable_recentsearch" id="${curNewznabProvider.getID()}_enable_recentsearch" <%= html_checked if curNewznabProvider.enable_recentsearch else '' %>/>
|
||||
<input type="checkbox" name="${curNewznabProvider.get_id()}_enable_recentsearch" id="${curNewznabProvider.get_id()}_enable_recentsearch" <%= html_checked if curNewznabProvider.enable_recentsearch else '' %>/>
|
||||
<p>perform recent searches at provider</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNewznabProvider, 'enable_backlog'):
|
||||
#if $hasattr($curNewznabProvider, 'enable_backlog') and $curNewznabProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curNewznabProvider.getID()}_enable_backlog">
|
||||
<label for="${curNewznabProvider.get_id()}_enable_backlog">
|
||||
<span class="component-title">Enable backlog searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNewznabProvider.getID()}_enable_backlog" id="${curNewznabProvider.getID()}_enable_backlog" <%= html_checked if curNewznabProvider.enable_backlog else '' %>/>
|
||||
<input type="checkbox" name="${curNewznabProvider.get_id()}_enable_backlog" id="${curNewznabProvider.get_id()}_enable_backlog" <%= html_checked if curNewznabProvider.enable_backlog else '' %>/>
|
||||
<p>perform backlog searches at provider</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNewznabProvider, 'search_mode'):
|
||||
#if $hasattr($curNewznabProvider, 'search_mode') and $curNewznabProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<span class="component-title">Season search mode</span>
|
||||
<span class="component-desc">
|
||||
<label class="space-right">
|
||||
<input type="radio" name="${curNewznabProvider.getID()}_search_mode" id="${curNewznabProvider.getID()}_search_mode_sponly" value="sponly" <%= html_checked if 'sponly' == curNewznabProvider.search_mode else '' %>/>season packs only
|
||||
<input type="radio" name="${curNewznabProvider.get_id()}_search_mode" id="${curNewznabProvider.get_id()}_search_mode_sponly" value="sponly" <%= html_checked if 'sponly' == curNewznabProvider.search_mode else '' %>/>season packs only
|
||||
</label>
|
||||
<label>
|
||||
<input type="radio" name="${curNewznabProvider.getID()}_search_mode" id="${curNewznabProvider.getID()}_search_mode_eponly" value="eponly" <%= html_checked if 'eponly' == curNewznabProvider.search_mode else '' %>/>episodes only
|
||||
<input type="radio" name="${curNewznabProvider.get_id()}_search_mode" id="${curNewznabProvider.get_id()}_search_mode_eponly" value="eponly" <%= html_checked if 'eponly' == curNewznabProvider.search_mode else '' %>/>episodes only
|
||||
</label>
|
||||
<p>when searching for complete seasons, search for packs or collect single episodes</p>
|
||||
</span>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNewznabProvider, 'search_fallback'):
|
||||
#if $hasattr($curNewznabProvider, 'search_fallback') and $curNewznabProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curNewznabProvider.getID()}_search_fallback">
|
||||
<label for="${curNewznabProvider.get_id()}_search_fallback">
|
||||
<span class="component-title">Season search fallback</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNewznabProvider.getID()}_search_fallback" id="${curNewznabProvider.getID()}_search_fallback" <%= html_checked if curNewznabProvider.search_fallback else '' %>/>
|
||||
<input type="checkbox" name="${curNewznabProvider.get_id()}_search_fallback" id="${curNewznabProvider.get_id()}_search_fallback" <%= html_checked if curNewznabProvider.search_fallback else '' %>/>
|
||||
<p>run the alternate season search mode when a complete season is not found</p>
|
||||
</span>
|
||||
</label>
|
||||
|
@ -264,218 +259,205 @@
|
|||
#end if
|
||||
</div>
|
||||
#end for
|
||||
##
|
||||
|
||||
##
|
||||
#for $curNzbProvider in [$curProvider for $curProvider in $sickbeard.providers.sortedProviderList() if $curProvider.providerType == $GenericProvider.NZB and $curProvider not in $sickbeard.newznabProviderList]:
|
||||
<div class="providerDiv" id="${curNzbProvider.getID()}Div">
|
||||
<div class="providerDiv" id="${curNzbProvider.get_id()}Div">
|
||||
#if $hasattr($curNzbProvider, 'username'):
|
||||
<div class="field-pair">
|
||||
<label for="${curNzbProvider.getID()}_username">
|
||||
<label for="${curNzbProvider.get_id()}_username">
|
||||
<span class="component-title">Username</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="${curNzbProvider.getID()}_username" value="$curNzbProvider.username" class="form-control input-sm input350" />
|
||||
<input type="text" name="${curNzbProvider.get_id()}_username" value="$curNzbProvider.username" class="form-control input-sm input350" />
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNzbProvider, 'api_key'):
|
||||
<div class="field-pair">
|
||||
<label for="${curNzbProvider.getID()}_api_key">
|
||||
<label for="${curNzbProvider.get_id()}_api_key">
|
||||
<span class="component-title">API key</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="${curNzbProvider.getID()}_api_key" value="<%= starify(curNzbProvider.api_key) %>" class="form-control input-sm input350" />
|
||||
#set $field_name = curNzbProvider.get_id() + '_api_key'
|
||||
<input type="text" name="$field_name" value="<%= starify(curNzbProvider.api_key) %>" class="form-control input-sm input350" />
|
||||
#if callable(getattr(curNzbProvider, 'ui_string'))
|
||||
<div class="clear-left"><p>${curNzbProvider.ui_string($field_name)}</p></div>
|
||||
#end if
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
|
||||
#if $hasattr($curNzbProvider, 'enable_recentsearch'):
|
||||
#if $hasattr($curNzbProvider, 'enable_recentsearch') and $curNzbProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curNzbProvider.getID()}_enable_recentsearch">
|
||||
<label for="${curNzbProvider.get_id()}_enable_recentsearch">
|
||||
<span class="component-title">Enable recent searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNzbProvider.getID()}_enable_recentsearch" id="${curNzbProvider.getID()}_enable_recentsearch" <%= html_checked if curNzbProvider.enable_recentsearch else '' %>/>
|
||||
<input type="checkbox" name="${curNzbProvider.get_id()}_enable_recentsearch" id="${curNzbProvider.get_id()}_enable_recentsearch" <%= html_checked if curNzbProvider.enable_recentsearch else '' %>/>
|
||||
<p>enable provider to perform recent searches.</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNzbProvider, 'enable_backlog'):
|
||||
#if $hasattr($curNzbProvider, 'enable_backlog') and $curNzbProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curNzbProvider.getID()}_enable_backlog">
|
||||
<label for="${curNzbProvider.get_id()}_enable_backlog">
|
||||
<span class="component-title">Enable backlog searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNzbProvider.getID()}_enable_backlog" id="${curNzbProvider.getID()}_enable_backlog" <%= html_checked if curNzbProvider.enable_backlog else '' %>/>
|
||||
<input type="checkbox" name="${curNzbProvider.get_id()}_enable_backlog" id="${curNzbProvider.get_id()}_enable_backlog" <%= html_checked if curNzbProvider.enable_backlog else '' %>/>
|
||||
<p>enable provider to perform backlog searches.</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNzbProvider, 'search_fallback'):
|
||||
#if $hasattr($curNzbProvider, 'search_mode') and $curNzbProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curNzbProvider.getID()}_search_fallback">
|
||||
<span class="component-title">Season search mode</span>
|
||||
<span class="component-desc">
|
||||
<label class="space-right">
|
||||
<input type="radio" name="${curNzbProvider.get_id()}_search_mode" id="${curNzbProvider.get_id()}_search_mode_sponly" value="sponly" <%= html_checked if 'sponly' == curNzbProvider.search_mode else '' %>/>season packs only
|
||||
</label>
|
||||
<label>
|
||||
<input type="radio" name="${curNzbProvider.get_id()}_search_mode" id="${curNzbProvider.get_id()}_search_mode_eponly" value="eponly" <%= html_checked if 'eponly' == curNzbProvider.search_mode else '' %>/>episodes only
|
||||
</label>
|
||||
<p>when searching for complete seasons, search for packs or collect single episodes</p>
|
||||
</span>
|
||||
</div>
|
||||
#end if
|
||||
#if $hasattr($curNzbProvider, 'search_fallback') and $curNzbProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curNzbProvider.get_id()}_search_fallback">
|
||||
<span class="component-title">Season search fallback</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNzbProvider.getID()}_search_fallback" id="${curNzbProvider.getID()}_search_fallback" <%= html_checked if curNzbProvider.search_fallback else '' %>/>
|
||||
<p>when searching for a complete season depending on search mode you may return no results, this helps by restarting the search using the opposite search mode.</p>
|
||||
<input type="checkbox" name="${curNzbProvider.get_id()}_search_fallback" id="${curNzbProvider.get_id()}_search_fallback" <%= html_checked if curNzbProvider.search_fallback else '' %>/>
|
||||
<p>run the alternate season search mode when a complete season is not found</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNzbProvider, 'search_mode'):
|
||||
#if not $curNzbProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label>
|
||||
<span class="component-title">Season search mode</span>
|
||||
<span class="component-desc">
|
||||
<p>when searching for complete seasons you can choose to have it look for season packs only, or choose to have it build a complete season from just single episodes.</p>
|
||||
</span>
|
||||
</label>
|
||||
<label>
|
||||
<span class="component-title"></span>
|
||||
<span class="component-desc">
|
||||
<input type="radio" name="${curNzbProvider.getID()}_search_mode" id="${curNzbProvider.getID()}_search_mode_sponly" value="sponly" <%= html_checked if 'sponly' == curNzbProvider.search_mode else '' %>/>season packs only.
|
||||
</span>
|
||||
</label>
|
||||
<label>
|
||||
<span class="component-title"></span>
|
||||
<span class="component-desc">
|
||||
<input type="radio" name="${curNzbProvider.getID()}_search_mode" id="${curNzbProvider.getID()}_search_mode_eponly" value="eponly" <%= html_checked if 'eponly' == curNzbProvider.search_mode else '' %>/>episodes only.
|
||||
</span>
|
||||
</label>
|
||||
<span class="component-desc">The latest releases are the focus of this provider, no backlog searching</span>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
</div>
|
||||
#end for
|
||||
##
|
||||
|
||||
##
|
||||
#for $curTorrentProvider in [$curProvider for $curProvider in $sickbeard.providers.sortedProviderList() if $curProvider.providerType == $GenericProvider.TORRENT]:
|
||||
<div class="providerDiv" id="${curTorrentProvider.getID()}Div">
|
||||
<div class="providerDiv" id="${curTorrentProvider.get_id()}Div">
|
||||
#if $hasattr($curTorrentProvider, 'api_key'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_api_key">
|
||||
<label for="${curTorrentProvider.get_id()}_api_key">
|
||||
<span class="component-title">Api key:</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="${curTorrentProvider.getID()}_api_key" id="${curTorrentProvider.getID()}_api_key" value="<%= starify(curTorrentProvider.api_key) %>" class="form-control input-sm input350" />
|
||||
<input type="text" name="${curTorrentProvider.get_id()}_api_key" id="${curTorrentProvider.get_id()}_api_key" value="<%= starify(curTorrentProvider.api_key) %>" class="form-control input-sm input350" />
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'digest'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_digest">
|
||||
<label for="${curTorrentProvider.get_id()}_digest">
|
||||
<span class="component-title">Digest:</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="${curTorrentProvider.getID()}_digest" id="${curTorrentProvider.getID()}_digest" value="$curTorrentProvider.digest" class="form-control input-sm input350" />
|
||||
<input type="text" name="${curTorrentProvider.get_id()}_digest" id="${curTorrentProvider.get_id()}_digest" value="$curTorrentProvider.digest" class="form-control input-sm input350" />
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'hash'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_hash">
|
||||
<label for="${curTorrentProvider.get_id()}_hash">
|
||||
<span class="component-title">Hash:</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="${curTorrentProvider.getID()}_hash" id="${curTorrentProvider.getID()}_hash" value="$curTorrentProvider.hash" class="form-control input-sm input350" />
|
||||
<input type="text" name="${curTorrentProvider.get_id()}_hash" id="${curTorrentProvider.get_id()}_hash" value="$curTorrentProvider.hash" class="form-control input-sm input350" />
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'username'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_username">
|
||||
<label for="${curTorrentProvider.get_id()}_username">
|
||||
<span class="component-title">Username:</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="${curTorrentProvider.getID()}_username" id="${curTorrentProvider.getID()}_username" value="$curTorrentProvider.username" class="form-control input-sm input350" />
|
||||
<input type="text" name="${curTorrentProvider.get_id()}_username" id="${curTorrentProvider.get_id()}_username" value="$curTorrentProvider.username" class="form-control input-sm input350" />
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'password'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_password">
|
||||
<label for="${curTorrentProvider.get_id()}_password">
|
||||
<span class="component-title">Password:</span>
|
||||
<span class="component-desc">
|
||||
<input type="password" name="${curTorrentProvider.getID()}_password" id="${curTorrentProvider.getID()}_password" value="#echo '*' * len($curTorrentProvider.password)#" class="form-control input-sm input350" />
|
||||
<input type="password" name="${curTorrentProvider.get_id()}_password" id="${curTorrentProvider.get_id()}_password" value="#echo '*' * len($curTorrentProvider.password)#" class="form-control input-sm input350" />
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'passkey'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_passkey">
|
||||
<label for="${curTorrentProvider.get_id()}_passkey">
|
||||
<span class="component-title">Passkey:</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="${curTorrentProvider.getID()}_passkey" id="${curTorrentProvider.getID()}_passkey" value="<%= starify(curTorrentProvider.passkey) %>" class="form-control input-sm input350" />
|
||||
<input type="text" name="${curTorrentProvider.get_id()}_passkey" id="${curTorrentProvider.get_id()}_passkey" value="<%= starify(curTorrentProvider.passkey) %>" class="form-control input-sm input350" />
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'ratio'):
|
||||
#if $hasattr($curTorrentProvider, '_seed_ratio') and 'blackhole' != $sickbeard.TORRENT_METHOD:
|
||||
#set $torrent_method_text = {'blackhole': 'Black hole', 'utorrent': 'uTorrent', 'transmission': 'Transmission', 'deluge': 'Deluge', 'download_station': 'Synology DS', 'rtorrent': 'rTorrent'}
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_ratio">
|
||||
<span class="component-title" id="${curTorrentProvider.getID()}_ratio_desc">Seed ratio:</span>
|
||||
<label for="${curTorrentProvider.get_id()}_ratio">
|
||||
<span class="component-title" id="${curTorrentProvider.get_id()}_ratio_desc">Seed until ratio (the goal)</span>
|
||||
<span class="component-desc">
|
||||
<input type="number" step="0.1" name="${curTorrentProvider.getID()}_ratio" id="${curTorrentProvider.getID()}_ratio" value="$curTorrentProvider.ratio" class="form-control input-sm input75" />
|
||||
</span>
|
||||
</label>
|
||||
<label>
|
||||
<span class="component-title"> </span>
|
||||
<span class="component-desc">
|
||||
<p>stop transfer when ratio is reached<br>(-1 SickGear default to seed forever, or leave blank for downloader default)</p>
|
||||
<input type="number" step="0.1" name="${curTorrentProvider.get_id()}_ratio" id="${curTorrentProvider.get_id()}_ratio" value="$curTorrentProvider._seed_ratio" class="form-control input-sm input75" />
|
||||
<p>this ratio is requested of each download sent to $torrent_method_text[$sickbeard.TORRENT_METHOD]</p>
|
||||
<div class="clear-left"><p>(set -1 to seed forever, or leave blank for the $torrent_method_text[$sickbeard.TORRENT_METHOD] default)</p></div>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'minseed'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_minseed">
|
||||
<span class="component-title" id="${curTorrentProvider.getID()}_minseed_desc">Minimum seeders:</span>
|
||||
<label for="${curTorrentProvider.get_id()}_minseed">
|
||||
<span class="component-title" id="${curTorrentProvider.get_id()}_minseed_desc">Minimum seeders</span>
|
||||
<span class="component-desc">
|
||||
<input type="number" name="${curTorrentProvider.getID()}_minseed" id="${curTorrentProvider.getID()}_minseed" value="$curTorrentProvider.minseed" class="form-control input-sm input75" />
|
||||
<input type="number" name="${curTorrentProvider.get_id()}_minseed" id="${curTorrentProvider.get_id()}_minseed" value="$curTorrentProvider.minseed" class="form-control input-sm input75" />
|
||||
<p>a release must have to be snatch worthy</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'minleech'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_minleech">
|
||||
<span class="component-title" id="${curTorrentProvider.getID()}_minleech_desc">Minimum leechers:</span>
|
||||
<label for="${curTorrentProvider.get_id()}_minleech">
|
||||
<span class="component-title" id="${curTorrentProvider.get_id()}_minleech_desc">Minimum leechers</span>
|
||||
<span class="component-desc">
|
||||
<input type="number" name="${curTorrentProvider.getID()}_minleech" id="${curTorrentProvider.getID()}_minleech" value="$curTorrentProvider.minleech" class="form-control input-sm input75" />
|
||||
<input type="number" name="${curTorrentProvider.get_id()}_minleech" id="${curTorrentProvider.get_id()}_minleech" value="$curTorrentProvider.minleech" class="form-control input-sm input75" />
|
||||
<p>a release must have to be snatch worthy</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'proxy'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_proxy">
|
||||
<label for="${curTorrentProvider.get_id()}_proxy">
|
||||
<span class="component-title">Access provider via proxy</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" class="enabler" name="${curTorrentProvider.getID()}_proxy" id="${curTorrentProvider.getID()}_proxy" <%= html_checked if curTorrentProvider.proxy.enabled else '' %>/>
|
||||
<input type="checkbox" class="enabler" name="${curTorrentProvider.get_id()}_proxy" id="${curTorrentProvider.get_id()}_proxy" <%= html_checked if curTorrentProvider.proxy.enabled else '' %>/>
|
||||
<p>to bypass country blocking mechanisms</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
||||
#if $hasattr($curTorrentProvider.proxy, 'url'):
|
||||
<div class="field-pair content_${curTorrentProvider.getID()}_proxy" id="content_${curTorrentProvider.getID()}_proxy">
|
||||
<label for="${curTorrentProvider.getID()}_proxy_url">
|
||||
<div class="field-pair content_${curTorrentProvider.get_id()}_proxy" id="content_${curTorrentProvider.get_id()}_proxy">
|
||||
<label for="${curTorrentProvider.get_id()}_proxy_url">
|
||||
<span class="component-title">Proxy URL:</span>
|
||||
<span class="component-desc">
|
||||
<select name="${curTorrentProvider.getID()}_proxy_url" id="${curTorrentProvider.getID()}_proxy_url" class="form-control input-sm">
|
||||
<select name="${curTorrentProvider.get_id()}_proxy_url" id="${curTorrentProvider.get_id()}_proxy_url" class="form-control input-sm">
|
||||
#for $i in $curTorrentProvider.proxy.urls.keys():
|
||||
<option value="$curTorrentProvider.proxy.urls[$i]" <%= html_selected if curTorrentProvider.proxy.url == curTorrentProvider.proxy.urls[i] else '' %>>$i</option>
|
||||
#end for
|
||||
|
@ -485,85 +467,71 @@
|
|||
</div>
|
||||
#end if
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'confirmed'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_confirmed">
|
||||
<label for="${curTorrentProvider.get_id()}_confirmed">
|
||||
<span class="component-title">Confirmed download</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curTorrentProvider.getID()}_confirmed" id="${curTorrentProvider.getID()}_confirmed" <%= html_checked if curTorrentProvider.confirmed else '' %>/>
|
||||
<input type="checkbox" name="${curTorrentProvider.get_id()}_confirmed" id="${curTorrentProvider.get_id()}_confirmed" <%= html_checked if curTorrentProvider.confirmed else '' %>/>
|
||||
<p>only download torrents from trusted or verified uploaders ?</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'freeleech'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_freeleech">
|
||||
<label for="${curTorrentProvider.get_id()}_freeleech">
|
||||
<span class="component-title">Freeleech</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curTorrentProvider.getID()}_freeleech" id="${curTorrentProvider.getID()}_freeleech" <%= html_checked if curTorrentProvider.freeleech else '' %>/>
|
||||
<input type="checkbox" name="${curTorrentProvider.get_id()}_freeleech" id="${curTorrentProvider.get_id()}_freeleech" <%= html_checked if curTorrentProvider.freeleech else '' %>/>
|
||||
<p>only download <b>[FreeLeech]</b> torrents.</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'enable_recentsearch'):
|
||||
#if $hasattr($curTorrentProvider, 'enable_recentsearch') and $curTorrentProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_enable_recentsearch">
|
||||
<label for="${curTorrentProvider.get_id()}_enable_recentsearch">
|
||||
<span class="component-title">Enable recent searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curTorrentProvider.getID()}_enable_recentsearch" id="${curTorrentProvider.getID()}_enable_recentsearch" <%= html_checked if curTorrentProvider.enable_recentsearch else '' %>/>
|
||||
<input type="checkbox" name="${curTorrentProvider.get_id()}_enable_recentsearch" id="${curTorrentProvider.get_id()}_enable_recentsearch" <%= html_checked if curTorrentProvider.enable_recentsearch else '' %>/>
|
||||
<p>enable provider to perform recent searches.</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'enable_backlog'):
|
||||
#if $hasattr($curTorrentProvider, 'enable_backlog') and $curTorrentProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_enable_backlog">
|
||||
<label for="${curTorrentProvider.get_id()}_enable_backlog">
|
||||
<span class="component-title">Enable backlog searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curTorrentProvider.getID()}_enable_backlog" id="${curTorrentProvider.getID()}_enable_backlog" <%= html_checked if curTorrentProvider.enable_backlog else '' %>/>
|
||||
<input type="checkbox" name="${curTorrentProvider.get_id()}_enable_backlog" id="${curTorrentProvider.get_id()}_enable_backlog" <%= html_checked if curTorrentProvider.enable_backlog else '' %>/>
|
||||
<p>enable provider to perform backlog searches.</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'search_fallback'):
|
||||
#if $hasattr($curTorrentProvider, 'search_mode') and $curTorrentProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_search_fallback">
|
||||
<span class="component-title">Season search fallback</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curTorrentProvider.getID()}_search_fallback" id="${curTorrentProvider.getID()}_search_fallback" <%= html_checked if curTorrentProvider.search_fallback else '' %>/>
|
||||
<p>when searching for a complete season depending on search mode you may return no results, this helps by restarting the search using the opposite search mode.</p>
|
||||
</span>
|
||||
</label>
|
||||
<span class="component-title">Season search mode</span>
|
||||
<span class="component-desc">
|
||||
<label class="space-right">
|
||||
<input type="radio" name="${curTorrentProvider.get_id()}_search_mode" id="${curTorrentProvider.get_id()}_search_mode_sponly" value="sponly" <%= html_checked if 'sponly' == curTorrentProvider.search_mode else '' %>/>season packs only
|
||||
</label>
|
||||
<label>
|
||||
<input type="radio" name="${curTorrentProvider.get_id()}_search_mode" id="${curTorrentProvider.get_id()}_search_mode_eponly" value="eponly" <%= html_checked if 'eponly' == curTorrentProvider.search_mode else '' %>/>episodes only
|
||||
</label>
|
||||
<p>when searching for complete seasons, search for packs or collect single episodes</p>
|
||||
</span>
|
||||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'search_mode'):
|
||||
#if $hasattr($curTorrentProvider, 'search_fallback') and $curTorrentProvider.supportsBacklog:
|
||||
<div class="field-pair">
|
||||
<label>
|
||||
<span class="component-title">Season search mode</span>
|
||||
<label for="${curTorrentProvider.get_id()}_search_fallback">
|
||||
<span class="component-title">Season search fallback</span>
|
||||
<span class="component-desc">
|
||||
<p>when searching for complete seasons you can choose to have it look for season packs only, or choose to have it build a complete season from just single episodes.</p>
|
||||
</span>
|
||||
</label>
|
||||
<label>
|
||||
<span class="component-title"></span>
|
||||
<span class="component-desc">
|
||||
<input type="radio" name="${curTorrentProvider.getID()}_search_mode" id="${curTorrentProvider.getID()}_search_mode_sponly" value="sponly" <%= html_checked if 'sponly' == curTorrentProvider.search_mode else '' %>/>season packs only.
|
||||
</span>
|
||||
</label>
|
||||
<label>
|
||||
<span class="component-title"></span>
|
||||
<span class="component-desc">
|
||||
<input type="radio" name="${curTorrentProvider.getID()}_search_mode" id="${curTorrentProvider.getID()}_search_mode_eponly" value="eponly" <%= html_checked if 'eponly' == curTorrentProvider.search_mode else '' %>/>episodes only.
|
||||
<input type="checkbox" name="${curTorrentProvider.get_id()}_search_fallback" id="${curTorrentProvider.get_id()}_search_fallback" <%= html_checked if curTorrentProvider.search_fallback else '' %>/>
|
||||
<p>run the alternate season search mode when a complete season is not found</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
|
|
@ -139,9 +139,9 @@
|
|||
#else
|
||||
#if 0 < $hItem['provider']
|
||||
#if $curStatus in [SNATCHED, FAILED]
|
||||
#set $provider = $providers.getProviderClass($generic.GenericProvider.makeID($hItem['provider']))
|
||||
#set $provider = $providers.getProviderClass($generic.GenericProvider.make_id($hItem['provider']))
|
||||
#if None is not $provider
|
||||
<img src="$sbRoot/images/providers/<%= provider.imageName() %>" width="16" height="16" /><span>$provider.name</span>
|
||||
<img src="$sbRoot/images/providers/<%= provider.image_name() %>" width="16" height="16" /><span>$provider.name</span>
|
||||
#else
|
||||
<img src="$sbRoot/images/providers/missing.png" width="16" height="16" title="missing provider" /><span>Missing Provider</span>
|
||||
#end if
|
||||
|
@ -186,10 +186,10 @@
|
|||
#set $curStatus, $curQuality = $Quality.splitCompositeStatus(int($action['action']))
|
||||
#set $basename = $os.path.basename($action['resource'])
|
||||
#if $curStatus in [SNATCHED, FAILED]
|
||||
#set $provider = $providers.getProviderClass($generic.GenericProvider.makeID($action['provider']))
|
||||
#set $provider = $providers.getProviderClass($generic.GenericProvider.make_id($action['provider']))
|
||||
#if None is not $provider
|
||||
#set $prov_list += ['<span%s><img class="help" src="%s/images/providers/%s" width="16" height="16" alt="%s" title="%s.. %s: %s" /></span>'\
|
||||
% (('', ' class="fail"')[FAILED == $curStatus], $sbRoot, $provider.imageName(), $provider.name,
|
||||
% (('', ' class="fail"')[FAILED == $curStatus], $sbRoot, $provider.image_name(), $provider.name,
|
||||
('%s%s' % ($order, 'th' if $order in [11, 12, 13] or str($order)[-1] not in $ordinal_indicators else $ordinal_indicators[str($order)[-1]]), 'Snatch failed')[FAILED == $curStatus],
|
||||
$provider.name, $basename)]
|
||||
#set $order += (0, 1)[SNATCHED == $curStatus]
|
||||
|
|
|
@ -72,9 +72,9 @@
|
|||
<tr>
|
||||
<td class="text-nowrap text-left">#echo re.sub('"', '', $hItem['release'])#</td>
|
||||
<td>#echo ($hItem['size'], '?')[-1 == $hItem['size']]#</td>
|
||||
#set $provider = $providers.getProviderClass($generic.GenericProvider.makeID($hItem['provider']))
|
||||
#set $provider = $providers.getProviderClass($generic.GenericProvider.make_id($hItem['provider']))
|
||||
#if None is not $provider:
|
||||
<td><img src="$sbRoot/images/providers/<%= provider.imageName() %>" width="16" height="16" alt="$provider.name" title="$provider.name" /></td>
|
||||
<td><img src="$sbRoot/images/providers/<%= provider.image_name() %>" width="16" height="16" alt="$provider.name" title="$provider.name" /></td>
|
||||
#else
|
||||
<td><img src="$sbRoot/images/providers/missing.png" width="16" height="16" alt="missing provider" title="missing provider" /></td>
|
||||
#end if
|
||||
|
|
|
@ -31,7 +31,6 @@ import sys
|
|||
import os.path
|
||||
import uuid
|
||||
import base64
|
||||
import sickbeard
|
||||
sys.path.insert(1, os.path.abspath('../lib'))
|
||||
from sickbeard import providers, metadata, config, webserveInit
|
||||
from sickbeard.providers.generic import GenericProvider
|
||||
|
@ -255,10 +254,6 @@ OMGWTFNZBS = False
|
|||
OMGWTFNZBS_USERNAME = None
|
||||
OMGWTFNZBS_APIKEY = None
|
||||
|
||||
NEWZBIN = False
|
||||
NEWZBIN_USERNAME = None
|
||||
NEWZBIN_PASSWORD = None
|
||||
|
||||
SAB_USERNAME = None
|
||||
SAB_PASSWORD = None
|
||||
SAB_APIKEY = None
|
||||
|
@ -485,10 +480,12 @@ COOKIE_SECRET = base64.b64encode(uuid.uuid4().bytes + uuid.uuid4().bytes)
|
|||
|
||||
__INITIALIZED__ = False
|
||||
|
||||
|
||||
def get_backlog_cycle_time():
|
||||
cycletime = RECENTSEARCH_FREQUENCY * 2 + 7
|
||||
return max([cycletime, 720])
|
||||
|
||||
|
||||
def initialize(consoleLogging=True):
|
||||
with INIT_LOCK:
|
||||
|
||||
|
@ -525,7 +522,7 @@ def initialize(consoleLogging=True):
|
|||
USE_SYNOLOGYNOTIFIER, SYNOLOGYNOTIFIER_NOTIFY_ONSNATCH, SYNOLOGYNOTIFIER_NOTIFY_ONDOWNLOAD, SYNOLOGYNOTIFIER_NOTIFY_ONSUBTITLEDOWNLOAD, \
|
||||
USE_EMAIL, EMAIL_HOST, EMAIL_PORT, EMAIL_TLS, EMAIL_USER, EMAIL_PASSWORD, EMAIL_FROM, EMAIL_NOTIFY_ONSNATCH, EMAIL_NOTIFY_ONDOWNLOAD, EMAIL_NOTIFY_ONSUBTITLEDOWNLOAD, EMAIL_LIST, \
|
||||
USE_LISTVIEW, METADATA_XBMC, METADATA_XBMC_12PLUS, METADATA_MEDIABROWSER, METADATA_PS3, METADATA_KODI, metadata_provider_dict, \
|
||||
NEWZBIN, NEWZBIN_USERNAME, NEWZBIN_PASSWORD, GIT_PATH, MOVE_ASSOCIATED_FILES, POSTPONE_IF_SYNC_FILES, recentSearchScheduler, NFO_RENAME, \
|
||||
GIT_PATH, MOVE_ASSOCIATED_FILES, POSTPONE_IF_SYNC_FILES, recentSearchScheduler, NFO_RENAME, \
|
||||
GUI_NAME, DEFAULT_HOME, HOME_LAYOUT, HISTORY_LAYOUT, DISPLAY_SHOW_SPECIALS, EPISODE_VIEW_LAYOUT, EPISODE_VIEW_SORT, EPISODE_VIEW_DISPLAY_PAUSED, EPISODE_VIEW_MISSED_RANGE, FUZZY_DATING, TRIM_ZERO, DATE_PRESET, TIME_PRESET, TIME_PRESET_W_SECONDS, THEME_NAME, \
|
||||
POSTER_SORTBY, POSTER_SORTDIR, \
|
||||
METADATA_WDTV, METADATA_TIVO, METADATA_MEDE8ER, IGNORE_WORDS, REQUIRE_WORDS, CALENDAR_UNPROTECTED, CREATE_MISSING_SHOW_DIRS, \
|
||||
|
@ -542,7 +539,6 @@ def initialize(consoleLogging=True):
|
|||
|
||||
CheckSection(CFG, 'General')
|
||||
CheckSection(CFG, 'Blackhole')
|
||||
CheckSection(CFG, 'Newzbin')
|
||||
CheckSection(CFG, 'SABnzbd')
|
||||
CheckSection(CFG, 'NZBget')
|
||||
CheckSection(CFG, 'XBMC')
|
||||
|
@ -587,7 +583,7 @@ def initialize(consoleLogging=True):
|
|||
CACHE_DIR = ACTUAL_CACHE_DIR
|
||||
|
||||
if not helpers.makeDir(CACHE_DIR):
|
||||
logger.log(u"!!! Creating local cache dir failed, using system default", logger.ERROR)
|
||||
logger.log(u'!!! Creating local cache dir failed, using system default', logger.ERROR)
|
||||
CACHE_DIR = None
|
||||
|
||||
# clean cache folders
|
||||
|
@ -604,7 +600,7 @@ def initialize(consoleLogging=True):
|
|||
TRIM_ZERO = bool(check_setting_int(CFG, 'GUI', 'trim_zero', 0))
|
||||
DATE_PRESET = check_setting_str(CFG, 'GUI', 'date_preset', '%x')
|
||||
TIME_PRESET_W_SECONDS = check_setting_str(CFG, 'GUI', 'time_preset', '%I:%M:%S %p')
|
||||
TIME_PRESET = TIME_PRESET_W_SECONDS.replace(u":%S", u"")
|
||||
TIME_PRESET = TIME_PRESET_W_SECONDS.replace(u':%S', u'')
|
||||
TIMEZONE_DISPLAY = check_setting_str(CFG, 'GUI', 'timezone_display', 'network')
|
||||
DISPLAY_BACKGROUND = bool(check_setting_int(CFG, 'General', 'display_background', 0))
|
||||
DISPLAY_BACKGROUND_TRANSPARENT = check_setting_str(CFG, 'General', 'display_background_transparent', 'transparent')
|
||||
|
@ -618,7 +614,7 @@ def initialize(consoleLogging=True):
|
|||
LOG_DIR = os.path.normpath(os.path.join(DATA_DIR, ACTUAL_LOG_DIR))
|
||||
|
||||
if not helpers.makeDir(LOG_DIR):
|
||||
logger.log(u"!!! No log folder, logging to screen only!", logger.ERROR)
|
||||
logger.log(u'!!! No log folder, logging to screen only!', logger.ERROR)
|
||||
|
||||
FILE_LOGGING_PRESET = check_setting_str(CFG, 'General', 'file_logging_preset', 'DB')
|
||||
|
||||
|
@ -635,7 +631,7 @@ def initialize(consoleLogging=True):
|
|||
|
||||
WEB_HOST = check_setting_str(CFG, 'General', 'web_host', '0.0.0.0')
|
||||
WEB_IPV6 = bool(check_setting_int(CFG, 'General', 'web_ipv6', 0))
|
||||
WEB_ROOT = check_setting_str(CFG, 'General', 'web_root', '').rstrip("/")
|
||||
WEB_ROOT = check_setting_str(CFG, 'General', 'web_root', '').rstrip('/')
|
||||
WEB_LOG = bool(check_setting_int(CFG, 'General', 'web_log', 0))
|
||||
ENCRYPTION_VERSION = check_setting_int(CFG, 'General', 'encryption_version', 0)
|
||||
WEB_USERNAME = check_setting_str(CFG, 'General', 'web_username', '')
|
||||
|
@ -767,10 +763,6 @@ def initialize(consoleLogging=True):
|
|||
NZBS_UID = check_setting_str(CFG, 'NZBs', 'nzbs_uid', '')
|
||||
NZBS_HASH = check_setting_str(CFG, 'NZBs', 'nzbs_hash', '')
|
||||
|
||||
NEWZBIN = bool(check_setting_int(CFG, 'Newzbin', 'newzbin', 0))
|
||||
NEWZBIN_USERNAME = check_setting_str(CFG, 'Newzbin', 'newzbin_username', '')
|
||||
NEWZBIN_PASSWORD = check_setting_str(CFG, 'Newzbin', 'newzbin_password', '')
|
||||
|
||||
SAB_USERNAME = check_setting_str(CFG, 'SABnzbd', 'sab_username', '')
|
||||
SAB_PASSWORD = check_setting_str(CFG, 'SABnzbd', 'sab_password', '')
|
||||
SAB_APIKEY = check_setting_str(CFG, 'SABnzbd', 'sab_apikey', '')
|
||||
|
@ -840,7 +832,7 @@ def initialize(consoleLogging=True):
|
|||
PROWL_NOTIFY_ONDOWNLOAD = bool(check_setting_int(CFG, 'Prowl', 'prowl_notify_ondownload', 0))
|
||||
PROWL_NOTIFY_ONSUBTITLEDOWNLOAD = bool(check_setting_int(CFG, 'Prowl', 'prowl_notify_onsubtitledownload', 0))
|
||||
PROWL_API = check_setting_str(CFG, 'Prowl', 'prowl_api', '')
|
||||
PROWL_PRIORITY = check_setting_str(CFG, 'Prowl', 'prowl_priority', "0")
|
||||
PROWL_PRIORITY = check_setting_str(CFG, 'Prowl', 'prowl_priority', '0')
|
||||
|
||||
USE_TWITTER = bool(check_setting_int(CFG, 'Twitter', 'use_twitter', 0))
|
||||
TWITTER_NOTIFY_ONSNATCH = bool(check_setting_int(CFG, 'Twitter', 'twitter_notify_onsnatch', 0))
|
||||
|
@ -924,7 +916,7 @@ def initialize(consoleLogging=True):
|
|||
NMA_NOTIFY_ONDOWNLOAD = bool(check_setting_int(CFG, 'NMA', 'nma_notify_ondownload', 0))
|
||||
NMA_NOTIFY_ONSUBTITLEDOWNLOAD = bool(check_setting_int(CFG, 'NMA', 'nma_notify_onsubtitledownload', 0))
|
||||
NMA_API = check_setting_str(CFG, 'NMA', 'nma_api', '')
|
||||
NMA_PRIORITY = check_setting_str(CFG, 'NMA', 'nma_priority', "0")
|
||||
NMA_PRIORITY = check_setting_str(CFG, 'NMA', 'nma_priority', '0')
|
||||
|
||||
USE_PUSHALOT = bool(check_setting_int(CFG, 'Pushalot', 'use_pushalot', 0))
|
||||
PUSHALOT_NOTIFY_ONSNATCH = bool(check_setting_int(CFG, 'Pushalot', 'pushalot_notify_onsnatch', 0))
|
||||
|
@ -1019,119 +1011,91 @@ def initialize(consoleLogging=True):
|
|||
torrentRssProviderList = providers.getTorrentRssProviderList(TORRENTRSS_DATA)
|
||||
|
||||
# dynamically load provider settings
|
||||
for curTorrentProvider in [curProvider for curProvider in providers.sortedProviderList() if
|
||||
curProvider.providerType == GenericProvider.TORRENT]:
|
||||
curTorrentProvider.enabled = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID(), 0))
|
||||
if hasattr(curTorrentProvider, 'api_key'):
|
||||
curTorrentProvider.api_key = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_api_key', '')
|
||||
if hasattr(curTorrentProvider, 'hash'):
|
||||
curTorrentProvider.hash = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_hash', '')
|
||||
if hasattr(curTorrentProvider, 'digest'):
|
||||
curTorrentProvider.digest = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_digest', '')
|
||||
if hasattr(curTorrentProvider, 'username'):
|
||||
curTorrentProvider.username = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_username', '')
|
||||
if hasattr(curTorrentProvider, 'password'):
|
||||
curTorrentProvider.password = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_password', '')
|
||||
if hasattr(curTorrentProvider, 'passkey'):
|
||||
curTorrentProvider.passkey = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_passkey', '')
|
||||
if hasattr(curTorrentProvider, 'proxy'):
|
||||
curTorrentProvider.proxy.enabled = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_proxy', 0))
|
||||
if hasattr(curTorrentProvider.proxy, 'url'):
|
||||
curTorrentProvider.proxy.url = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_proxy_url', '')
|
||||
if hasattr(curTorrentProvider, 'confirmed'):
|
||||
curTorrentProvider.confirmed = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_confirmed', 0))
|
||||
if hasattr(curTorrentProvider, 'options'):
|
||||
curTorrentProvider.options = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_options', '')
|
||||
if hasattr(curTorrentProvider, 'ratio'):
|
||||
curTorrentProvider.ratio = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_ratio', '')
|
||||
if hasattr(curTorrentProvider, 'minseed'):
|
||||
curTorrentProvider.minseed = check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_minseed', 0)
|
||||
if hasattr(curTorrentProvider, 'minleech'):
|
||||
curTorrentProvider.minleech = check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_minleech', 0)
|
||||
if hasattr(curTorrentProvider, 'freeleech'):
|
||||
curTorrentProvider.freeleech = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_freeleech', 0))
|
||||
if hasattr(curTorrentProvider, 'search_mode'):
|
||||
curTorrentProvider.search_mode = check_setting_str(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_search_mode',
|
||||
'eponly')
|
||||
if hasattr(curTorrentProvider, 'search_fallback'):
|
||||
curTorrentProvider.search_fallback = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_search_fallback',
|
||||
0))
|
||||
for torrent_prov in [curProvider for curProvider in providers.sortedProviderList()
|
||||
if GenericProvider.TORRENT == curProvider.providerType]:
|
||||
prov_id = torrent_prov.get_id()
|
||||
prov_id_uc = torrent_prov.get_id().upper()
|
||||
torrent_prov.enabled = bool(check_setting_int(CFG, prov_id_uc, prov_id, 0))
|
||||
if hasattr(torrent_prov, 'api_key'):
|
||||
torrent_prov.api_key = check_setting_str(CFG, prov_id_uc, prov_id + '_api_key', '')
|
||||
if hasattr(torrent_prov, 'hash'):
|
||||
torrent_prov.hash = check_setting_str(CFG, prov_id_uc, prov_id + '_hash', '')
|
||||
if hasattr(torrent_prov, 'digest'):
|
||||
torrent_prov.digest = check_setting_str(CFG, prov_id_uc, prov_id + '_digest', '')
|
||||
if hasattr(torrent_prov, 'username'):
|
||||
torrent_prov.username = check_setting_str(CFG, prov_id_uc, prov_id + '_username', '')
|
||||
if hasattr(torrent_prov, 'password'):
|
||||
torrent_prov.password = check_setting_str(CFG, prov_id_uc, prov_id + '_password', '')
|
||||
if hasattr(torrent_prov, 'passkey'):
|
||||
torrent_prov.passkey = check_setting_str(CFG, prov_id_uc, prov_id + '_passkey', '')
|
||||
if hasattr(torrent_prov, 'proxy'):
|
||||
torrent_prov.proxy.enabled = bool(check_setting_int(CFG, prov_id_uc, prov_id + '_proxy', 0))
|
||||
if hasattr(torrent_prov.proxy, 'url'):
|
||||
torrent_prov.proxy.url = check_setting_str(CFG, prov_id_uc, prov_id + '_proxy_url', '')
|
||||
if hasattr(torrent_prov, 'confirmed'):
|
||||
torrent_prov.confirmed = bool(check_setting_int(CFG, prov_id_uc, prov_id + '_confirmed', 0))
|
||||
if hasattr(torrent_prov, 'options'):
|
||||
torrent_prov.options = check_setting_str(CFG, prov_id_uc, prov_id + '_options', '')
|
||||
if hasattr(torrent_prov, 'ratio'):
|
||||
torrent_prov.ratio = check_setting_str(CFG, prov_id_uc, prov_id + '_ratio', '')
|
||||
if hasattr(torrent_prov, 'minseed'):
|
||||
torrent_prov.minseed = check_setting_int(CFG, prov_id_uc, prov_id + '_minseed', 0)
|
||||
if hasattr(torrent_prov, 'minleech'):
|
||||
torrent_prov.minleech = check_setting_int(CFG, prov_id_uc, prov_id + '_minleech', 0)
|
||||
if hasattr(torrent_prov, 'freeleech'):
|
||||
torrent_prov.freeleech = bool(check_setting_int(CFG, prov_id_uc, prov_id + '_freeleech', 0))
|
||||
if hasattr(torrent_prov, 'search_mode'):
|
||||
torrent_prov.search_mode = check_setting_str(CFG, prov_id_uc, prov_id + '_search_mode', 'eponly')
|
||||
if hasattr(torrent_prov, 'search_fallback'):
|
||||
torrent_prov.search_fallback = bool(check_setting_int(CFG, prov_id_uc, prov_id + '_search_fallback', 0))
|
||||
if hasattr(torrent_prov, 'enable_recentsearch'):
|
||||
torrent_prov.enable_recentsearch = bool(check_setting_int(CFG, prov_id_uc,
|
||||
prov_id + '_enable_recentsearch', 1))
|
||||
if hasattr(torrent_prov, 'enable_backlog'):
|
||||
torrent_prov.enable_backlog = bool(check_setting_int(CFG, prov_id_uc, prov_id + '_enable_backlog', 1))
|
||||
|
||||
if hasattr(curTorrentProvider, 'enable_recentsearch'):
|
||||
curTorrentProvider.enable_recentsearch = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() +
|
||||
'_enable_recentsearch', 1))
|
||||
if hasattr(curTorrentProvider, 'enable_backlog'):
|
||||
curTorrentProvider.enable_backlog = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_enable_backlog',
|
||||
1))
|
||||
|
||||
for curNzbProvider in [curProvider for curProvider in providers.sortedProviderList() if
|
||||
curProvider.providerType == GenericProvider.NZB]:
|
||||
curNzbProvider.enabled = bool(
|
||||
check_setting_int(CFG, curNzbProvider.getID().upper(), curNzbProvider.getID(), 0))
|
||||
if hasattr(curNzbProvider, 'api_key'):
|
||||
curNzbProvider.api_key = check_setting_str(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_api_key', '')
|
||||
if hasattr(curNzbProvider, 'username'):
|
||||
curNzbProvider.username = check_setting_str(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_username', '')
|
||||
if hasattr(curNzbProvider, 'search_mode'):
|
||||
curNzbProvider.search_mode = check_setting_str(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_search_mode',
|
||||
'eponly')
|
||||
if hasattr(curNzbProvider, 'search_fallback'):
|
||||
curNzbProvider.search_fallback = bool(check_setting_int(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_search_fallback',
|
||||
0))
|
||||
if hasattr(curNzbProvider, 'enable_recentsearch'):
|
||||
curNzbProvider.enable_recentsearch = bool(check_setting_int(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_enable_recentsearch',
|
||||
1))
|
||||
if hasattr(curNzbProvider, 'enable_backlog'):
|
||||
curNzbProvider.enable_backlog = bool(check_setting_int(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_enable_backlog',
|
||||
1))
|
||||
for nzb_prov in [curProvider for curProvider in providers.sortedProviderList()
|
||||
if GenericProvider.NZB == curProvider.providerType]:
|
||||
prov_id = nzb_prov.get_id()
|
||||
prov_id_uc = nzb_prov.get_id().upper()
|
||||
nzb_prov.enabled = bool(
|
||||
check_setting_int(CFG, prov_id_uc, prov_id, 0))
|
||||
if hasattr(nzb_prov, 'api_key'):
|
||||
nzb_prov.api_key = check_setting_str(CFG, prov_id_uc, prov_id + '_api_key', '')
|
||||
if hasattr(nzb_prov, 'username'):
|
||||
nzb_prov.username = check_setting_str(CFG, prov_id_uc, prov_id + '_username', '')
|
||||
if hasattr(nzb_prov, 'search_mode'):
|
||||
nzb_prov.search_mode = check_setting_str(CFG, prov_id_uc, prov_id + '_search_mode', 'eponly')
|
||||
if hasattr(nzb_prov, 'search_fallback'):
|
||||
nzb_prov.search_fallback = bool(check_setting_int(CFG, prov_id_uc, prov_id + '_search_fallback', 0))
|
||||
if hasattr(nzb_prov, 'enable_recentsearch'):
|
||||
nzb_prov.enable_recentsearch = bool(check_setting_int(CFG, prov_id_uc,
|
||||
prov_id + '_enable_recentsearch', 1))
|
||||
if hasattr(nzb_prov, 'enable_backlog'):
|
||||
nzb_prov.enable_backlog = bool(check_setting_int(CFG, prov_id_uc, prov_id + '_enable_backlog', 1))
|
||||
|
||||
if not os.path.isfile(CONFIG_FILE):
|
||||
logger.log(u"Unable to find '" + CONFIG_FILE + "', all settings will be default!", logger.DEBUG)
|
||||
logger.log(u'Unable to find \'' + CONFIG_FILE + '\', all settings will be default!', logger.DEBUG)
|
||||
save_config()
|
||||
|
||||
# start up all the threads
|
||||
logger.sb_log_instance.initLogging(consoleLogging=consoleLogging)
|
||||
|
||||
# initialize the main SB database
|
||||
myDB = db.DBConnection()
|
||||
db.MigrationCode(myDB)
|
||||
my_db = db.DBConnection()
|
||||
db.MigrationCode(my_db)
|
||||
|
||||
# initialize the cache database
|
||||
myDB = db.DBConnection('cache.db')
|
||||
db.upgradeDatabase(myDB, cache_db.InitialSchema)
|
||||
my_db = db.DBConnection('cache.db')
|
||||
db.upgradeDatabase(my_db, cache_db.InitialSchema)
|
||||
|
||||
# initialize the failed downloads database
|
||||
myDB = db.DBConnection('failed.db')
|
||||
db.upgradeDatabase(myDB, failed_db.InitialSchema)
|
||||
my_db = db.DBConnection('failed.db')
|
||||
db.upgradeDatabase(my_db, failed_db.InitialSchema)
|
||||
|
||||
# fix up any db problems
|
||||
myDB = db.DBConnection()
|
||||
db.sanityCheckDatabase(myDB, mainDB.MainSanityCheck)
|
||||
my_db = db.DBConnection()
|
||||
db.sanityCheckDatabase(my_db, mainDB.MainSanityCheck)
|
||||
|
||||
# migrate the config if it needs it
|
||||
migrator = ConfigMigrator(CFG)
|
||||
|
@ -1147,7 +1111,7 @@ def initialize(consoleLogging=True):
|
|||
(METADATA_TIVO, metadata.tivo),
|
||||
(METADATA_MEDE8ER, metadata.mede8er),
|
||||
(METADATA_KODI, metadata.kodi),
|
||||
]:
|
||||
]:
|
||||
(cur_metadata_config, cur_metadata_class) = cur_metadata_tuple
|
||||
tmp_provider = cur_metadata_class.metadata_class()
|
||||
tmp_provider.set_config(cur_metadata_config)
|
||||
|
@ -1169,7 +1133,7 @@ def initialize(consoleLogging=True):
|
|||
cycleTime=datetime.timedelta(hours=1),
|
||||
threadName='SHOWUPDATER',
|
||||
start_time=datetime.time(hour=SHOW_UPDATE_HOUR),
|
||||
prevent_cycle_run=sickbeard.showQueueScheduler.action.isShowUpdateRunning) # 3 AM
|
||||
prevent_cycle_run=showQueueScheduler.action.isShowUpdateRunning) # 3 AM
|
||||
|
||||
# searchers
|
||||
searchQueueScheduler = scheduler.Scheduler(search_queue.SearchQueue(),
|
||||
|
@ -1178,18 +1142,18 @@ def initialize(consoleLogging=True):
|
|||
|
||||
update_interval = datetime.timedelta(minutes=RECENTSEARCH_FREQUENCY)
|
||||
recentSearchScheduler = scheduler.Scheduler(searchRecent.RecentSearcher(),
|
||||
cycleTime=update_interval,
|
||||
threadName='RECENTSEARCHER',
|
||||
run_delay=update_now if RECENTSEARCH_STARTUP
|
||||
else datetime.timedelta(minutes=5),
|
||||
prevent_cycle_run=sickbeard.searchQueueScheduler.action.is_recentsearch_in_progress)
|
||||
cycleTime=update_interval,
|
||||
threadName='RECENTSEARCHER',
|
||||
run_delay=update_now if RECENTSEARCH_STARTUP
|
||||
else datetime.timedelta(minutes=5),
|
||||
prevent_cycle_run=searchQueueScheduler.action.is_recentsearch_in_progress)
|
||||
|
||||
backlogSearchScheduler = searchBacklog.BacklogSearchScheduler(searchBacklog.BacklogSearcher(),
|
||||
cycleTime=datetime.timedelta(minutes=get_backlog_cycle_time()),
|
||||
threadName='BACKLOG',
|
||||
run_delay=update_now if BACKLOG_STARTUP
|
||||
else datetime.timedelta(minutes=10),
|
||||
prevent_cycle_run=sickbeard.searchQueueScheduler.action.is_standard_backlog_in_progress)
|
||||
prevent_cycle_run=searchQueueScheduler.action.is_standard_backlog_in_progress)
|
||||
|
||||
search_intervals = {'15m': 15, '45m': 45, '90m': 90, '4h': 4 * 60, 'daily': 24 * 60}
|
||||
if CHECK_PROPERS_INTERVAL in search_intervals:
|
||||
|
@ -1204,7 +1168,7 @@ def initialize(consoleLogging=True):
|
|||
threadName='FINDPROPERS',
|
||||
start_time=run_at,
|
||||
run_delay=update_interval,
|
||||
prevent_cycle_run=sickbeard.searchQueueScheduler.action.is_propersearch_in_progress)
|
||||
prevent_cycle_run=searchQueueScheduler.action.is_propersearch_in_progress)
|
||||
|
||||
# processors
|
||||
autoPostProcesserScheduler = scheduler.Scheduler(autoPostProcesser.PostProcesser(),
|
||||
|
@ -1290,52 +1254,52 @@ def halt():
|
|||
|
||||
if __INITIALIZED__:
|
||||
|
||||
logger.log(u"Aborting all threads")
|
||||
logger.log(u'Aborting all threads')
|
||||
|
||||
events.stop.set()
|
||||
logger.log(u"Waiting for the EVENTS thread to exit")
|
||||
logger.log(u'Waiting for the EVENTS thread to exit')
|
||||
try:
|
||||
events.join(10)
|
||||
except:
|
||||
pass
|
||||
|
||||
recentSearchScheduler.stop.set()
|
||||
logger.log(u"Waiting for the RECENTSEARCH thread to exit")
|
||||
logger.log(u'Waiting for the RECENTSEARCH thread to exit')
|
||||
try:
|
||||
recentSearchScheduler.join(10)
|
||||
except:
|
||||
pass
|
||||
|
||||
backlogSearchScheduler.stop.set()
|
||||
logger.log(u"Waiting for the BACKLOG thread to exit")
|
||||
logger.log(u'Waiting for the BACKLOG thread to exit')
|
||||
try:
|
||||
backlogSearchScheduler.join(10)
|
||||
except:
|
||||
pass
|
||||
|
||||
showUpdateScheduler.stop.set()
|
||||
logger.log(u"Waiting for the SHOWUPDATER thread to exit")
|
||||
logger.log(u'Waiting for the SHOWUPDATER thread to exit')
|
||||
try:
|
||||
showUpdateScheduler.join(10)
|
||||
except:
|
||||
pass
|
||||
|
||||
versionCheckScheduler.stop.set()
|
||||
logger.log(u"Waiting for the VERSIONCHECKER thread to exit")
|
||||
logger.log(u'Waiting for the VERSIONCHECKER thread to exit')
|
||||
try:
|
||||
versionCheckScheduler.join(10)
|
||||
except:
|
||||
pass
|
||||
|
||||
showQueueScheduler.stop.set()
|
||||
logger.log(u"Waiting for the SHOWQUEUE thread to exit")
|
||||
logger.log(u'Waiting for the SHOWQUEUE thread to exit')
|
||||
try:
|
||||
showQueueScheduler.join(10)
|
||||
except:
|
||||
pass
|
||||
|
||||
searchQueueScheduler.stop.set()
|
||||
logger.log(u"Waiting for the SEARCHQUEUE thread to exit")
|
||||
logger.log(u'Waiting for the SEARCHQUEUE thread to exit')
|
||||
try:
|
||||
searchQueueScheduler.join(10)
|
||||
except:
|
||||
|
@ -1343,7 +1307,7 @@ def halt():
|
|||
|
||||
if PROCESS_AUTOMATICALLY:
|
||||
autoPostProcesserScheduler.stop.set()
|
||||
logger.log(u"Waiting for the POSTPROCESSER thread to exit")
|
||||
logger.log(u'Waiting for the POSTPROCESSER thread to exit')
|
||||
try:
|
||||
autoPostProcesserScheduler.join(10)
|
||||
except:
|
||||
|
@ -1351,7 +1315,7 @@ def halt():
|
|||
|
||||
if USE_TRAKT:
|
||||
traktCheckerScheduler.stop.set()
|
||||
logger.log(u"Waiting for the TRAKTCHECKER thread to exit")
|
||||
logger.log(u'Waiting for the TRAKTCHECKER thread to exit')
|
||||
try:
|
||||
traktCheckerScheduler.join(10)
|
||||
except:
|
||||
|
@ -1359,7 +1323,7 @@ def halt():
|
|||
|
||||
if DOWNLOAD_PROPERS:
|
||||
properFinderScheduler.stop.set()
|
||||
logger.log(u"Waiting for the PROPERFINDER thread to exit")
|
||||
logger.log(u'Waiting for the PROPERFINDER thread to exit')
|
||||
try:
|
||||
properFinderScheduler.join(10)
|
||||
except:
|
||||
|
@ -1367,7 +1331,7 @@ def halt():
|
|||
|
||||
if USE_SUBTITLES:
|
||||
subtitlesFinderScheduler.stop.set()
|
||||
logger.log(u"Waiting for the SUBTITLESFINDER thread to exit")
|
||||
logger.log(u'Waiting for the SUBTITLESFINDER thread to exit')
|
||||
try:
|
||||
subtitlesFinderScheduler.join(10)
|
||||
except:
|
||||
|
@ -1375,7 +1339,7 @@ def halt():
|
|||
|
||||
if ADBA_CONNECTION:
|
||||
ADBA_CONNECTION.logout()
|
||||
logger.log(u"Waiting for the ANIDB CONNECTION thread to exit")
|
||||
logger.log(u'Waiting for the ANIDB CONNECTION thread to exit')
|
||||
try:
|
||||
ADBA_CONNECTION.join(10)
|
||||
except:
|
||||
|
@ -1387,7 +1351,7 @@ def halt():
|
|||
|
||||
def sig_handler(signum=None, frame=None):
|
||||
if type(signum) != type(None):
|
||||
logger.log(u"Signal %i caught, saving and exiting..." % int(signum))
|
||||
logger.log(u'Signal %i caught, saving and exiting...' % int(signum))
|
||||
events.put(events.SystemEvent.SHUTDOWN)
|
||||
|
||||
|
||||
|
@ -1395,12 +1359,12 @@ def saveAll():
|
|||
global showList
|
||||
|
||||
# write all shows
|
||||
logger.log(u"Saving all shows to the database")
|
||||
logger.log(u'Saving all shows to the database')
|
||||
for show in showList:
|
||||
show.saveToDB()
|
||||
|
||||
# save config
|
||||
logger.log(u"Saving config file to disk")
|
||||
logger.log(u'Saving config file to disk')
|
||||
save_config()
|
||||
|
||||
|
||||
|
@ -1408,7 +1372,7 @@ def restart(soft=True):
|
|||
if soft:
|
||||
halt()
|
||||
saveAll()
|
||||
logger.log(u"Re-initializing all data")
|
||||
logger.log(u'Re-initializing all data')
|
||||
initialize()
|
||||
else:
|
||||
events.put(events.SystemEvent.RESTART)
|
||||
|
@ -1517,7 +1481,7 @@ def save_config():
|
|||
new_config['General']['keep_processed_dir'] = int(KEEP_PROCESSED_DIR)
|
||||
new_config['General']['process_method'] = PROCESS_METHOD
|
||||
new_config['General']['move_associated_files'] = int(MOVE_ASSOCIATED_FILES)
|
||||
new_config['General']['postpone_if_sync_files'] = int (POSTPONE_IF_SYNC_FILES)
|
||||
new_config['General']['postpone_if_sync_files'] = int(POSTPONE_IF_SYNC_FILES)
|
||||
new_config['General']['nfo_rename'] = int(NFO_RENAME)
|
||||
new_config['General']['process_automatically'] = int(PROCESS_AUTOMATICALLY)
|
||||
new_config['General']['unpack'] = int(UNPACK)
|
||||
|
@ -1537,99 +1501,74 @@ def save_config():
|
|||
new_config['Blackhole']['torrent_dir'] = TORRENT_DIR
|
||||
|
||||
# dynamically save provider settings
|
||||
for curTorrentProvider in [curProvider for curProvider in providers.sortedProviderList() if
|
||||
curProvider.providerType == GenericProvider.TORRENT]:
|
||||
new_config[curTorrentProvider.getID().upper()] = {}
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID()] = int(curTorrentProvider.enabled)
|
||||
if hasattr(curTorrentProvider, 'digest'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_digest'] = curTorrentProvider.digest
|
||||
if hasattr(curTorrentProvider, 'hash'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_hash'] = curTorrentProvider.hash
|
||||
if hasattr(curTorrentProvider, 'api_key'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_api_key'] = curTorrentProvider.api_key
|
||||
if hasattr(curTorrentProvider, 'username'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_username'] = curTorrentProvider.username
|
||||
if hasattr(curTorrentProvider, 'password'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_password'] = helpers.encrypt(
|
||||
curTorrentProvider.password, ENCRYPTION_VERSION)
|
||||
if hasattr(curTorrentProvider, 'passkey'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_passkey'] = curTorrentProvider.passkey
|
||||
if hasattr(curTorrentProvider, 'confirmed'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_confirmed'] = int(
|
||||
curTorrentProvider.confirmed)
|
||||
if hasattr(curTorrentProvider, 'ratio'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_ratio'] = curTorrentProvider.ratio
|
||||
if hasattr(curTorrentProvider, 'minseed'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_minseed'] = int(
|
||||
curTorrentProvider.minseed)
|
||||
if hasattr(curTorrentProvider, 'minleech'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_minleech'] = int(
|
||||
curTorrentProvider.minleech)
|
||||
if hasattr(curTorrentProvider, 'options'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_options'] = curTorrentProvider.options
|
||||
if hasattr(curTorrentProvider, 'proxy'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_proxy'] = int(
|
||||
curTorrentProvider.proxy.enabled)
|
||||
if hasattr(curTorrentProvider.proxy, 'url'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_proxy_url'] = curTorrentProvider.proxy.url
|
||||
if hasattr(curTorrentProvider, 'freeleech'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_freeleech'] = int(
|
||||
curTorrentProvider.freeleech)
|
||||
if hasattr(curTorrentProvider, 'search_mode'):
|
||||
new_config[curTorrentProvider.getID().upper()][
|
||||
curTorrentProvider.getID() + '_search_mode'] = curTorrentProvider.search_mode
|
||||
if hasattr(curTorrentProvider, 'search_fallback'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_search_fallback'] = int(
|
||||
curTorrentProvider.search_fallback)
|
||||
if hasattr(curTorrentProvider, 'enable_recentsearch'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_enable_recentsearch'] = int(
|
||||
curTorrentProvider.enable_recentsearch)
|
||||
if hasattr(curTorrentProvider, 'enable_backlog'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_enable_backlog'] = int(
|
||||
curTorrentProvider.enable_backlog)
|
||||
for torrent_prov in [curProvider for curProvider in providers.sortedProviderList()
|
||||
if GenericProvider.TORRENT == curProvider.providerType]:
|
||||
prov_id = torrent_prov.get_id()
|
||||
prov_id_uc = torrent_prov.get_id().upper()
|
||||
new_config[prov_id_uc] = {}
|
||||
new_config[prov_id_uc][prov_id] = int(torrent_prov.enabled)
|
||||
if hasattr(torrent_prov, 'digest'):
|
||||
new_config[prov_id_uc][prov_id + '_digest'] = torrent_prov.digest
|
||||
if hasattr(torrent_prov, 'hash'):
|
||||
new_config[prov_id_uc][prov_id + '_hash'] = torrent_prov.hash
|
||||
if hasattr(torrent_prov, 'api_key'):
|
||||
new_config[prov_id_uc][prov_id + '_api_key'] = torrent_prov.api_key
|
||||
if hasattr(torrent_prov, 'username'):
|
||||
new_config[prov_id_uc][prov_id + '_username'] = torrent_prov.username
|
||||
if hasattr(torrent_prov, 'password'):
|
||||
new_config[prov_id_uc][prov_id + '_password'] = helpers.encrypt(torrent_prov.password, ENCRYPTION_VERSION)
|
||||
if hasattr(torrent_prov, 'passkey'):
|
||||
new_config[prov_id_uc][prov_id + '_passkey'] = torrent_prov.passkey
|
||||
if hasattr(torrent_prov, 'confirmed'):
|
||||
new_config[prov_id_uc][prov_id + '_confirmed'] = int(torrent_prov.confirmed)
|
||||
if hasattr(torrent_prov, 'ratio'):
|
||||
new_config[prov_id_uc][prov_id + '_ratio'] = torrent_prov.ratio
|
||||
if hasattr(torrent_prov, 'minseed'):
|
||||
new_config[prov_id_uc][prov_id + '_minseed'] = int(torrent_prov.minseed)
|
||||
if hasattr(torrent_prov, 'minleech'):
|
||||
new_config[prov_id_uc][prov_id + '_minleech'] = int(torrent_prov.minleech)
|
||||
if hasattr(torrent_prov, 'options'):
|
||||
new_config[prov_id_uc][prov_id + '_options'] = torrent_prov.options
|
||||
if hasattr(torrent_prov, 'proxy'):
|
||||
new_config[prov_id_uc][prov_id + '_proxy'] = int(torrent_prov.proxy.enabled)
|
||||
if hasattr(torrent_prov.proxy, 'url'):
|
||||
new_config[prov_id_uc][prov_id + '_proxy_url'] = torrent_prov.proxy.url
|
||||
if hasattr(torrent_prov, 'freeleech'):
|
||||
new_config[prov_id_uc][prov_id + '_freeleech'] = int(torrent_prov.freeleech)
|
||||
if hasattr(torrent_prov, 'search_mode'):
|
||||
new_config[prov_id_uc][prov_id + '_search_mode'] = torrent_prov.search_mode
|
||||
if hasattr(torrent_prov, 'search_fallback'):
|
||||
new_config[prov_id_uc][prov_id + '_search_fallback'] = int(torrent_prov.search_fallback)
|
||||
if hasattr(torrent_prov, 'enable_recentsearch'):
|
||||
new_config[prov_id_uc][prov_id + '_enable_recentsearch'] = int(torrent_prov.enable_recentsearch)
|
||||
if hasattr(torrent_prov, 'enable_backlog'):
|
||||
new_config[prov_id_uc][prov_id + '_enable_backlog'] = int(torrent_prov.enable_backlog)
|
||||
|
||||
for curNzbProvider in [curProvider for curProvider in providers.sortedProviderList() if
|
||||
curProvider.providerType == GenericProvider.NZB]:
|
||||
new_config[curNzbProvider.getID().upper()] = {}
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID()] = int(curNzbProvider.enabled)
|
||||
for nzb_prov in [curProvider for curProvider in providers.sortedProviderList()
|
||||
if GenericProvider.NZB == curProvider.providerType]:
|
||||
prov_id = nzb_prov.get_id()
|
||||
prov_id_uc = nzb_prov.get_id().upper()
|
||||
new_config[prov_id_uc] = {}
|
||||
new_config[prov_id_uc][prov_id] = int(nzb_prov.enabled)
|
||||
|
||||
if hasattr(curNzbProvider, 'api_key'):
|
||||
new_config[curNzbProvider.getID().upper()][
|
||||
curNzbProvider.getID() + '_api_key'] = curNzbProvider.api_key
|
||||
if hasattr(curNzbProvider, 'username'):
|
||||
new_config[curNzbProvider.getID().upper()][
|
||||
curNzbProvider.getID() + '_username'] = curNzbProvider.username
|
||||
if hasattr(curNzbProvider, 'search_mode'):
|
||||
new_config[curNzbProvider.getID().upper()][
|
||||
curNzbProvider.getID() + '_search_mode'] = curNzbProvider.search_mode
|
||||
if hasattr(curNzbProvider, 'search_fallback'):
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID() + '_search_fallback'] = int(
|
||||
curNzbProvider.search_fallback)
|
||||
if hasattr(curNzbProvider, 'enable_recentsearch'):
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID() + '_enable_recentsearch'] = int(
|
||||
curNzbProvider.enable_recentsearch)
|
||||
if hasattr(curNzbProvider, 'enable_backlog'):
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID() + '_enable_backlog'] = int(
|
||||
curNzbProvider.enable_backlog)
|
||||
if hasattr(nzb_prov, 'api_key'):
|
||||
new_config[prov_id_uc][prov_id + '_api_key'] = nzb_prov.api_key
|
||||
if hasattr(nzb_prov, 'username'):
|
||||
new_config[prov_id_uc][prov_id + '_username'] = nzb_prov.username
|
||||
if hasattr(nzb_prov, 'search_mode'):
|
||||
new_config[prov_id_uc][prov_id + '_search_mode'] = nzb_prov.search_mode
|
||||
if hasattr(nzb_prov, 'search_fallback'):
|
||||
new_config[prov_id_uc][prov_id + '_search_fallback'] = int(nzb_prov.search_fallback)
|
||||
if hasattr(nzb_prov, 'enable_recentsearch'):
|
||||
new_config[prov_id_uc][prov_id + '_enable_recentsearch'] = int(nzb_prov.enable_recentsearch)
|
||||
if hasattr(nzb_prov, 'enable_backlog'):
|
||||
new_config[prov_id_uc][prov_id + '_enable_backlog'] = int(nzb_prov.enable_backlog)
|
||||
|
||||
new_config['NZBs'] = {}
|
||||
new_config['NZBs']['nzbs'] = int(NZBS)
|
||||
new_config['NZBs']['nzbs_uid'] = NZBS_UID
|
||||
new_config['NZBs']['nzbs_hash'] = NZBS_HASH
|
||||
|
||||
new_config['Newzbin'] = {}
|
||||
new_config['Newzbin']['newzbin'] = int(NEWZBIN)
|
||||
new_config['Newzbin']['newzbin_username'] = NEWZBIN_USERNAME
|
||||
new_config['Newzbin']['newzbin_password'] = helpers.encrypt(NEWZBIN_PASSWORD, ENCRYPTION_VERSION)
|
||||
|
||||
new_config['SABnzbd'] = {}
|
||||
new_config['SABnzbd']['sab_username'] = SAB_USERNAME
|
||||
new_config['SABnzbd']['sab_password'] = helpers.encrypt(SAB_PASSWORD, ENCRYPTION_VERSION)
|
||||
|
@ -1881,41 +1820,41 @@ def save_config():
|
|||
new_config.write()
|
||||
|
||||
|
||||
def launchBrowser(startPort=None):
|
||||
if not startPort:
|
||||
startPort = WEB_PORT
|
||||
def launchBrowser(start_port=None):
|
||||
if not start_port:
|
||||
start_port = WEB_PORT
|
||||
if ENABLE_HTTPS:
|
||||
browserURL = 'https://localhost:%d%s' % (startPort, WEB_ROOT)
|
||||
browser_url = 'https://localhost:%d%s' % (start_port, WEB_ROOT)
|
||||
else:
|
||||
browserURL = 'http://localhost:%d%s' % (startPort, WEB_ROOT)
|
||||
browser_url = 'http://localhost:%d%s' % (start_port, WEB_ROOT)
|
||||
try:
|
||||
webbrowser.open(browserURL, 2, 1)
|
||||
webbrowser.open(browser_url, 2, 1)
|
||||
except:
|
||||
try:
|
||||
webbrowser.open(browserURL, 1, 1)
|
||||
webbrowser.open(browser_url, 1, 1)
|
||||
except:
|
||||
logger.log(u"Unable to launch a browser", logger.ERROR)
|
||||
logger.log(u'Unable to launch a browser', logger.ERROR)
|
||||
|
||||
|
||||
def getEpList(epIDs, showid=None):
|
||||
if epIDs == None or len(epIDs) == 0:
|
||||
def getEpList(ep_ids, showid=None):
|
||||
if None is ep_ids or 0 == len(ep_ids):
|
||||
return []
|
||||
|
||||
query = "SELECT * FROM tv_episodes WHERE indexerid in (%s)" % (",".join(['?'] * len(epIDs)),)
|
||||
params = epIDs
|
||||
query = 'SELECT * FROM tv_episodes WHERE indexerid in (%s)' % (','.join(['?'] * len(ep_ids)),)
|
||||
params = ep_ids
|
||||
|
||||
if showid != None:
|
||||
query += " AND showid = ?"
|
||||
if None is not showid:
|
||||
query += ' AND showid = ?'
|
||||
params.append(showid)
|
||||
|
||||
myDB = db.DBConnection()
|
||||
sqlResults = myDB.select(query, params)
|
||||
my_db = db.DBConnection()
|
||||
sql_results = my_db.select(query, params)
|
||||
|
||||
epList = []
|
||||
ep_list = []
|
||||
|
||||
for curEp in sqlResults:
|
||||
curShowObj = helpers.findCertainShow(showList, int(curEp["showid"]))
|
||||
curEpObj = curShowObj.getEpisode(int(curEp["season"]), int(curEp["episode"]))
|
||||
epList.append(curEpObj)
|
||||
for curEp in sql_results:
|
||||
cur_show_obj = helpers.findCertainShow(showList, int(curEp['showid']))
|
||||
cur_ep_obj = cur_show_obj.getEpisode(int(curEp['season']), int(curEp['episode']))
|
||||
ep_list.append(cur_ep_obj)
|
||||
|
||||
return epList
|
||||
return ep_list
|
||||
|
|
|
@ -168,7 +168,7 @@ class GenericClient(object):
|
|||
|
||||
try:
|
||||
# Sets per provider seed ratio
|
||||
result.ratio = result.provider.seedRatio()
|
||||
result.ratio = result.provider.seed_ratio()
|
||||
|
||||
result = self._get_torrent_hash(result)
|
||||
|
||||
|
|
|
@ -721,7 +721,7 @@ class ConfigMigrator():
|
|||
for curProvider in providers.sortedProviderList():
|
||||
if hasattr(curProvider, 'enable_recentsearch'):
|
||||
curProvider.enable_recentsearch = bool(check_setting_int(
|
||||
self.config_obj, curProvider.getID().upper(), curProvider.getID() + '_enable_dailysearch', 1))
|
||||
self.config_obj, curProvider.get_id().upper(), curProvider.get_id() + '_enable_dailysearch', 1))
|
||||
|
||||
def _migrate_v7(self):
|
||||
sickbeard.EPISODE_VIEW_LAYOUT = check_setting_str(self.config_obj, 'GUI', 'coming_eps_layout', 'banner')
|
||||
|
|
|
@ -100,10 +100,6 @@ class EpisodeNotFoundException(SickBeardException):
|
|||
"The episode wasn't found on the Indexer"
|
||||
|
||||
|
||||
class NewzbinAPIThrottled(SickBeardException):
|
||||
"Newzbin has throttled us, deal with it"
|
||||
|
||||
|
||||
class ShowDirNotFoundException(SickBeardException):
|
||||
"The show dir doesn't exist"
|
||||
|
||||
|
|
|
@ -1400,7 +1400,7 @@ def check_port(host, port, timeout=1.0):
|
|||
|
||||
|
||||
def clear_unused_providers():
|
||||
providers = [x.cache.providerID for x in sickbeard.providers.sortedProviderList() if x.isActive()]
|
||||
providers = [x.cache.providerID for x in sickbeard.providers.sortedProviderList() if x.is_active()]
|
||||
|
||||
if providers:
|
||||
myDB = db.DBConnection('cache.db')
|
||||
|
|
|
@ -104,7 +104,7 @@ def sendNZB(nzb, proper=False):
|
|||
else:
|
||||
if nzb.resultType == "nzb":
|
||||
genProvider = GenericProvider("")
|
||||
data = genProvider.getURL(nzb.url)
|
||||
data = genProvider.get_url(nzb.url)
|
||||
if (data == None):
|
||||
return False
|
||||
nzbcontent64 = standard_b64encode(data)
|
||||
|
|
|
@ -66,14 +66,14 @@ def _getProperList():
|
|||
|
||||
# for each provider get a list of the
|
||||
origThreadName = threading.currentThread().name
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.isActive()]
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.is_active()]
|
||||
for curProvider in providers:
|
||||
threading.currentThread().name = origThreadName + ' :: [' + curProvider.name + ']'
|
||||
|
||||
logger.log(u'Searching for any new PROPER releases from ' + curProvider.name)
|
||||
|
||||
try:
|
||||
curPropers = curProvider.findPropers(search_date)
|
||||
curPropers = curProvider.find_propers(search_date)
|
||||
except exceptions.AuthException as e:
|
||||
logger.log(u'Authentication error: ' + ex(e), logger.ERROR)
|
||||
continue
|
||||
|
@ -243,8 +243,9 @@ def _downloadPropers(properList):
|
|||
epObj = showObj.getEpisode(curProper.season, curProper.episode)
|
||||
|
||||
# make the result object
|
||||
result = curProper.provider.getResult([epObj])
|
||||
result.url = curProper.url
|
||||
result = curProper.provider.get_result([epObj], curProper.url)
|
||||
if None is result:
|
||||
continue
|
||||
result.name = curProper.name
|
||||
result.quality = curProper.quality
|
||||
result.version = curProper.version
|
||||
|
|
|
@ -53,7 +53,7 @@ from sickbeard import logger
|
|||
|
||||
def sortedProviderList():
|
||||
initialList = sickbeard.providerList + sickbeard.newznabProviderList + sickbeard.torrentRssProviderList
|
||||
providerDict = dict(zip([x.getID() for x in initialList], initialList))
|
||||
providerDict = dict(zip([x.get_id() for x in initialList], initialList))
|
||||
|
||||
newList = []
|
||||
|
||||
|
@ -207,7 +207,7 @@ def getProviderModule(name):
|
|||
def getProviderClass(id):
|
||||
providerMatch = [x for x in
|
||||
sickbeard.providerList + sickbeard.newznabProviderList + sickbeard.torrentRssProviderList if
|
||||
x.getID() == id]
|
||||
x.get_id() == id]
|
||||
|
||||
if len(providerMatch) != 1:
|
||||
return None
|
||||
|
|
|
@ -47,13 +47,13 @@ class AlphaRatioProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = AlphaRatioCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'session' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'keeplogged': '1', 'login': 'Login'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -66,10 +66,10 @@ class AlphaRatioProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -82,7 +82,7 @@ class AlphaRatioProvider(generic.TorrentProvider):
|
|||
search_string = unidecode(search_string)
|
||||
|
||||
search_url = self.urls['search'] % search_string
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -124,7 +124,7 @@ class AlphaRatioProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -16,73 +16,67 @@
|
|||
# You should have received a copy of the GNU General Public License
|
||||
# along with Sick Beard. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
import urllib
|
||||
import datetime
|
||||
import urllib
|
||||
|
||||
import generic
|
||||
from . import generic
|
||||
from sickbeard import classes, show_name_helpers, logger, tvcache
|
||||
|
||||
|
||||
class animenzb(generic.NZBProvider):
|
||||
class AnimeNZBProvider(generic.NZBProvider):
|
||||
|
||||
def __init__(self):
|
||||
generic.NZBProvider.__init__(self, 'animenzb', True, True)
|
||||
self.cache = animenzbCache(self)
|
||||
generic.NZBProvider.__init__(self, 'AnimeNZB', anime_only=True)
|
||||
|
||||
self.url = 'http://animenzb.com/'
|
||||
|
||||
self.cache = AnimeNZBCache(self)
|
||||
|
||||
def _get_season_search_strings(self, ep_obj):
|
||||
|
||||
return [x for x in show_name_helpers.makeSceneSeasonSearchString(self.show, ep_obj)]
|
||||
|
||||
def _get_episode_search_strings(self, ep_obj, add_string=''):
|
||||
|
||||
return [x for x in show_name_helpers.makeSceneSearchString(self.show, ep_obj)]
|
||||
|
||||
def _doSearch(self, search_string, search_mode='eponly', epcount=0, age=0):
|
||||
if self.show and not self.show.is_anime:
|
||||
logger.log(u'%s is not an anime skipping ...' % self.show.name)
|
||||
return []
|
||||
def _do_search(self, search_string, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
params = {
|
||||
'cat': 'anime',
|
||||
'q': search_string.encode('utf-8'),
|
||||
'max': '100'
|
||||
}
|
||||
results = []
|
||||
if self.show and not self.show.is_anime:
|
||||
return results
|
||||
|
||||
params = {'cat': 'anime',
|
||||
'q': search_string.encode('utf-8'),
|
||||
'max': '100'}
|
||||
|
||||
search_url = self.url + 'rss?' + urllib.urlencode(params)
|
||||
|
||||
logger.log(u'Search url: %s' % search_url, logger.DEBUG)
|
||||
|
||||
data = self.cache.getRSSFeed(search_url)
|
||||
if not data:
|
||||
return []
|
||||
|
||||
if 'entries' in data:
|
||||
if data and 'entries' in data:
|
||||
|
||||
items = data.entries
|
||||
results = []
|
||||
|
||||
for curItem in items:
|
||||
(title, url) = self._get_title_and_url(curItem)
|
||||
|
||||
if title and url:
|
||||
results.append(curItem)
|
||||
else:
|
||||
logger.log(
|
||||
u'The data returned from %s is incomplete, this result is unusable' % self.name,
|
||||
logger.DEBUG)
|
||||
logger.log(u'The data returned from %s is incomplete, this result is unusable' % self.name,
|
||||
logger.DEBUG)
|
||||
|
||||
return results
|
||||
return results
|
||||
|
||||
return []
|
||||
|
||||
def findPropers(self, date=None):
|
||||
def find_propers(self, date=None):
|
||||
|
||||
results = []
|
||||
|
||||
for item in self._doSearch('v2|v3|v4|v5'):
|
||||
for item in self._do_search('v2|v3|v4|v5'):
|
||||
|
||||
(title, url) = self._get_title_and_url(item)
|
||||
|
||||
if item.has_key('published_parsed') and item['published_parsed']:
|
||||
if 'published_parsed' in item and item['published_parsed']:
|
||||
result_date = item.published_parsed
|
||||
if result_date:
|
||||
result_date = datetime.datetime(*result_date[0:6])
|
||||
|
@ -97,32 +91,25 @@ class animenzb(generic.NZBProvider):
|
|||
return results
|
||||
|
||||
|
||||
class animenzbCache(tvcache.TVCache):
|
||||
class AnimeNZBCache(tvcache.TVCache):
|
||||
|
||||
def __init__(self, provider):
|
||||
def __init__(self, this_provider):
|
||||
tvcache.TVCache.__init__(self, this_provider)
|
||||
|
||||
tvcache.TVCache.__init__(self, provider)
|
||||
|
||||
|
||||
self.minTime = 20
|
||||
self.minTime = 20 # cache update frequency
|
||||
|
||||
def _getRSSData(self):
|
||||
|
||||
params = {
|
||||
'cat': 'anime'.encode('utf-8'),
|
||||
'max': '100'.encode('utf-8')
|
||||
}
|
||||
params = {'cat': 'anime'.encode('utf-8'),
|
||||
'max': '100'.encode('utf-8')}
|
||||
|
||||
rss_url = self.provider.url + 'rss?' + urllib.urlencode(params)
|
||||
|
||||
logger.log(u'%s cache update URL: %s' % (self.provider.name, rss_url), logger.DEBUG)
|
||||
|
||||
data = self.getRSSFeed(rss_url)
|
||||
|
||||
if data and 'entries' in data:
|
||||
return data.entries
|
||||
else:
|
||||
return []
|
||||
return []
|
||||
|
||||
|
||||
provider = animenzb()
|
||||
provider = AnimeNZBProvider()
|
||||
|
|
|
@ -15,6 +15,7 @@
|
|||
# You should have received a copy of the GNU General Public License
|
||||
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
import re
|
||||
import datetime
|
||||
import time
|
||||
|
||||
|
@ -52,10 +53,10 @@ class BeyondHDProvider(generic.TorrentProvider):
|
|||
logger.log(u'Incorrect authentication credentials for %s : %s' % (self.name, data_json['error']), logger.DEBUG)
|
||||
raise AuthException('Authentication credentials for %s are incorrect, check your config' % self.name)
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._checkAuth():
|
||||
if not self._check_auth():
|
||||
return results
|
||||
|
||||
for mode in search_params.keys():
|
||||
|
@ -72,9 +73,9 @@ class BeyondHDProvider(generic.TorrentProvider):
|
|||
|
||||
search_url = self.urls['cache'] % (self.passkey, self.categories[mode])
|
||||
if 'Cache' != mode:
|
||||
search_url += self.urls['search'] % search_string
|
||||
search_url += self.urls['search'] % re.sub('[\.\s]+', ' ', search_string)
|
||||
|
||||
data_json = self.getURL(search_url, json=True)
|
||||
data_json = self.get_url(search_url, json=True)
|
||||
|
||||
cnt = len(results)
|
||||
if data_json and 'results' in data_json and self._check_auth_from_data(data_json):
|
||||
|
@ -97,7 +98,7 @@ class BeyondHDProvider(generic.TorrentProvider):
|
|||
|
||||
return generic.TorrentProvider._get_episode_search_strings(self, ep_obj, add_string, scene=False, use_or=False)
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date, ['proper', 'repack'])
|
||||
|
||||
|
|
|
@ -43,13 +43,13 @@ class BitSoupProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = BitSoupCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'ssl': 'yes'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -62,10 +62,10 @@ class BitSoupProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -78,7 +78,7 @@ class BitSoupProvider(generic.TorrentProvider):
|
|||
search_string = unidecode(search_string)
|
||||
|
||||
search_url = self.urls['search'] % (search_string, self.categories)
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -121,7 +121,7 @@ class BitSoupProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -43,7 +43,7 @@ class BTNProvider(generic.TorrentProvider):
|
|||
def _check_auth_from_data(self, data_json):
|
||||
|
||||
if data_json is None:
|
||||
return self._checkAuth()
|
||||
return self._check_auth()
|
||||
|
||||
if 'api-error' not in data_json:
|
||||
return True
|
||||
|
@ -52,9 +52,9 @@ class BTNProvider(generic.TorrentProvider):
|
|||
logger.DEBUG)
|
||||
raise AuthException('Your authentication credentials for %s are incorrect, check your config.' % self.name)
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
self._checkAuth()
|
||||
self._check_auth()
|
||||
|
||||
params = {}
|
||||
|
||||
|
@ -236,14 +236,14 @@ class BTNProvider(generic.TorrentProvider):
|
|||
|
||||
return to_return
|
||||
|
||||
def findPropers(self, search_date=None):
|
||||
def find_propers(self, search_date=None):
|
||||
|
||||
results = []
|
||||
|
||||
search_terms = ['%.proper.%', '%.repack.%']
|
||||
|
||||
for term in search_terms:
|
||||
for item in self._doSearch({'release': term}, age=4 * 24 * 60 * 60):
|
||||
for item in self._do_search({'release': term}, age=4 * 24 * 60 * 60):
|
||||
if item['Time']:
|
||||
try:
|
||||
result_date = datetime.datetime.fromtimestamp(float(item['Time']))
|
||||
|
@ -273,7 +273,7 @@ class BTNProvider(generic.TorrentProvider):
|
|||
% self.name, logger.WARNING)
|
||||
seconds_since_last_update = 86400
|
||||
|
||||
return self._doSearch(search_params=None, age=seconds_since_last_update)
|
||||
return self._do_search(search_params=None, age=seconds_since_last_update)
|
||||
|
||||
|
||||
class BTNCache(tvcache.TVCache):
|
||||
|
|
|
@ -42,13 +42,13 @@ class FreshOnTVProvider(generic.TorrentProvider):
|
|||
self.freeleech = False
|
||||
self.cache = FreshOnTVCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'login': 'Do it!'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -64,10 +64,10 @@ class FreshOnTVProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -83,7 +83,7 @@ class FreshOnTVProvider(generic.TorrentProvider):
|
|||
|
||||
# returns top 15 results by default, expandable in user profile to 100
|
||||
search_url = self.urls['search'] % (freeleech, search_string)
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -129,7 +129,7 @@ class FreshOnTVProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -28,6 +28,7 @@ from base64 import b16encode, b32decode
|
|||
|
||||
import sickbeard
|
||||
import requests
|
||||
import requests.cookies
|
||||
from sickbeard import helpers, classes, logger, db, tvcache, encodingKludge as ek
|
||||
from sickbeard.common import Quality, MULTI_EP_RESULT, SEASON_RESULT, USER_AGENT
|
||||
from sickbeard.exceptions import SickBeardException, AuthException, ex
|
||||
|
@ -42,8 +43,8 @@ class HaltParseException(SickBeardException):
|
|||
|
||||
|
||||
class GenericProvider:
|
||||
NZB = "nzb"
|
||||
TORRENT = "torrent"
|
||||
NZB = 'nzb'
|
||||
TORRENT = 'torrent'
|
||||
|
||||
def __init__(self, name, supports_backlog=False, anime_only=False):
|
||||
# these need to be set in the subclass
|
||||
|
@ -67,85 +68,86 @@ class GenericProvider:
|
|||
|
||||
self.headers = {
|
||||
# Using USER_AGENT instead of Mozilla to keep same user agent along authentication and download phases,
|
||||
#otherwise session might be broken and download fail, asking again for authentication
|
||||
#'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/32.0.1700.107 Safari/537.36'}
|
||||
# otherwise session might be broken and download fail, asking again for authentication
|
||||
# 'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/32.0.1700.107 Safari/537.36'}
|
||||
'User-Agent': USER_AGENT}
|
||||
|
||||
def getID(self):
|
||||
return GenericProvider.makeID(self.name)
|
||||
def get_id(self):
|
||||
return GenericProvider.make_id(self.name)
|
||||
|
||||
@staticmethod
|
||||
def makeID(name):
|
||||
return re.sub("[^\w\d_]", "_", name.strip().lower())
|
||||
def make_id(name):
|
||||
return re.sub('[^\w\d_]', '_', name.strip().lower())
|
||||
|
||||
def imageName(self, *default_name):
|
||||
def image_name(self, *default_name):
|
||||
|
||||
for name in ['%s.%s' % (self.getID(), image_ext) for image_ext in ['png', 'gif', 'jpg']]:
|
||||
for name in ['%s.%s' % (self.get_id(), image_ext) for image_ext in ['png', 'gif', 'jpg']]:
|
||||
if ek.ek(os.path.isfile,
|
||||
ek.ek(os.path.join, sickbeard.PROG_DIR, 'gui', sickbeard.GUI_NAME, 'images', 'providers', name)):
|
||||
return name
|
||||
|
||||
return '%s.png' % ('newznab', default_name[0])[any(default_name)]
|
||||
|
||||
def _checkAuth(self):
|
||||
def _check_auth(self):
|
||||
return True
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
return True
|
||||
|
||||
def isActive(self):
|
||||
if self.providerType == GenericProvider.NZB and sickbeard.USE_NZBS:
|
||||
return self.isEnabled()
|
||||
elif self.providerType == GenericProvider.TORRENT and sickbeard.USE_TORRENTS:
|
||||
return self.isEnabled()
|
||||
def is_active(self):
|
||||
if GenericProvider.NZB == self.providerType and sickbeard.USE_NZBS:
|
||||
return self.is_enabled()
|
||||
elif GenericProvider.TORRENT == self.providerType and sickbeard.USE_TORRENTS:
|
||||
return self.is_enabled()
|
||||
else:
|
||||
return False
|
||||
|
||||
def isEnabled(self):
|
||||
def is_enabled(self):
|
||||
"""
|
||||
This should be overridden and should return the config setting eg. sickbeard.MYPROVIDER
|
||||
"""
|
||||
return self.enabled
|
||||
|
||||
def getResult(self, episodes):
|
||||
def get_result(self, episodes, url):
|
||||
"""
|
||||
Returns a result of the correct type for this provider
|
||||
"""
|
||||
|
||||
if self.providerType == GenericProvider.NZB:
|
||||
if GenericProvider.NZB == self.providerType:
|
||||
result = classes.NZBSearchResult(episodes)
|
||||
elif self.providerType == GenericProvider.TORRENT:
|
||||
elif GenericProvider.TORRENT == self.providerType:
|
||||
result = classes.TorrentSearchResult(episodes)
|
||||
else:
|
||||
result = classes.SearchResult(episodes)
|
||||
|
||||
result.provider = self
|
||||
result.url = url
|
||||
|
||||
return result
|
||||
|
||||
def getURL(self, url, post_data=None, params=None, timeout=30, json=False):
|
||||
def get_url(self, url, post_data=None, params=None, timeout=30, json=False):
|
||||
"""
|
||||
By default this is just a simple urlopen call but this method should be overridden
|
||||
for providers with special URL requirements (like cookies)
|
||||
"""
|
||||
|
||||
# check for auth
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return
|
||||
|
||||
return helpers.getURL(url, post_data=post_data, params=params, headers=self.headers, timeout=timeout,
|
||||
session=self.session, json=json)
|
||||
|
||||
def downloadResult(self, result):
|
||||
def download_result(self, result):
|
||||
"""
|
||||
Save the result to disk.
|
||||
"""
|
||||
|
||||
# check for auth
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return False
|
||||
|
||||
if self.providerType == GenericProvider.TORRENT:
|
||||
if GenericProvider.TORRENT == self.providerType:
|
||||
try:
|
||||
torrent_hash = re.findall('urn:btih:([0-9a-f]{32,40})', result.url)[0].upper()
|
||||
|
||||
|
@ -153,7 +155,7 @@ class GenericProvider:
|
|||
torrent_hash = b16encode(b32decode(torrent_hash)).lower()
|
||||
|
||||
if not torrent_hash:
|
||||
logger.log("Unable to extract torrent hash from link: " + ex(result.url), logger.ERROR)
|
||||
logger.log('Unable to extract torrent hash from link: ' + ex(result.url), logger.ERROR)
|
||||
return False
|
||||
|
||||
urls = ['https://%s/%s.torrent' % (u, torrent_hash)
|
||||
|
@ -163,7 +165,7 @@ class GenericProvider:
|
|||
|
||||
filename = ek.ek(os.path.join, sickbeard.TORRENT_DIR,
|
||||
helpers.sanitizeFileName(result.name) + '.' + self.providerType)
|
||||
elif self.providerType == GenericProvider.NZB:
|
||||
elif GenericProvider.NZB == self.providerType:
|
||||
urls = [result.url]
|
||||
|
||||
filename = ek.ek(os.path.join, sickbeard.NZB_DIR,
|
||||
|
@ -173,19 +175,19 @@ class GenericProvider:
|
|||
|
||||
for url in urls:
|
||||
if helpers.download_file(url, filename, session=self.session):
|
||||
logger.log(u"Downloading a result from " + self.name + " at " + url)
|
||||
logger.log(u'Downloading a result from ' + self.name + ' at ' + url)
|
||||
|
||||
if self.providerType == GenericProvider.TORRENT:
|
||||
logger.log(u"Saved magnet link to " + filename, logger.MESSAGE)
|
||||
if GenericProvider.TORRENT == self.providerType:
|
||||
logger.log(u'Saved magnet link to ' + filename, logger.MESSAGE)
|
||||
else:
|
||||
logger.log(u"Saved result to " + filename, logger.MESSAGE)
|
||||
logger.log(u'Saved result to ' + filename, logger.MESSAGE)
|
||||
|
||||
if self._verify_download(filename):
|
||||
return True
|
||||
elif ek.ek(os.path.isfile, filename):
|
||||
ek.ek(os.remove, filename)
|
||||
|
||||
logger.log(u"Failed to download result", logger.ERROR)
|
||||
logger.log(u'Failed to download result', logger.ERROR)
|
||||
return False
|
||||
|
||||
def _verify_download(self, file_name=None):
|
||||
|
@ -194,7 +196,7 @@ class GenericProvider:
|
|||
"""
|
||||
|
||||
# primitive verification of torrents, just make sure we didn't get a text file or something
|
||||
if self.providerType == GenericProvider.TORRENT:
|
||||
if GenericProvider.TORRENT == self.providerType:
|
||||
parser = createParser(file_name)
|
||||
if parser:
|
||||
mime_type = parser._getMimeType()
|
||||
|
@ -202,18 +204,18 @@ class GenericProvider:
|
|||
parser.stream._input.close()
|
||||
except:
|
||||
pass
|
||||
if mime_type == 'application/x-bittorrent':
|
||||
if 'application/x-bittorrent' == mime_type:
|
||||
return True
|
||||
|
||||
logger.log(u"Result is not a valid torrent file", logger.WARNING)
|
||||
logger.log(u'Result is not a valid torrent file', logger.WARNING)
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
def searchRSS(self, episodes):
|
||||
def search_rss(self, episodes):
|
||||
return self.cache.findNeededEpisodes(episodes)
|
||||
|
||||
def getQuality(self, item, anime=False):
|
||||
def get_quality(self, item, anime=False):
|
||||
"""
|
||||
Figures out the quality of the given RSS item node
|
||||
|
||||
|
@ -225,13 +227,13 @@ class GenericProvider:
|
|||
quality = Quality.sceneQuality(title, anime)
|
||||
return quality
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
return []
|
||||
|
||||
def _get_season_search_strings(self, episode):
|
||||
return []
|
||||
|
||||
def _get_episode_search_strings(self, eb_obj, add_string=''):
|
||||
def _get_episode_search_strings(self, *args, **kwargs):
|
||||
return []
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
|
@ -259,223 +261,216 @@ class GenericProvider:
|
|||
pass
|
||||
|
||||
if title:
|
||||
title = re.sub(r'\s+', '.', u'' + title)
|
||||
title = re.sub(r'\s+', '.', u'%s' % title)
|
||||
|
||||
if url:
|
||||
url = url.replace('&', '&')
|
||||
url = str(url).replace('&', '&')
|
||||
|
||||
return title, url
|
||||
|
||||
def findSearchResults(self, show, episodes, search_mode, manualSearch=False):
|
||||
def find_search_results(self, show, episodes, search_mode, manual_search=False):
|
||||
|
||||
self._checkAuth()
|
||||
self._check_auth()
|
||||
self.show = show
|
||||
|
||||
results = {}
|
||||
itemList = []
|
||||
item_list = []
|
||||
|
||||
searched_scene_season = None
|
||||
for epObj in episodes:
|
||||
for ep_obj in episodes:
|
||||
# search cache for episode result
|
||||
cacheResult = self.cache.searchCache(epObj, manualSearch)
|
||||
if cacheResult:
|
||||
if epObj.episode not in results:
|
||||
results[epObj.episode] = cacheResult
|
||||
cache_result = self.cache.searchCache(ep_obj, manual_search)
|
||||
if cache_result:
|
||||
if ep_obj.episode not in results:
|
||||
results[ep_obj.episode] = cache_result
|
||||
else:
|
||||
results[epObj.episode].extend(cacheResult)
|
||||
results[ep_obj.episode].extend(cache_result)
|
||||
|
||||
# found result, search next episode
|
||||
continue
|
||||
|
||||
# skip if season already searched
|
||||
if len(episodes) > 1 and searched_scene_season == epObj.scene_season:
|
||||
if 1 < len(episodes) and ep_obj.scene_season == searched_scene_season:
|
||||
continue
|
||||
|
||||
# mark season searched for season pack searches so we can skip later on
|
||||
searched_scene_season = epObj.scene_season
|
||||
searched_scene_season = ep_obj.scene_season
|
||||
|
||||
if 'sponly' == search_mode:
|
||||
# get season search results
|
||||
for curString in self._get_season_search_strings(epObj):
|
||||
itemList += self._doSearch(curString, search_mode, len(episodes))
|
||||
for curString in self._get_season_search_strings(ep_obj):
|
||||
item_list += self._do_search(curString, search_mode, len(episodes))
|
||||
else:
|
||||
# get single episode search results
|
||||
for curString in self._get_episode_search_strings(epObj):
|
||||
itemList += self._doSearch(curString, 'eponly', len(episodes))
|
||||
for curString in self._get_episode_search_strings(ep_obj):
|
||||
item_list += self._do_search(curString, 'eponly', len(episodes))
|
||||
|
||||
# if we found what we needed already from cache then return results and exit
|
||||
if len(results) == len(episodes):
|
||||
return results
|
||||
|
||||
# sort list by quality
|
||||
if len(itemList):
|
||||
if len(item_list):
|
||||
items = {}
|
||||
itemsUnknown = []
|
||||
for item in itemList:
|
||||
quality = self.getQuality(item, anime=show.is_anime)
|
||||
if quality == Quality.UNKNOWN:
|
||||
itemsUnknown += [item]
|
||||
items_unknown = []
|
||||
for item in item_list:
|
||||
quality = self.get_quality(item, anime=show.is_anime)
|
||||
if Quality.UNKNOWN == quality:
|
||||
items_unknown += [item]
|
||||
else:
|
||||
if quality not in items:
|
||||
items[quality] = [item]
|
||||
else:
|
||||
items[quality].append(item)
|
||||
|
||||
itemList = list(itertools.chain(*[v for (k, v) in sorted(items.items(), reverse=True)]))
|
||||
itemList += itemsUnknown if itemsUnknown else []
|
||||
item_list = list(itertools.chain(*[v for (k, v) in sorted(items.items(), reverse=True)]))
|
||||
item_list += items_unknown if items_unknown else []
|
||||
|
||||
# filter results
|
||||
cl = []
|
||||
for item in itemList:
|
||||
for item in item_list:
|
||||
(title, url) = self._get_title_and_url(item)
|
||||
|
||||
# parse the file name
|
||||
try:
|
||||
myParser = NameParser(False, convert=True)
|
||||
parse_result = myParser.parse(title)
|
||||
parser = NameParser(False, convert=True)
|
||||
parse_result = parser.parse(title)
|
||||
except InvalidNameException:
|
||||
logger.log(u"Unable to parse the filename " + title + " into a valid episode", logger.DEBUG)
|
||||
logger.log(u'Unable to parse the filename ' + title + ' into a valid episode', logger.DEBUG)
|
||||
continue
|
||||
except InvalidShowException:
|
||||
logger.log(u'No show name or scene exception matched the parsed filename ' + title, logger.DEBUG)
|
||||
continue
|
||||
|
||||
showObj = parse_result.show
|
||||
show_obj = parse_result.show
|
||||
quality = parse_result.quality
|
||||
release_group = parse_result.release_group
|
||||
version = parse_result.version
|
||||
|
||||
addCacheEntry = False
|
||||
if not (showObj.air_by_date or showObj.sports):
|
||||
if search_mode == 'sponly':
|
||||
add_cache_entry = False
|
||||
if not (show_obj.air_by_date or show_obj.sports):
|
||||
if 'sponly' == search_mode:
|
||||
if len(parse_result.episode_numbers):
|
||||
logger.log(
|
||||
u"This is supposed to be a season pack search but the result " + title + " is not a valid season pack, skipping it",
|
||||
logger.DEBUG)
|
||||
addCacheEntry = True
|
||||
if len(parse_result.episode_numbers) and (
|
||||
parse_result.season_number not in set([ep.season for ep in episodes]) or not [ep for ep in episodes if
|
||||
ep.scene_episode in parse_result.episode_numbers]):
|
||||
logger.log(
|
||||
u"The result " + title + " doesn't seem to be a valid episode that we are trying to snatch, ignoring",
|
||||
logger.DEBUG)
|
||||
addCacheEntry = True
|
||||
logger.log(u'This is supposed to be a season pack search but the result ' + title
|
||||
+ u' is not a valid season pack, skipping it', logger.DEBUG)
|
||||
add_cache_entry = True
|
||||
if len(parse_result.episode_numbers)\
|
||||
and (parse_result.season_number not in set([ep.season for ep in episodes])
|
||||
or not [ep for ep in episodes if ep.scene_episode in parse_result.episode_numbers]):
|
||||
logger.log(u'The result ' + title + u' doesn\'t seem to be a valid episode that we are trying'
|
||||
+ u' to snatch, ignoring', logger.DEBUG)
|
||||
add_cache_entry = True
|
||||
else:
|
||||
if not len(parse_result.episode_numbers) and parse_result.season_number and not [ep for ep in
|
||||
episodes if
|
||||
ep.season == parse_result.season_number and ep.episode in parse_result.episode_numbers]:
|
||||
logger.log(
|
||||
u"The result " + title + " doesn't seem to be a valid season that we are trying to snatch, ignoring",
|
||||
logger.DEBUG)
|
||||
addCacheEntry = True
|
||||
if not len(parse_result.episode_numbers)\
|
||||
and parse_result.season_number\
|
||||
and not [ep for ep in episodes
|
||||
if ep.season == parse_result.season_number
|
||||
and ep.episode in parse_result.episode_numbers]:
|
||||
logger.log(u'The result ' + title + u' doesn\'t seem to be a valid season that we are trying'
|
||||
+ u' to snatch, ignoring', logger.DEBUG)
|
||||
add_cache_entry = True
|
||||
elif len(parse_result.episode_numbers) and not [ep for ep in episodes if
|
||||
ep.season == parse_result.season_number and ep.episode in parse_result.episode_numbers]:
|
||||
logger.log(
|
||||
u"The result " + title + " doesn't seem to be a valid episode that we are trying to snatch, ignoring",
|
||||
logger.DEBUG)
|
||||
addCacheEntry = True
|
||||
logger.log(u'The result ' + title + ' doesn\'t seem to be a valid episode that we are trying'
|
||||
+ u' to snatch, ignoring', logger.DEBUG)
|
||||
add_cache_entry = True
|
||||
|
||||
if not addCacheEntry:
|
||||
if not add_cache_entry:
|
||||
# we just use the existing info for normal searches
|
||||
actual_season = parse_result.season_number
|
||||
actual_episodes = parse_result.episode_numbers
|
||||
else:
|
||||
if not (parse_result.is_air_by_date):
|
||||
logger.log(
|
||||
u"This is supposed to be a date search but the result " + title + " didn't parse as one, skipping it",
|
||||
logger.DEBUG)
|
||||
addCacheEntry = True
|
||||
if not parse_result.is_air_by_date:
|
||||
logger.log(u'This is supposed to be a date search but the result ' + title
|
||||
+ u' didn\'t parse as one, skipping it', logger.DEBUG)
|
||||
add_cache_entry = True
|
||||
else:
|
||||
airdate = parse_result.air_date.toordinal()
|
||||
myDB = db.DBConnection()
|
||||
sql_results = myDB.select(
|
||||
"SELECT season, episode FROM tv_episodes WHERE showid = ? AND airdate = ?",
|
||||
[showObj.indexerid, airdate])
|
||||
my_db = db.DBConnection()
|
||||
sql_results = my_db.select('SELECT season, episode FROM tv_episodes WHERE showid = ? AND airdate = ?',
|
||||
[show_obj.indexerid, airdate])
|
||||
|
||||
if len(sql_results) != 1:
|
||||
logger.log(
|
||||
u"Tried to look up the date for the episode " + title + " but the database didn't give proper results, skipping it",
|
||||
logger.WARNING)
|
||||
addCacheEntry = True
|
||||
if 1 != len(sql_results):
|
||||
logger.log(u'Tried to look up the date for the episode ' + title + ' but the database didn\'t'
|
||||
+ u' give proper results, skipping it', logger.WARNING)
|
||||
add_cache_entry = True
|
||||
|
||||
if not addCacheEntry:
|
||||
actual_season = int(sql_results[0]["season"])
|
||||
actual_episodes = [int(sql_results[0]["episode"])]
|
||||
if not add_cache_entry:
|
||||
actual_season = int(sql_results[0]['season'])
|
||||
actual_episodes = [int(sql_results[0]['episode'])]
|
||||
|
||||
# add parsed result to cache for usage later on
|
||||
if addCacheEntry:
|
||||
logger.log(u"Adding item from search to cache: " + title, logger.DEBUG)
|
||||
ci = self.cache._addCacheEntry(title, url, parse_result=parse_result)
|
||||
if ci is not None:
|
||||
if add_cache_entry:
|
||||
logger.log(u'Adding item from search to cache: ' + title, logger.DEBUG)
|
||||
ci = self.cache.add_cache_entry(title, url, parse_result=parse_result)
|
||||
if None is not ci:
|
||||
cl.append(ci)
|
||||
continue
|
||||
|
||||
# make sure we want the episode
|
||||
wantEp = True
|
||||
want_ep = True
|
||||
for epNo in actual_episodes:
|
||||
if not showObj.wantEpisode(actual_season, epNo, quality, manualSearch):
|
||||
wantEp = False
|
||||
if not show_obj.wantEpisode(actual_season, epNo, quality, manual_search):
|
||||
want_ep = False
|
||||
break
|
||||
|
||||
if not wantEp:
|
||||
logger.log(
|
||||
u"Ignoring result " + title + " because we don't want an episode that is " +
|
||||
Quality.qualityStrings[
|
||||
quality], logger.DEBUG)
|
||||
|
||||
if not want_ep:
|
||||
logger.log(u'Ignoring result %s because we don\'t want an episode that is %s'
|
||||
% (title, Quality.qualityStrings[quality]), logger.DEBUG)
|
||||
continue
|
||||
|
||||
logger.log(u"Found result " + title + " at " + url, logger.DEBUG)
|
||||
logger.log(u'Found result %s at %s' % (title, url), logger.DEBUG)
|
||||
|
||||
# make a result object
|
||||
epObj = []
|
||||
ep_obj = []
|
||||
for curEp in actual_episodes:
|
||||
epObj.append(showObj.getEpisode(actual_season, curEp))
|
||||
ep_obj.append(show_obj.getEpisode(actual_season, curEp))
|
||||
|
||||
result = self.getResult(epObj)
|
||||
result.show = showObj
|
||||
result.url = url
|
||||
result = self.get_result(ep_obj, url)
|
||||
if None is result:
|
||||
continue
|
||||
result.show = show_obj
|
||||
result.name = title
|
||||
result.quality = quality
|
||||
result.release_group = release_group
|
||||
result.content = None
|
||||
result.version = version
|
||||
|
||||
if len(epObj) == 1:
|
||||
epNum = epObj[0].episode
|
||||
logger.log(u"Single episode result.", logger.DEBUG)
|
||||
elif len(epObj) > 1:
|
||||
epNum = MULTI_EP_RESULT
|
||||
logger.log(u"Separating multi-episode result to check for later - result contains episodes: " + str(
|
||||
parse_result.episode_numbers), logger.DEBUG)
|
||||
elif len(epObj) == 0:
|
||||
epNum = SEASON_RESULT
|
||||
logger.log(u"Separating full season result to check for later", logger.DEBUG)
|
||||
if 1 == len(ep_obj):
|
||||
ep_num = ep_obj[0].episode
|
||||
logger.log(u'Single episode result.', logger.DEBUG)
|
||||
elif 1 < len(ep_obj):
|
||||
ep_num = MULTI_EP_RESULT
|
||||
logger.log(u'Separating multi-episode result to check for later - result contains episodes: '
|
||||
+ str(parse_result.episode_numbers), logger.DEBUG)
|
||||
elif 0 == len(ep_obj):
|
||||
ep_num = SEASON_RESULT
|
||||
logger.log(u'Separating full season result to check for later', logger.DEBUG)
|
||||
|
||||
if epNum not in results:
|
||||
results[epNum] = [result]
|
||||
if ep_num not in results:
|
||||
results[ep_num] = [result]
|
||||
else:
|
||||
results[epNum].append(result)
|
||||
results[ep_num].append(result)
|
||||
|
||||
# check if we have items to add to cache
|
||||
if len(cl) > 0:
|
||||
myDB = self.cache._getDB()
|
||||
myDB.mass_action(cl)
|
||||
if 0 < len(cl):
|
||||
my_db = self.cache.get_db()
|
||||
my_db.mass_action(cl)
|
||||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=None):
|
||||
def find_propers(self, search_date=None):
|
||||
|
||||
results = self.cache.listPropers(search_date)
|
||||
|
||||
return [classes.Proper(x['name'], x['url'], datetime.datetime.fromtimestamp(x['time']), self.show) for x in
|
||||
results]
|
||||
|
||||
def seedRatio(self):
|
||||
'''
|
||||
def seed_ratio(self):
|
||||
"""
|
||||
Provider should override this value if custom seed ratio enabled
|
||||
It should return the value of the provider seed ratio
|
||||
'''
|
||||
"""
|
||||
return ''
|
||||
|
||||
@staticmethod
|
||||
|
@ -489,6 +484,30 @@ class GenericProvider:
|
|||
logger.log(u'%s in response from %s' % (('No %s items' % mode,
|
||||
'%s %s item%s' % (count, mode, maybe_plural(count)))[0 < count], url))
|
||||
|
||||
def check_auth_cookie(self):
|
||||
|
||||
if hasattr(self, 'cookies'):
|
||||
cookies = self.cookies
|
||||
|
||||
if not re.match('^(\w+=\w+[;\s]*)+$', cookies):
|
||||
return False
|
||||
|
||||
cj = requests.utils.add_dict_to_cookiejar(self.session.cookies,
|
||||
dict([x.strip().split('=') for x in cookies.split(';')
|
||||
if x != ''])),
|
||||
for item in cj:
|
||||
if not isinstance(item, requests.cookies.RequestsCookieJar):
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
def _check_cookie(self):
|
||||
|
||||
if self.check_auth_cookie():
|
||||
return True, None
|
||||
|
||||
return False, 'Cookies not correctly formatted key=value pairs e.g. uid=xx;pass=yy)'
|
||||
|
||||
|
||||
class NZBProvider(GenericProvider):
|
||||
|
||||
|
@ -497,9 +516,30 @@ class NZBProvider(GenericProvider):
|
|||
|
||||
self.providerType = GenericProvider.NZB
|
||||
|
||||
def imageName(self):
|
||||
def image_name(self):
|
||||
|
||||
return GenericProvider.imageName(self, 'newznab')
|
||||
return GenericProvider.image_name(self, 'newznab')
|
||||
|
||||
def maybe_apikey(self):
|
||||
|
||||
if hasattr(self, 'needs_auth'):
|
||||
if hasattr(self, 'key') and 0 < len(self.key):
|
||||
return self.key
|
||||
if hasattr(self, 'api_key') and 0 < len(self.api_key):
|
||||
return self.api_key
|
||||
return None
|
||||
return False
|
||||
|
||||
def _check_auth(self):
|
||||
|
||||
has_key = self.maybe_apikey()
|
||||
if has_key:
|
||||
return has_key
|
||||
if None is has_key:
|
||||
raise AuthException('%s for %s is empty in config provider options'
|
||||
% ('API key' + ('', ' and/or Username')[hasattr(self, 'username')], self.name))
|
||||
|
||||
return GenericProvider._check_auth(self)
|
||||
|
||||
def _find_propers(self, search_date=None):
|
||||
|
||||
|
@ -508,7 +548,7 @@ class NZBProvider(GenericProvider):
|
|||
cache_results]
|
||||
|
||||
index = 0
|
||||
alt_search = ('nzbs_org' == self.getID())
|
||||
alt_search = ('nzbs_org' == self.get_id())
|
||||
term_items_found = False
|
||||
do_search_alt = False
|
||||
|
||||
|
@ -534,7 +574,7 @@ class NZBProvider(GenericProvider):
|
|||
else:
|
||||
index += 1
|
||||
|
||||
for item in self._doSearch(search_params, age=4):
|
||||
for item in self._do_search(search_params, age=4):
|
||||
|
||||
(title, url) = self._get_title_and_url(item)
|
||||
|
||||
|
@ -549,7 +589,7 @@ class NZBProvider(GenericProvider):
|
|||
logger.log(u'Unable to figure out the date for entry %s, skipping it', title)
|
||||
continue
|
||||
|
||||
if not search_date or result_date > search_date:
|
||||
if not search_date or search_date < result_date:
|
||||
search_result = classes.Proper(title, url, result_date, self.show)
|
||||
results.append(search_result)
|
||||
term_items_found = True
|
||||
|
@ -569,15 +609,15 @@ class TorrentProvider(GenericProvider):
|
|||
|
||||
self._seed_ratio = None
|
||||
|
||||
def imageName(self):
|
||||
def image_name(self):
|
||||
|
||||
return GenericProvider.imageName(self, 'torrent')
|
||||
return GenericProvider.image_name(self, 'torrent')
|
||||
|
||||
def seedRatio(self):
|
||||
def seed_ratio(self):
|
||||
|
||||
return self._seed_ratio
|
||||
|
||||
def getQuality(self, item, anime=False):
|
||||
def get_quality(self, item, anime=False):
|
||||
|
||||
if isinstance(item, tuple):
|
||||
name = item[0]
|
||||
|
@ -655,7 +695,7 @@ class TorrentProvider(GenericProvider):
|
|||
search_params += [crop.sub(r'\1', '%s %s' % (name, detail) + ('', ' ' + x)[any(x)]) for x in append]
|
||||
return search_params
|
||||
|
||||
def _checkAuth(self):
|
||||
def _check_auth(self):
|
||||
|
||||
if hasattr(self, 'username') and hasattr(self, 'password'):
|
||||
if self.username and self.password:
|
||||
|
@ -674,7 +714,7 @@ class TorrentProvider(GenericProvider):
|
|||
return True
|
||||
setting = 'Passkey'
|
||||
else:
|
||||
return GenericProvider._checkAuth(self)
|
||||
return GenericProvider._check_auth(self)
|
||||
|
||||
raise AuthException('%s for %s is empty in config provider options' % (setting, self.name))
|
||||
|
||||
|
@ -720,7 +760,7 @@ class TorrentProvider(GenericProvider):
|
|||
proper_check = re.compile(r'(?i)(?:%s)' % clean_term.sub('', proper_term))
|
||||
|
||||
search_string = self._get_episode_search_strings(cur_ep, add_string=proper_term)
|
||||
for item in self._doSearch(search_string[0]):
|
||||
for item in self._do_search(search_string[0]):
|
||||
title, url = self._get_title_and_url(item)
|
||||
if not proper_check.search(title):
|
||||
continue
|
||||
|
@ -741,4 +781,4 @@ class TorrentProvider(GenericProvider):
|
|||
def get_cache_data(self, *args, **kwargs):
|
||||
|
||||
search_params = {'Cache': ['']}
|
||||
return self._doSearch(search_params)
|
||||
return self._do_search(search_params)
|
||||
|
|
|
@ -44,13 +44,13 @@ class GFTrackerProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = GFTrackerCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'gft_uid' in self.session.cookies and 'gft_pass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
helpers.getURL(self.urls['login_get'], session=self.session)
|
||||
login_params = {'username': self.username, 'password': self.password}
|
||||
response = helpers.getURL(self.urls['login_post'], post_data=login_params, session=self.session)
|
||||
|
@ -61,16 +61,16 @@ class GFTrackerProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
||||
rc = dict((k, re.compile('(?i)' + v)) for (k, v) in {'info': 'details', 'get': 'download',
|
||||
'seeders': '(^\d+)', 'leechers': '(\d+)$'}.items())
|
||||
'seeders': r'(^\d+)', 'leechers': r'(\d+)$'}.items())
|
||||
for mode in search_params.keys():
|
||||
for search_string in search_params[mode]:
|
||||
|
||||
|
@ -81,7 +81,7 @@ class GFTrackerProvider(generic.TorrentProvider):
|
|||
if 'Cache' != mode:
|
||||
search_url += self.urls['search'] % search_string
|
||||
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -126,7 +126,7 @@ class GFTrackerProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -73,27 +73,27 @@ class HDBitsProvider(generic.TorrentProvider):
|
|||
|
||||
return title, url
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
self._checkAuth()
|
||||
self._check_auth()
|
||||
|
||||
logger.log(u'Search url: %s search_params: %s' % (self.urls['search'], search_params), logger.DEBUG)
|
||||
|
||||
response_json = self.getURL(self.urls['search'], post_data=search_params, json=True)
|
||||
response_json = self.get_url(self.urls['search'], post_data=search_params, json=True)
|
||||
if response_json and 'data' in response_json and self.check_auth_from_data(response_json):
|
||||
return response_json['data']
|
||||
|
||||
logger.log(u'Resulting JSON from %s isn\'t correct, not parsing it' % self.name, logger.ERROR)
|
||||
return []
|
||||
|
||||
def findPropers(self, search_date=None):
|
||||
def find_propers(self, search_date=None):
|
||||
|
||||
results = []
|
||||
|
||||
search_terms = [' proper ', ' repack ']
|
||||
|
||||
for term in search_terms:
|
||||
for item in self._doSearch(self._build_search_strings(search_term=term)):
|
||||
for item in self._do_search(self._build_search_strings(search_term=term)):
|
||||
if item['utadded']:
|
||||
try:
|
||||
result_date = datetime.datetime.fromtimestamp(int(item['utadded']))
|
||||
|
@ -142,9 +142,9 @@ class HDBitsProvider(generic.TorrentProvider):
|
|||
|
||||
def get_cache_data(self):
|
||||
|
||||
self._checkAuth()
|
||||
self._check_auth()
|
||||
|
||||
response_json = self.getURL(self.urls['search'], post_data=self._build_search_strings(), json=True)
|
||||
response_json = self.get_url(self.urls['search'], post_data=self._build_search_strings(), json=True)
|
||||
if response_json and 'data' in response_json and self.check_auth_from_data(response_json):
|
||||
return response_json['data']
|
||||
|
||||
|
|
|
@ -42,13 +42,13 @@ class IPTorrentsProvider(generic.TorrentProvider):
|
|||
self.freeleech = False
|
||||
self.cache = IPTorrentsCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'login': 'submit'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -58,10 +58,10 @@ class IPTorrentsProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -76,7 +76,7 @@ class IPTorrentsProvider(generic.TorrentProvider):
|
|||
search_string = unidecode(search_string)
|
||||
search_url = '%s%s' % (self.urls['search'] % (freeleech, search_string),
|
||||
(';o=seeders', '')['Cache' == mode])
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -120,7 +120,7 @@ class IPTorrentsProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date, '')
|
||||
|
||||
|
|
|
@ -55,7 +55,7 @@ class KATProvider(generic.TorrentProvider):
|
|||
|
||||
file_name = None
|
||||
|
||||
data = self.getURL(torrent_link)
|
||||
data = self.get_url(torrent_link)
|
||||
if not data:
|
||||
return None
|
||||
|
||||
|
@ -144,7 +144,7 @@ class KATProvider(generic.TorrentProvider):
|
|||
|
||||
return [{'Episode': self._build_search_strings(ep_detail, append=(add_string, '')[self.show.anime])}]
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -216,7 +216,7 @@ class KATProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date, '')
|
||||
|
||||
|
|
|
@ -44,13 +44,13 @@ class MoreThanProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = MoreThanCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'session' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'login': 'submit'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -63,10 +63,10 @@ class MoreThanProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -82,7 +82,7 @@ class MoreThanProvider(generic.TorrentProvider):
|
|||
search_url = self.urls['search'] % search_string
|
||||
|
||||
# fetches 15 results by default, and up to 100 if allowed in user profile
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
if not html or self._has_no_results(html):
|
||||
|
@ -133,7 +133,7 @@ class MoreThanProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -43,18 +43,10 @@ class NewznabProvider(generic.NZBProvider):
|
|||
self.default = False
|
||||
self.cache = NewznabCache(self)
|
||||
|
||||
def _checkAuth(self):
|
||||
|
||||
if self.needs_auth and not self.key:
|
||||
logger.log(u'Incorrect authentication credentials for %s : API key is missing' % self.name, logger.DEBUG)
|
||||
raise AuthException('Your authentication credentials for %s are missing, check your config.' % self.name)
|
||||
|
||||
return True
|
||||
|
||||
def check_auth_from_data(self, data):
|
||||
|
||||
if data is None:
|
||||
return self._checkAuth()
|
||||
return self._check_auth()
|
||||
|
||||
if 'error' in data.feed:
|
||||
code = data.feed['error']['code']
|
||||
|
@ -84,22 +76,20 @@ class NewznabProvider(generic.NZBProvider):
|
|||
"""
|
||||
return_categories = []
|
||||
|
||||
self._checkAuth()
|
||||
api_key = self._check_auth()
|
||||
|
||||
params = {'t': 'caps'}
|
||||
if self.needs_auth and self.key:
|
||||
params['apikey'] = self.key
|
||||
if isinstance(api_key, basestring):
|
||||
params['apikey'] = api_key
|
||||
|
||||
try:
|
||||
categories = self.getURL('%s/api' % self.url, params=params, timeout=10)
|
||||
except:
|
||||
logger.log(u'Error getting html for [%s]' %
|
||||
('%s/api?%s' % (self.url, '&'.join('%s=%s' % (x, y) for x, y in params.items()))), logger.DEBUG)
|
||||
categories = self.get_url('%s/api' % self.url, params=params, timeout=10)
|
||||
if not categories:
|
||||
logger.log(u'Error getting html for [%s/api?%s]' %
|
||||
(self.url, '&'.join('%s=%s' % (x, y) for x, y in params.items())), logger.DEBUG)
|
||||
return (False, return_categories, 'Error getting html for [%s]' %
|
||||
('%s/api?%s' % (self.url, '&'.join('%s=%s' % (x, y) for x, y in params.items()))))
|
||||
|
||||
xml_categories = helpers.parse_xml(categories)
|
||||
|
||||
if not xml_categories:
|
||||
logger.log(u'Error parsing xml for [%s]' % self.name, logger.DEBUG)
|
||||
return False, return_categories, 'Error parsing xml for [%s]' % self.name
|
||||
|
@ -117,7 +107,7 @@ class NewznabProvider(generic.NZBProvider):
|
|||
|
||||
def config_str(self):
|
||||
return '%s|%s|%s|%s|%i|%s|%i|%i|%i' \
|
||||
% (self.name or '', self.url or '', self.key or '', self.cat_ids or '', self.enabled,
|
||||
% (self.name or '', self.url or '', self.maybe_apikey() or '', self.cat_ids or '', self.enabled,
|
||||
self.search_mode or '', self.search_fallback, self.enable_recentsearch, self.enable_backlog)
|
||||
|
||||
def _get_season_search_strings(self, ep_obj):
|
||||
|
@ -154,7 +144,7 @@ class NewznabProvider(generic.NZBProvider):
|
|||
|
||||
return to_return
|
||||
|
||||
def _get_episode_search_strings(self, ep_obj, add_string=''):
|
||||
def _get_episode_search_strings(self, ep_obj):
|
||||
to_return = []
|
||||
params = {}
|
||||
|
||||
|
@ -169,8 +159,7 @@ class NewznabProvider(generic.NZBProvider):
|
|||
params['ep'] = '%i' % int(
|
||||
ep_obj.scene_absolute_number if int(ep_obj.scene_absolute_number) > 0 else ep_obj.scene_episode)
|
||||
else:
|
||||
params['season'] = ep_obj.scene_season
|
||||
params['ep'] = ep_obj.scene_episode
|
||||
params['season'], params['ep'] = ep_obj.scene_season, ep_obj.scene_episode
|
||||
|
||||
# search
|
||||
rid = helpers.mapIndexersToShow(ep_obj.show)[2]
|
||||
|
@ -203,9 +192,9 @@ class NewznabProvider(generic.NZBProvider):
|
|||
|
||||
return to_return
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
self._checkAuth()
|
||||
api_key = self._check_auth()
|
||||
|
||||
if 'rid' not in search_params and 'q' not in search_params:
|
||||
logger.log('Error no rid or search term given.')
|
||||
|
@ -233,8 +222,8 @@ class NewznabProvider(generic.NZBProvider):
|
|||
if search_params:
|
||||
params.update(search_params)
|
||||
|
||||
if self.needs_auth and self.key:
|
||||
params['apikey'] = self.key
|
||||
if isinstance(api_key, basestring):
|
||||
params['apikey'] = api_key
|
||||
|
||||
results = []
|
||||
offset = total = 0
|
||||
|
@ -288,7 +277,7 @@ class NewznabProvider(generic.NZBProvider):
|
|||
% (items, helpers.maybe_plural(items), params['limit']), logger.DEBUG)
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=None):
|
||||
def find_propers(self, search_date=None):
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
||||
|
@ -305,8 +294,9 @@ class NewznabCache(tvcache.TVCache):
|
|||
'cat': self.provider.cat_ids + ',5060,5070',
|
||||
'attrs': 'rageid'}
|
||||
|
||||
if self.provider.needs_auth and self.provider.key:
|
||||
params['apikey'] = self.provider.key
|
||||
has_apikey = self.provider.maybe_apikey()
|
||||
if has_apikey:
|
||||
params['apikey'] = has_apikey
|
||||
|
||||
rss_url = '%sapi?%s' % (self.provider.url, urllib.urlencode(params))
|
||||
|
||||
|
@ -314,13 +304,14 @@ class NewznabCache(tvcache.TVCache):
|
|||
|
||||
return self.getRSSFeed(rss_url)
|
||||
|
||||
def _checkAuth(self, *data):
|
||||
|
||||
return self.provider.check_auth_from_data(data[0])
|
||||
|
||||
def updateCache(self):
|
||||
|
||||
if self.shouldUpdate() and self._checkAuth(None):
|
||||
if self.shouldUpdate():
|
||||
try:
|
||||
self._checkAuth()
|
||||
except Exception:
|
||||
return []
|
||||
|
||||
data = self._getRSSData()
|
||||
|
||||
# as long as the http request worked we count this as an update
|
||||
|
@ -332,7 +323,7 @@ class NewznabCache(tvcache.TVCache):
|
|||
|
||||
self.setLastUpdate()
|
||||
|
||||
if self._checkAuth(data):
|
||||
if self.provider.check_auth_from_data(data):
|
||||
items = data.entries
|
||||
cl = []
|
||||
for item in items:
|
||||
|
@ -341,7 +332,7 @@ class NewznabCache(tvcache.TVCache):
|
|||
cl.append(ci)
|
||||
|
||||
if 0 < len(cl):
|
||||
my_db = self._getDB()
|
||||
my_db = self.get_db()
|
||||
my_db.mass_action(cl)
|
||||
|
||||
else:
|
||||
|
@ -376,4 +367,4 @@ class NewznabCache(tvcache.TVCache):
|
|||
url = self._translateLinkURL(url)
|
||||
|
||||
logger.log(u'Attempting to add item from RSS to cache: ' + title, logger.DEBUG)
|
||||
return self._addCacheEntry(title, url, indexer_id=tvrageid)
|
||||
return self.add_cache_entry(title, url, indexer_id=tvrageid)
|
||||
|
|
|
@ -17,113 +17,82 @@
|
|||
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
import urllib
|
||||
import re
|
||||
|
||||
import generic
|
||||
from . import generic
|
||||
from sickbeard import logger, tvcache, show_name_helpers
|
||||
from sickbeard.common import Quality
|
||||
|
||||
|
||||
class NyaaProvider(generic.TorrentProvider):
|
||||
|
||||
def __init__(self):
|
||||
generic.TorrentProvider.__init__(self, 'NyaaTorrents', True, True)
|
||||
self.ratio = None
|
||||
self.cache = NyaaCache(self)
|
||||
generic.TorrentProvider.__init__(self, 'NyaaTorrents', anime_only=True)
|
||||
|
||||
self.url = 'http://www.nyaa.se/'
|
||||
|
||||
def getQuality(self, item, anime=False):
|
||||
title = item.title
|
||||
quality = Quality.sceneQuality(title, anime)
|
||||
return quality
|
||||
self.cache = NyaaCache(self)
|
||||
|
||||
def findSearchResults(self, show, episodes, search_mode, manualSearch=False):
|
||||
return generic.TorrentProvider.findSearchResults(self, show, episodes, search_mode, manualSearch)
|
||||
def _do_search(self, search_string, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
def _get_season_search_strings(self, ep_obj):
|
||||
return show_name_helpers.makeSceneShowSearchStrings(self.show)
|
||||
|
||||
def _get_episode_search_strings(self, ep_obj, add_string=''):
|
||||
return self._get_season_search_strings(ep_obj)
|
||||
|
||||
def _doSearch(self, search_string, search_mode='eponly', epcount=0, age=0):
|
||||
results = []
|
||||
if self.show and not self.show.is_anime:
|
||||
logger.log(u"" + str(self.show.name) + " is not an anime skipping " + str(self.name))
|
||||
return []
|
||||
return results
|
||||
|
||||
params = {
|
||||
"term": search_string.encode('utf-8'),
|
||||
"cats": '1_37', # Limit to English-translated Anime (for now)
|
||||
"sort": '2', # Sort Descending By Seeders
|
||||
}
|
||||
params = {'term': search_string.encode('utf-8'),
|
||||
'cats': '1_37', # Limit to English-translated Anime (for now)
|
||||
# 'sort': '2', # Sort Descending By Seeders
|
||||
}
|
||||
|
||||
searchURL = self.url + '?page=rss&' + urllib.urlencode(params)
|
||||
search_url = self.url + '?page=rss&' + urllib.urlencode(params)
|
||||
|
||||
logger.log(u"Search string: " + searchURL, logger.DEBUG)
|
||||
logger.log(u'Search string: ' + search_url, logger.DEBUG)
|
||||
|
||||
data = self.cache.getRSSFeed(searchURL)
|
||||
if not data:
|
||||
return []
|
||||
|
||||
if 'entries' in data:
|
||||
data = self.cache.getRSSFeed(search_url)
|
||||
if data and 'entries' in data:
|
||||
items = data.entries
|
||||
|
||||
results = []
|
||||
|
||||
for curItem in items:
|
||||
|
||||
(title, url) = self._get_title_and_url(curItem)
|
||||
title, url = self._get_title_and_url(curItem)
|
||||
|
||||
if title and url:
|
||||
results.append(curItem)
|
||||
else:
|
||||
logger.log(
|
||||
u"The data returned from " + self.name + " is incomplete, this result is unusable",
|
||||
logger.DEBUG)
|
||||
logger.log(u'The data returned from ' + self.name + ' is incomplete, this result is unusable',
|
||||
logger.DEBUG)
|
||||
|
||||
return results
|
||||
return results
|
||||
|
||||
return []
|
||||
def find_search_results(self, show, episodes, search_mode, manual_search=False):
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
return generic.TorrentProvider.find_search_results(self, show, episodes, search_mode, manual_search)
|
||||
|
||||
return generic.TorrentProvider._get_title_and_url(self, item)
|
||||
def _get_season_search_strings(self, ep_obj, **kwargs):
|
||||
|
||||
def _extract_name_from_filename(self, filename):
|
||||
name_regex = '(.*?)\.?(\[.*]|\d+\.TPB)\.torrent$'
|
||||
logger.log(u"Comparing " + name_regex + " against " + filename, logger.DEBUG)
|
||||
match = re.match(name_regex, filename, re.I)
|
||||
if match:
|
||||
return match.group(1)
|
||||
return None
|
||||
return show_name_helpers.makeSceneShowSearchStrings(self.show)
|
||||
|
||||
def seedRatio(self):
|
||||
return self.ratio
|
||||
def _get_episode_search_strings(self, ep_obj, **kwargs):
|
||||
|
||||
return self._get_season_search_strings(ep_obj)
|
||||
|
||||
|
||||
class NyaaCache(tvcache.TVCache):
|
||||
def __init__(self, provider):
|
||||
tvcache.TVCache.__init__(self, provider)
|
||||
|
||||
# only poll NyaaTorrents every 15 minutes max
|
||||
self.minTime = 15
|
||||
def __init__(self, this_provider):
|
||||
tvcache.TVCache.__init__(self, this_provider)
|
||||
|
||||
self.minTime = 15 # cache update frequency
|
||||
|
||||
def _getRSSData(self):
|
||||
params = {
|
||||
"page": 'rss', # Use RSS page
|
||||
"order": '1', # Sort Descending By Date
|
||||
"cats": '1_37', # Limit to English-translated Anime (for now)
|
||||
}
|
||||
params = {'page': 'rss', # Use RSS page
|
||||
'order': '1', # Sort Descending By Date
|
||||
'cats': '1_37'} # Limit to English-translated Anime (for now)
|
||||
|
||||
url = self.provider.url + '?' + urllib.urlencode(params)
|
||||
|
||||
logger.log(u"NyaaTorrents cache update URL: " + url, logger.DEBUG)
|
||||
logger.log(u'NyaaTorrents cache update URL: ' + url, logger.DEBUG)
|
||||
|
||||
data = self.getRSSFeed(url)
|
||||
|
||||
if data and 'entries' in data:
|
||||
return data.entries
|
||||
else:
|
||||
return []
|
||||
return []
|
||||
|
||||
|
||||
provider = NyaaProvider()
|
||||
|
|
|
@ -16,117 +16,202 @@
|
|||
# You should have received a copy of the GNU General Public License
|
||||
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
import urllib
|
||||
import re
|
||||
from datetime import datetime
|
||||
import time
|
||||
import traceback
|
||||
|
||||
import generic
|
||||
import sickbeard
|
||||
import urllib
|
||||
from sickbeard import tvcache, classes, logger, show_name_helpers
|
||||
from sickbeard.exceptions import AuthException
|
||||
|
||||
|
||||
try:
|
||||
import xml.etree.cElementTree as etree
|
||||
except ImportError:
|
||||
import elementtree.ElementTree as etree
|
||||
|
||||
try:
|
||||
import json
|
||||
except ImportError:
|
||||
from lib import simplejson as json
|
||||
from sickbeard.rssfeeds import RSSFeeds
|
||||
from sickbeard.bs4_parser import BS4Parser
|
||||
|
||||
|
||||
class OmgwtfnzbsProvider(generic.NZBProvider):
|
||||
|
||||
def __init__(self):
|
||||
generic.NZBProvider.__init__(self, 'omgwtfnzbs', True, False)
|
||||
self.username = None
|
||||
self.api_key = None
|
||||
self.cache = OmgwtfnzbsCache(self)
|
||||
generic.NZBProvider.__init__(self, 'omgwtfnzbs')
|
||||
|
||||
self.url = 'https://omgwtfnzbs.org/'
|
||||
|
||||
def _checkAuth(self):
|
||||
self.url_base = 'https://omgwtfnzbs.org/'
|
||||
self.urls = {'config_provider_home_uri': self.url_base,
|
||||
'cache': 'https://rss.omgwtfnzbs.org/rss-download.php?%s',
|
||||
'search': self.url_base + 'json/?%s',
|
||||
'get': self.url_base + '%s',
|
||||
'cache_html': self.url_base + 'browse.php?cat=tv%s',
|
||||
'search_html': self.url_base + 'browse.php?cat=tv&search=%s'}
|
||||
|
||||
if not self.username or not self.api_key:
|
||||
raise AuthException("Your authentication credentials for " + self.name + " are missing, check your config.")
|
||||
self.url = self.urls['config_provider_home_uri']
|
||||
|
||||
return True
|
||||
self.needs_auth = True
|
||||
self.username, self.api_key, self.cookies = 3 * [None]
|
||||
self.cache = OmgwtfnzbsCache(self)
|
||||
|
||||
def _checkAuthFromData(self, parsed_data, is_XML=True):
|
||||
def _check_auth_from_data(self, parsed_data, is_xml=True):
|
||||
|
||||
if parsed_data is None:
|
||||
return self._checkAuth()
|
||||
return self._check_auth()
|
||||
|
||||
if is_XML:
|
||||
if is_xml:
|
||||
# provider doesn't return xml on error
|
||||
return True
|
||||
else:
|
||||
parsedJSON = parsed_data
|
||||
data_json = parsed_data
|
||||
|
||||
if 'notice' in parsedJSON:
|
||||
description_text = parsedJSON.get('notice')
|
||||
if 'notice' in data_json:
|
||||
description_text = data_json.get('notice')
|
||||
|
||||
if 'information is incorrect' in parsedJSON.get('notice'):
|
||||
logger.log(u"Incorrect authentication credentials for " + self.name + " : " + str(description_text),
|
||||
if 'information is incorrect' in data_json.get('notice'):
|
||||
logger.log(u'Incorrect authentication credentials for ' + self.name + ' : ' + str(description_text),
|
||||
logger.DEBUG)
|
||||
raise AuthException(
|
||||
"Your authentication credentials for " + self.name + " are incorrect, check your config.")
|
||||
'Your authentication credentials for ' + self.name + ' are incorrect, check your config.')
|
||||
|
||||
elif '0 results matched your terms' in parsedJSON.get('notice'):
|
||||
elif '0 results matched your terms' in data_json.get('notice'):
|
||||
return True
|
||||
|
||||
else:
|
||||
logger.log(u"Unknown error given from " + self.name + " : " + str(description_text), logger.DEBUG)
|
||||
logger.log(u'Unknown error given from ' + self.name + ' : ' + str(description_text), logger.DEBUG)
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
def _get_season_search_strings(self, ep_obj):
|
||||
|
||||
return [x for x in show_name_helpers.makeSceneSeasonSearchString(self.show, ep_obj)]
|
||||
|
||||
def _get_episode_search_strings(self, ep_obj, add_string=''):
|
||||
def _get_episode_search_strings(self, ep_obj):
|
||||
|
||||
return [x for x in show_name_helpers.makeSceneSearchString(self.show, ep_obj)]
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
return (item['release'], item['getnzb'])
|
||||
|
||||
def _doSearch(self, search, search_mode='eponly', epcount=0, retention=0):
|
||||
return item['release'], item['getnzb']
|
||||
|
||||
self._checkAuth()
|
||||
def get_result(self, episodes, url):
|
||||
|
||||
params = {'user': self.username,
|
||||
'api': self.api_key,
|
||||
'eng': 1,
|
||||
'catid': '19,20', # SD,HD
|
||||
'retention': sickbeard.USENET_RETENTION,
|
||||
'search': search}
|
||||
result = None
|
||||
if url and False is self._init_api():
|
||||
data = self.get_url(url)
|
||||
if data:
|
||||
if '</nzb>' not in data or 'seem to be logged in' in data:
|
||||
logger.log(u'Failed nzb data response: %s' % data, logger.DEBUG)
|
||||
return result
|
||||
result = classes.NZBDataSearchResult(episodes)
|
||||
result.extraInfo += [data]
|
||||
|
||||
if retention or not params['retention']:
|
||||
params['retention'] = retention
|
||||
if None is result:
|
||||
result = classes.NZBSearchResult(episodes)
|
||||
result.url = url
|
||||
|
||||
search_url = 'https://api.omgwtfnzbs.org/json/?' + urllib.urlencode(params)
|
||||
logger.log(u"Search url: " + search_url, logger.DEBUG)
|
||||
result.provider = self
|
||||
|
||||
parsedJSON = self.getURL(search_url, json=True)
|
||||
if not parsedJSON:
|
||||
return []
|
||||
return result
|
||||
|
||||
if self._checkAuthFromData(parsedJSON, is_XML=False):
|
||||
results = []
|
||||
def get_cache_data(self):
|
||||
|
||||
for item in parsedJSON:
|
||||
if 'release' in item and 'getnzb' in item:
|
||||
results.append(item)
|
||||
api_key = self._init_api()
|
||||
if False is api_key:
|
||||
return self.search_html()
|
||||
if None is not api_key:
|
||||
params = {'user': self.username,
|
||||
'api': api_key,
|
||||
'eng': 1,
|
||||
'catid': '19,20'} # SD,HD
|
||||
|
||||
return results
|
||||
rss_url = self.urls['cache'] % urllib.urlencode(params)
|
||||
|
||||
logger.log(self.name + u' cache update URL: ' + rss_url, logger.DEBUG)
|
||||
|
||||
data = RSSFeeds(self).get_feed(rss_url)
|
||||
if data and 'entries' in data:
|
||||
return data.entries
|
||||
return []
|
||||
|
||||
def findPropers(self, search_date=None):
|
||||
def _do_search(self, search, search_mode='eponly', epcount=0, retention=0):
|
||||
|
||||
api_key = self._init_api()
|
||||
if False is api_key:
|
||||
return self.search_html(search)
|
||||
results = []
|
||||
if None is not api_key:
|
||||
params = {'user': self.username,
|
||||
'api': api_key,
|
||||
'eng': 1,
|
||||
'catid': '19,20', # SD,HD
|
||||
'retention': (sickbeard.USENET_RETENTION, retention)[retention or not sickbeard.USENET_RETENTION],
|
||||
'search': search}
|
||||
|
||||
search_url = self.urls['search'] % urllib.urlencode(params)
|
||||
logger.log(u'Search url: ' + search_url, logger.DEBUG)
|
||||
|
||||
data_json = self.get_url(search_url, json=True)
|
||||
if data_json and self._check_auth_from_data(data_json, is_xml=False):
|
||||
for item in data_json:
|
||||
if 'release' in item and 'getnzb' in item:
|
||||
results.append(item)
|
||||
return results
|
||||
|
||||
def search_html(self, search=''):
|
||||
|
||||
results = []
|
||||
if None is self.cookies:
|
||||
return results
|
||||
|
||||
rc = dict((k, re.compile('(?i)' + v)) for (k, v) in {'info': 'detail', 'get': r'send\?', 'nuked': r'\bnuked',
|
||||
'cat': 'cat=(?:19|20)'}.items())
|
||||
mode = ('search', 'cache')['' == search]
|
||||
search_url = self.urls[mode + '_html'] % search
|
||||
html = self.get_url(search_url)
|
||||
cnt = len(results)
|
||||
try:
|
||||
if not html:
|
||||
raise generic.HaltParseException
|
||||
|
||||
with BS4Parser(html, features=['html5lib', 'permissive']) as soup:
|
||||
torrent_table = soup.find('table', attrs={'id': 'table_table'})
|
||||
torrent_rows = []
|
||||
if torrent_table:
|
||||
torrent_rows = torrent_table.find('tbody').find_all('tr')
|
||||
|
||||
if 1 > len(torrent_rows):
|
||||
raise generic.HaltParseException
|
||||
|
||||
for tr in torrent_rows:
|
||||
try:
|
||||
if tr.find('img', src=rc['nuked']) or not tr.find('a', href=rc['cat']):
|
||||
continue
|
||||
|
||||
title = tr.find('a', href=rc['info'])['title']
|
||||
download_url = tr.find('a', href=rc['get'])
|
||||
age = tr.find_all('td')[-1]['data-sort']
|
||||
except (AttributeError, TypeError):
|
||||
continue
|
||||
|
||||
if title and download_url and age:
|
||||
results.append({'release': title, 'getnzb': self.urls['get'] % download_url['href'].lstrip('/'),
|
||||
'usenetage': int(age.strip())})
|
||||
|
||||
except generic.HaltParseException:
|
||||
time.sleep(1.1)
|
||||
pass
|
||||
except Exception:
|
||||
logger.log(u'Failed to parse. Traceback: %s' % traceback.format_exc(), logger.ERROR)
|
||||
|
||||
self._log_result(mode, len(results) - cnt, search_url)
|
||||
return results
|
||||
|
||||
def find_propers(self, search_date=None):
|
||||
|
||||
search_terms = ['.PROPER.', '.REPACK.']
|
||||
results = []
|
||||
|
||||
for term in search_terms:
|
||||
for item in self._doSearch(term, retention=4):
|
||||
for item in self._do_search(term, retention=4):
|
||||
if 'usenetage' in item:
|
||||
|
||||
title, url = self._get_title_and_url(item)
|
||||
|
@ -140,47 +225,40 @@ class OmgwtfnzbsProvider(generic.NZBProvider):
|
|||
|
||||
return results
|
||||
|
||||
def _init_api(self):
|
||||
|
||||
try:
|
||||
api_key = self._check_auth()
|
||||
if not api_key.startswith('cookie:'):
|
||||
return api_key
|
||||
except Exception:
|
||||
return None
|
||||
|
||||
self.cookies = re.sub(r'(?i)([\s\']+|cookie\s*:)', '', api_key)
|
||||
success, msg = self._check_cookie()
|
||||
if not success:
|
||||
logger.log(u'%s: %s' % (msg, self.cookies), logger.WARNING)
|
||||
self.cookies = None
|
||||
return None
|
||||
return False
|
||||
|
||||
@staticmethod
|
||||
def ui_string(key):
|
||||
result = ''
|
||||
if 'omgwtfnzbs_api_key' == key:
|
||||
result = 'Or use... \'cookie: cookname=xx; cookpass=yy\''
|
||||
return result
|
||||
|
||||
|
||||
class OmgwtfnzbsCache(tvcache.TVCache):
|
||||
def __init__(self, provider):
|
||||
tvcache.TVCache.__init__(self, provider)
|
||||
|
||||
def __init__(self, this_provider):
|
||||
tvcache.TVCache.__init__(self, this_provider)
|
||||
|
||||
self.minTime = 20
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
"""
|
||||
Retrieves the title and URL data from the item XML node
|
||||
|
||||
item: An elementtree.ElementTree element representing the <item> tag of the RSS feed
|
||||
|
||||
Returns: A tuple containing two strings representing title and URL respectively
|
||||
"""
|
||||
|
||||
title = item.title if item.title else None
|
||||
if title:
|
||||
title = u'' + title
|
||||
title = title.replace(' ', '.')
|
||||
|
||||
url = item.link if item.link else None
|
||||
if url:
|
||||
url = url.replace('&', '&')
|
||||
|
||||
return (title, url)
|
||||
|
||||
def _getRSSData(self):
|
||||
params = {'user': provider.username,
|
||||
'api': provider.api_key,
|
||||
'eng': 1,
|
||||
'catid': '19,20'} # SD,HD
|
||||
|
||||
rss_url = 'https://rss.omgwtfnzbs.org/rss-download.php?' + urllib.urlencode(params)
|
||||
|
||||
logger.log(self.provider.name + u" cache update URL: " + rss_url, logger.DEBUG)
|
||||
|
||||
data = self.getRSSFeed(rss_url)
|
||||
|
||||
if data and 'entries' in data:
|
||||
return data.entries
|
||||
else:
|
||||
return []
|
||||
return self.provider.get_cache_data()
|
||||
|
||||
provider = OmgwtfnzbsProvider()
|
||||
|
|
|
@ -39,14 +39,14 @@ class PiSexyProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = PiSexyCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies and\
|
||||
'pcode' in self.session.cookies and 'pisexy' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -59,24 +59,24 @@ class PiSexyProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
||||
rc = dict((k, re.compile('(?i)' + v))
|
||||
for (k, v) in {'info': 'download', 'get': 'download', 'valid_cat': 'cat=(?:0|50[12])',
|
||||
'title': 'Download\s([^\s]+).*', 'seeders': '(^\d+)', 'leechers': '(\d+)$'}.items())
|
||||
'title': r'Download\s([^\s]+).*', 'seeders': r'(^\d+)', 'leechers': r'(\d+)$'}.items())
|
||||
for mode in search_params.keys():
|
||||
for search_string in search_params[mode]:
|
||||
if isinstance(search_string, unicode):
|
||||
search_string = unidecode(search_string)
|
||||
|
||||
search_url = self.urls['search'] % search_string
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -129,7 +129,7 @@ class PiSexyProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -55,7 +55,7 @@ class RarbgProvider(generic.TorrentProvider):
|
|||
self.request_throttle = datetime.datetime.now()
|
||||
self.cache = RarbgCache(self)
|
||||
|
||||
def _doLogin(self, reset=False):
|
||||
def _do_login(self, reset=False):
|
||||
|
||||
if not reset and self.token and self.token_expiry and datetime.datetime.now() < self.token_expiry:
|
||||
return True
|
||||
|
@ -69,10 +69,10 @@ class RarbgProvider(generic.TorrentProvider):
|
|||
logger.log(u'No usable API token returned from: %s' % self.urls['api_token'], logger.ERROR)
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin(reset=True):
|
||||
if not self._do_login(reset=True):
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -122,7 +122,7 @@ class RarbgProvider(generic.TorrentProvider):
|
|||
time_out += 1
|
||||
time.sleep(1)
|
||||
|
||||
data = self.getURL(search_url % {'ranked': int(self.confirmed), 'token': self.token}, json=True)
|
||||
data = self.get_url(search_url % {'ranked': int(self.confirmed), 'token': self.token}, json=True)
|
||||
self.token_expiry = datetime.datetime.now() + datetime.timedelta(minutes=14)
|
||||
self.request_throttle = datetime.datetime.now() + datetime.timedelta(seconds=3)
|
||||
|
||||
|
@ -131,7 +131,7 @@ class RarbgProvider(generic.TorrentProvider):
|
|||
continue
|
||||
|
||||
elif 2 == data['error_code']: # Invalid token set
|
||||
if self._doLogin(reset=True):
|
||||
if self._do_login(reset=True):
|
||||
continue
|
||||
self._log_result(mode_base, len(items[mode_base]) - cnt, search_url)
|
||||
return results
|
||||
|
@ -160,7 +160,7 @@ class RarbgProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date, '{{.proper.|.repack.}}')
|
||||
|
||||
|
|
|
@ -41,9 +41,9 @@ class TorrentRssProvider(generic.TorrentProvider):
|
|||
self.feeder = RSSFeeds(self)
|
||||
self.cache = TorrentRssCache(self)
|
||||
|
||||
def imageName(self):
|
||||
def image_name(self):
|
||||
|
||||
return generic.GenericProvider.imageName(self, 'torrentrss')
|
||||
return generic.GenericProvider.image_name(self, 'torrentrss')
|
||||
|
||||
def config_str(self):
|
||||
return '%s|%s|%s|%d|%s|%d|%d|%d' % (self.name or '',
|
||||
|
@ -81,9 +81,9 @@ class TorrentRssProvider(generic.TorrentProvider):
|
|||
|
||||
def validate_feed(self):
|
||||
|
||||
succ, err_msg = self.feeder.check_cookie()
|
||||
if not succ:
|
||||
return succ, err_msg
|
||||
success, err_msg = self._check_cookie()
|
||||
if not success:
|
||||
return success, err_msg
|
||||
|
||||
try:
|
||||
items = self.get_cache_data()
|
||||
|
@ -96,7 +96,7 @@ class TorrentRssProvider(generic.TorrentProvider):
|
|||
if re.search('urn:btih:([0-9a-f]{32,40})', url):
|
||||
break
|
||||
else:
|
||||
torrent_file = self.getURL(url)
|
||||
torrent_file = self.get_url(url)
|
||||
try:
|
||||
bdecode(torrent_file)
|
||||
break
|
||||
|
|
|
@ -44,13 +44,13 @@ class SCCProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = SCCCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'submit': 'come on in'}
|
||||
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
|
@ -61,12 +61,12 @@ class SCCProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
rc = dict((k, re.compile('(?i)' + v)) for (k, v) in {'info': 'detail', 'get': 'download'}.items())
|
||||
|
@ -83,7 +83,7 @@ class SCCProvider(generic.TorrentProvider):
|
|||
self.urls['nonscene'] % search_string]
|
||||
|
||||
for search_url in searches:
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -128,7 +128,7 @@ class SCCProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -44,13 +44,13 @@ class SpeedCDProvider(generic.TorrentProvider):
|
|||
self.freeleech = False
|
||||
self.cache = SpeedCDCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'inSpeed_speedian' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -63,10 +63,10 @@ class SpeedCDProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -78,7 +78,7 @@ class SpeedCDProvider(generic.TorrentProvider):
|
|||
post_data = dict({'/browse.php?': None, 'cata': 'yes', 'jxt': 4, 'jxw': 'b', 'search': search_string},
|
||||
**self.categories[mode])
|
||||
|
||||
data_json = self.getURL(self.urls['search'], post_data=post_data, json=True)
|
||||
data_json = self.get_url(self.urls['search'], post_data=post_data, json=True)
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
if not data_json:
|
||||
|
@ -111,7 +111,7 @@ class SpeedCDProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@
|
|||
# You should have received a copy of the GNU General Public License
|
||||
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
import re
|
||||
import datetime
|
||||
|
||||
from . import generic
|
||||
|
@ -33,17 +34,17 @@ class StrikeProvider(generic.TorrentProvider):
|
|||
|
||||
self.minseed, self.minleech = 2 * [None]
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._checkAuth():
|
||||
if not self._check_auth():
|
||||
return results
|
||||
|
||||
for mode in search_params.keys():
|
||||
for search_string in search_params[mode]:
|
||||
|
||||
search_url = self.urls['search'] % search_string.replace(' ', '+')
|
||||
data_json = self.getURL(search_url, json=True)
|
||||
search_url = self.urls['search'] % re.sub('[\.\s]+', ' ', search_string)
|
||||
data_json = self.get_url(search_url, json=True)
|
||||
|
||||
cnt = len(results)
|
||||
try:
|
||||
|
@ -66,7 +67,7 @@ class StrikeProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -36,8 +36,8 @@ class ThePirateBayProvider(generic.TorrentProvider):
|
|||
def __init__(self):
|
||||
generic.TorrentProvider.__init__(self, 'The Pirate Bay')
|
||||
|
||||
self.urls = {'config_provider_home_uri': ['https://thepiratebay.gd/', 'https://thepiratebay.mn/',
|
||||
'https://thepiratebay.am/', 'https://thepiratebay.vg/',
|
||||
self.urls = {'config_provider_home_uri': ['https://thepiratebay.se/', 'https://thepiratebay.gd/',
|
||||
'https://thepiratebay.mn/', 'https://thepiratebay.vg/',
|
||||
'https://thepiratebay.la/'],
|
||||
'search': 'search/%s/0/7/200',
|
||||
'cache': 'tv/latest/'} # order by seed
|
||||
|
@ -57,7 +57,7 @@ class ThePirateBayProvider(generic.TorrentProvider):
|
|||
has_signature = False
|
||||
details_url = '/ajax_details_filelist.php?id=%s' % torrent_id
|
||||
for idx, url in enumerate(self.urls['config_provider_home_uri']):
|
||||
data = self.getURL(url + details_url)
|
||||
data = self.get_url(url + details_url)
|
||||
if data and re.search(r'<title>The\sPirate\sBay', data[33:200:]):
|
||||
has_signature = True
|
||||
break
|
||||
|
@ -144,13 +144,13 @@ class ThePirateBayProvider(generic.TorrentProvider):
|
|||
|
||||
return [{'Episode': self._build_search_strings(ep_detail, append=(add_string, '')[self.show.anime])}]
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
||||
rc = dict((k, re.compile('(?i)' + v))
|
||||
for (k, v) in {'info': 'detail', 'get': 'download[^"]+magnet', 'tid': '.*/(\d{5,}).*',
|
||||
for (k, v) in {'info': 'detail', 'get': 'download[^"]+magnet', 'tid': r'.*/(\d{5,}).*',
|
||||
'verify': '(?:helper|moderator|trusted|vip)'}.items())
|
||||
has_signature = False
|
||||
for mode in search_params.keys():
|
||||
|
@ -165,7 +165,7 @@ class ThePirateBayProvider(generic.TorrentProvider):
|
|||
|
||||
log_url = u'(%s/%s): %s' % (idx + 1, len(self.urls['config_provider_home_uri']), search_url)
|
||||
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
if html and re.search(r'Pirate\sBay', html[33:7632:]):
|
||||
has_signature = True
|
||||
break
|
||||
|
@ -227,7 +227,7 @@ class ThePirateBayProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date, '')
|
||||
|
||||
|
|
|
@ -16,140 +16,86 @@
|
|||
# You should have received a copy of the GNU General Public License
|
||||
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
import urllib
|
||||
import traceback
|
||||
import urllib
|
||||
|
||||
import generic
|
||||
from . import generic
|
||||
from sickbeard import logger, tvcache, show_name_helpers
|
||||
from sickbeard.common import Quality
|
||||
from sickbeard.bs4_parser import BS4Parser
|
||||
|
||||
|
||||
class TokyoToshokanProvider(generic.TorrentProvider):
|
||||
|
||||
def __init__(self):
|
||||
generic.TorrentProvider.__init__(self, 'TokyoToshokan', True, True)
|
||||
self.ratio = None
|
||||
self.cache = TokyoToshokanCache(self)
|
||||
generic.TorrentProvider.__init__(self, 'TokyoToshokan', anime_only=True)
|
||||
|
||||
self.url = 'http://tokyotosho.info/'
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
self.cache = TokyoToshokanCache(self)
|
||||
|
||||
title, url = item
|
||||
|
||||
if title:
|
||||
title = u'' + title
|
||||
title = title.replace(' ', '.')
|
||||
|
||||
if url:
|
||||
url = url.replace('&', '&')
|
||||
|
||||
return (title, url)
|
||||
|
||||
def seedRatio(self):
|
||||
return self.ratio
|
||||
|
||||
def getQuality(self, item, anime=False):
|
||||
quality = Quality.sceneQuality(item[0], anime)
|
||||
return quality
|
||||
|
||||
def findSearchResults(self, show, episodes, search_mode, manualSearch=False):
|
||||
return generic.TorrentProvider.findSearchResults(self, show, episodes, search_mode, manualSearch)
|
||||
|
||||
def _get_season_search_strings(self, ep_obj):
|
||||
return [x.replace('.', ' ') for x in show_name_helpers.makeSceneSeasonSearchString(self.show, ep_obj)]
|
||||
|
||||
def _get_episode_search_strings(self, ep_obj, add_string=''):
|
||||
return [x.replace('.', ' ') for x in show_name_helpers.makeSceneSearchString(self.show, ep_obj)]
|
||||
|
||||
def _doSearch(self, search_string, search_mode='eponly', epcount=0, age=0):
|
||||
if self.show and not self.show.is_anime:
|
||||
logger.log(u"" + str(self.show.name) + " is not an anime skipping " + str(self.name))
|
||||
return []
|
||||
|
||||
params = {
|
||||
"terms": search_string.encode('utf-8'),
|
||||
"type": 1, # get anime types
|
||||
}
|
||||
|
||||
searchURL = self.url + 'search.php?' + urllib.urlencode(params)
|
||||
|
||||
data = self.getURL(searchURL)
|
||||
|
||||
logger.log(u"Search string: " + searchURL, logger.DEBUG)
|
||||
|
||||
if not data:
|
||||
return []
|
||||
def _do_search(self, search_string, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
try:
|
||||
with BS4Parser(data, features=["html5lib", "permissive"]) as soup:
|
||||
torrent_table = soup.find('table', attrs={'class': 'listing'})
|
||||
torrent_rows = torrent_table.find_all('tr') if torrent_table else []
|
||||
if torrent_rows:
|
||||
if torrent_rows[0].find('td', attrs={'class': 'centertext'}):
|
||||
a = 1
|
||||
else:
|
||||
a = 0
|
||||
if self.show and not self.show.is_anime:
|
||||
return results
|
||||
|
||||
for top, bottom in zip(torrent_rows[a::2], torrent_rows[a::2]):
|
||||
title = top.find('td', attrs={'class': 'desc-top'}).text
|
||||
url = top.find('td', attrs={'class': 'desc-top'}).find('a')['href']
|
||||
params = {'terms': search_string.encode('utf-8'),
|
||||
'type': 1} # get anime types
|
||||
|
||||
if not title or not url:
|
||||
continue
|
||||
search_url = self.url + 'search.php?' + urllib.urlencode(params)
|
||||
logger.log(u'Search string: ' + search_url, logger.DEBUG)
|
||||
|
||||
item = title.lstrip(), url
|
||||
results.append(item)
|
||||
html = self.get_url(search_url)
|
||||
if html:
|
||||
try:
|
||||
with BS4Parser(html, features=['html5lib', 'permissive']) as soup:
|
||||
torrent_table = soup.find('table', attrs={'class': 'listing'})
|
||||
torrent_rows = torrent_table.find_all('tr') if torrent_table else []
|
||||
if torrent_rows:
|
||||
a = (0, 1)[torrent_rows[0].find('td', attrs={'class': 'centertext'})]
|
||||
|
||||
except Exception as e:
|
||||
logger.log(u"Failed to parsing " + self.name + " Traceback: " + traceback.format_exc(), logger.ERROR)
|
||||
for top, bottom in zip(torrent_rows[a::2], torrent_rows[a::2]):
|
||||
title = top.find('td', attrs={'class': 'desc-top'}).text
|
||||
url = top.find('td', attrs={'class': 'desc-top'}).find('a')['href']
|
||||
|
||||
if title and url:
|
||||
results.append((title.lstrip(), url))
|
||||
|
||||
except Exception:
|
||||
logger.log(u'Failed to parsing ' + self.name + ' Traceback: ' + traceback.format_exc(), logger.ERROR)
|
||||
|
||||
return results
|
||||
|
||||
def find_search_results(self, show, episodes, search_mode, manual_search=False):
|
||||
|
||||
return generic.TorrentProvider.find_search_results(self, show, episodes, search_mode, manual_search)
|
||||
|
||||
def _get_season_search_strings(self, ep_obj, **kwargs):
|
||||
|
||||
return [x.replace('.', ' ') for x in show_name_helpers.makeSceneSeasonSearchString(self.show, ep_obj)]
|
||||
|
||||
def _get_episode_search_strings(self, ep_obj, **kwargs):
|
||||
|
||||
return [x.replace('.', ' ') for x in show_name_helpers.makeSceneSearchString(self.show, ep_obj)]
|
||||
|
||||
|
||||
class TokyoToshokanCache(tvcache.TVCache):
|
||||
def __init__(self, provider):
|
||||
tvcache.TVCache.__init__(self, provider)
|
||||
|
||||
# only poll NyaaTorrents every 15 minutes max
|
||||
self.minTime = 15
|
||||
def __init__(self, this_provider):
|
||||
tvcache.TVCache.__init__(self, this_provider)
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
"""
|
||||
Retrieves the title and URL data from the item XML node
|
||||
|
||||
item: An elementtree.ElementTree element representing the <item> tag of the RSS feed
|
||||
|
||||
Returns: A tuple containing two strings representing title and URL respectively
|
||||
"""
|
||||
|
||||
title = item.title if item.title else None
|
||||
if title:
|
||||
title = u'' + title
|
||||
title = title.replace(' ', '.')
|
||||
|
||||
url = item.link if item.link else None
|
||||
if url:
|
||||
url = url.replace('&', '&')
|
||||
|
||||
return (title, url)
|
||||
self.minTime = 15 # cache update frequency
|
||||
|
||||
def _getRSSData(self):
|
||||
params = {
|
||||
"filter": '1',
|
||||
}
|
||||
params = {'filter': '1'}
|
||||
|
||||
url = self.provider.url + 'rss.php?' + urllib.urlencode(params)
|
||||
|
||||
logger.log(u"TokyoToshokan cache update URL: " + url, logger.DEBUG)
|
||||
logger.log(u'TokyoToshokan cache update URL: ' + url, logger.DEBUG)
|
||||
|
||||
data = self.getRSSFeed(url)
|
||||
|
||||
if data and 'entries' in data:
|
||||
return data.entries
|
||||
else:
|
||||
return []
|
||||
return []
|
||||
|
||||
|
||||
provider = TokyoToshokanProvider()
|
||||
|
|
|
@ -43,13 +43,13 @@ class TorrentBytesProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = TorrentBytesCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'login': 'Log in!'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -62,10 +62,10 @@ class TorrentBytesProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -77,7 +77,7 @@ class TorrentBytesProvider(generic.TorrentProvider):
|
|||
search_string = unidecode(search_string)
|
||||
|
||||
search_url = self.urls['search'] % (search_string, self.categories)
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -121,7 +121,7 @@ class TorrentBytesProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -44,13 +44,13 @@ class TorrentDayProvider(generic.TorrentProvider):
|
|||
self.freeleech = False
|
||||
self.cache = TorrentDayCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'uid' in self.session.cookies and 'pass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'submit.x': 0, 'submit.y': 0}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -63,10 +63,10 @@ class TorrentDayProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -80,7 +80,7 @@ class TorrentDayProvider(generic.TorrentProvider):
|
|||
if self.freeleech:
|
||||
post_data.update({'free': 'on'})
|
||||
|
||||
data_json = self.getURL(self.urls['search'], post_data=post_data, json=True)
|
||||
data_json = self.get_url(self.urls['search'], post_data=post_data, json=True)
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
if not data_json:
|
||||
|
@ -111,7 +111,7 @@ class TorrentDayProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date, '')
|
||||
|
||||
|
|
|
@ -44,13 +44,13 @@ class TorrentLeechProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = TorrentLeechCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'tluid' in self.session.cookies and 'tlpass' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'remember_me': 'on', 'login': 'submit'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -60,10 +60,10 @@ class TorrentLeechProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -80,7 +80,7 @@ class TorrentLeechProvider(generic.TorrentProvider):
|
|||
else:
|
||||
search_url = self.urls['search'] % (search_string, self.categories)
|
||||
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -123,7 +123,7 @@ class TorrentLeechProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -46,13 +46,13 @@ class TorrentShackProvider(generic.TorrentProvider):
|
|||
self.username, self.password, self.minseed, self.minleech = 4 * [None]
|
||||
self.cache = TorrentShackCache(self)
|
||||
|
||||
def _doLogin(self):
|
||||
def _do_login(self):
|
||||
|
||||
logged_in = lambda: 'session' in self.session.cookies
|
||||
if logged_in():
|
||||
return True
|
||||
|
||||
if self._checkAuth():
|
||||
if self._check_auth():
|
||||
login_params = {'username': self.username, 'password': self.password, 'keeplogged': '1', 'login': 'Login'}
|
||||
response = helpers.getURL(self.urls['login'], post_data=login_params, session=self.session)
|
||||
if response and logged_in():
|
||||
|
@ -65,10 +65,10 @@ class TorrentShackProvider(generic.TorrentProvider):
|
|||
|
||||
return False
|
||||
|
||||
def _doSearch(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, search_mode='eponly', epcount=0, age=0):
|
||||
|
||||
results = []
|
||||
if not self._doLogin():
|
||||
if not self._do_login():
|
||||
return results
|
||||
|
||||
items = {'Season': [], 'Episode': [], 'Cache': []}
|
||||
|
@ -83,7 +83,7 @@ class TorrentShackProvider(generic.TorrentProvider):
|
|||
|
||||
# fetch 15 results by default, and up to 100 if allowed in user profile
|
||||
search_url = self.urls['search'] % search_string
|
||||
html = self.getURL(search_url)
|
||||
html = self.get_url(search_url)
|
||||
|
||||
cnt = len(items[mode])
|
||||
try:
|
||||
|
@ -128,7 +128,7 @@ class TorrentShackProvider(generic.TorrentProvider):
|
|||
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
|
|
@ -48,9 +48,9 @@ class ToTVProvider(generic.TorrentProvider):
|
|||
logger.DEBUG)
|
||||
raise AuthException('Your authentication credentials for %s are incorrect, check your config.' % self.name)
|
||||
|
||||
def _doSearch(self, search_params, mode='eponly', epcount=0, age=0):
|
||||
def _do_search(self, search_params, mode='eponly', epcount=0, age=0):
|
||||
|
||||
self._checkAuth()
|
||||
self._check_auth()
|
||||
self.headers.update({'X-Authorization': self.api_key})
|
||||
results = []
|
||||
params = {'limit': 100}
|
||||
|
@ -62,7 +62,7 @@ class ToTVProvider(generic.TorrentProvider):
|
|||
|
||||
search_url = self.urls['search'] % urllib.urlencode(params)
|
||||
|
||||
data_json = self.getURL(search_url, json=True)
|
||||
data_json = self.get_url(search_url, json=True)
|
||||
|
||||
cnt = len(results)
|
||||
if data_json and 'results' in data_json and self._check_auth_from_data(data_json):
|
||||
|
@ -82,7 +82,7 @@ class ToTVProvider(generic.TorrentProvider):
|
|||
self._log_result(mode, len(results) - cnt, search_url)
|
||||
return results
|
||||
|
||||
def findPropers(self, search_date=datetime.datetime.today()):
|
||||
def find_propers(self, search_date=datetime.datetime.today()):
|
||||
|
||||
return self._find_propers(search_date)
|
||||
|
||||
|
@ -113,7 +113,7 @@ class ToTVProvider(generic.TorrentProvider):
|
|||
|
||||
def get_cache_data(self, *args, **kwargs):
|
||||
|
||||
return self._doSearch({})
|
||||
return self._do_search({})
|
||||
|
||||
|
||||
class ToTVCache(tvcache.TVCache):
|
||||
|
|
|
@ -66,7 +66,7 @@ class WombleCache(tvcache.TVCache):
|
|||
cl.append(ci)
|
||||
|
||||
if 0 < len(cl):
|
||||
my_db = self._getDB()
|
||||
my_db = self.get_db()
|
||||
my_db.mass_action(cl)
|
||||
|
||||
# set last updated
|
||||
|
|
|
@ -3,9 +3,6 @@
|
|||
# This file is part of SickGear.
|
||||
#
|
||||
|
||||
import re
|
||||
import requests
|
||||
import requests.cookies
|
||||
from feedparser import feedparser
|
||||
|
||||
from sickbeard import helpers, logger
|
||||
|
@ -20,28 +17,10 @@ class RSSFeeds:
|
|||
|
||||
def _check_auth_cookie(self):
|
||||
|
||||
if self.provider and hasattr(self.provider, 'cookies'):
|
||||
cookies = self.provider.cookies
|
||||
|
||||
if not re.match('^(\w+=\w+[;\s]*)+$', cookies):
|
||||
return False
|
||||
|
||||
cj = requests.utils.add_dict_to_cookiejar(self.provider.session.cookies,
|
||||
dict([x.strip().split('=') for x in cookies.split(';')
|
||||
if x != ''])),
|
||||
for item in cj:
|
||||
if not isinstance(item, requests.cookies.RequestsCookieJar):
|
||||
return False
|
||||
|
||||
if self.provider:
|
||||
return self.provider.check_auth_cookie()
|
||||
return True
|
||||
|
||||
def check_cookie(self):
|
||||
|
||||
if self._check_auth_cookie():
|
||||
return True, None
|
||||
|
||||
return False, 'Cookies not correctly formatted key=value pairs e.g. uid=xx;pass=yy): ' + self.provider.cookies
|
||||
|
||||
def get_feed(self, url, request_headers=None):
|
||||
|
||||
if not self._check_auth_cookie():
|
||||
|
|
|
@ -59,17 +59,8 @@ def sendNZB(nzb):
|
|||
|
||||
# if it's a normal result we just pass SAB the URL
|
||||
if nzb.resultType == "nzb":
|
||||
# for newzbin results send the ID to sab specifically
|
||||
if nzb.provider.getID() == 'newzbin':
|
||||
id = nzb.provider.getIDFromURL(nzb.url)
|
||||
if not id:
|
||||
logger.log("Unable to send NZB to SABnzbd, can't find ID in URL " + str(nzb.url), logger.ERROR)
|
||||
return False
|
||||
params['mode'] = 'addid'
|
||||
params['name'] = id
|
||||
else:
|
||||
params['mode'] = 'addurl'
|
||||
params['name'] = nzb.url
|
||||
params['mode'] = 'addurl'
|
||||
params['name'] = nzb.url
|
||||
|
||||
# if we get a raw data result we want to upload it to SAB
|
||||
elif nzb.resultType == "nzbdata":
|
||||
|
|
|
@ -60,7 +60,7 @@ def _downloadResult(result):
|
|||
|
||||
# nzbs with an URL can just be downloaded from the provider
|
||||
if result.resultType == "nzb":
|
||||
newResult = resProvider.downloadResult(result)
|
||||
newResult = resProvider.download_result(result)
|
||||
# if it's an nzb data result
|
||||
elif result.resultType == "nzbdata":
|
||||
|
||||
|
@ -82,7 +82,7 @@ def _downloadResult(result):
|
|||
logger.log(u"Error trying to save NZB to black hole: " + ex(e), logger.ERROR)
|
||||
newResult = False
|
||||
elif resProvider.providerType == "torrent":
|
||||
newResult = resProvider.downloadResult(result)
|
||||
newResult = resProvider.download_result(result)
|
||||
else:
|
||||
logger.log(u"Invalid provider type - this is a coding error, report it please", logger.ERROR)
|
||||
newResult = False
|
||||
|
@ -134,7 +134,7 @@ def snatchEpisode(result, endStatus=SNATCHED):
|
|||
else:
|
||||
# make sure we have the torrent file content
|
||||
if not result.content and not result.url.startswith('magnet'):
|
||||
result.content = result.provider.getURL(result.url)
|
||||
result.content = result.provider.get_url(result.url)
|
||||
if not result.content:
|
||||
logger.log(u'Torrent content failed to download from ' + result.url, logger.ERROR)
|
||||
return False
|
||||
|
@ -401,12 +401,12 @@ def searchForNeededEpisodes(episodes):
|
|||
|
||||
origThreadName = threading.currentThread().name
|
||||
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.isActive() and x.enable_recentsearch]
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.is_active() and x.enable_recentsearch]
|
||||
|
||||
for curProvider in providers:
|
||||
threading.currentThread().name = origThreadName + " :: [" + curProvider.name + "]"
|
||||
|
||||
curFoundResults = curProvider.searchRSS(episodes)
|
||||
curFoundResults = curProvider.search_rss(episodes)
|
||||
|
||||
didSearch = True
|
||||
|
||||
|
@ -435,7 +435,7 @@ def searchForNeededEpisodes(episodes):
|
|||
if bestResult.resultType == "torrent" and sickbeard.TORRENT_METHOD != "blackhole":
|
||||
bestResult.content = None
|
||||
if not bestResult.url.startswith('magnet'):
|
||||
bestResult.content = bestResult.provider.getURL(bestResult.url)
|
||||
bestResult.content = bestResult.provider.get_url(bestResult.url)
|
||||
if not bestResult.content:
|
||||
continue
|
||||
|
||||
|
@ -451,7 +451,7 @@ def searchForNeededEpisodes(episodes):
|
|||
return foundResults.values()
|
||||
|
||||
|
||||
def searchProviders(show, episodes, manualSearch=False):
|
||||
def searchProviders(show, episodes, manual_search=False):
|
||||
foundResults = {}
|
||||
finalResults = []
|
||||
|
||||
|
@ -459,7 +459,7 @@ def searchProviders(show, episodes, manualSearch=False):
|
|||
|
||||
origThreadName = threading.currentThread().name
|
||||
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.isActive() and x.enable_backlog]
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.is_active() and x.enable_backlog]
|
||||
for providerNum, curProvider in enumerate(providers):
|
||||
if curProvider.anime_only and not show.is_anime:
|
||||
logger.log(u"" + str(show.name) + " is not an anime, skipping", logger.DEBUG)
|
||||
|
@ -482,7 +482,7 @@ def searchProviders(show, episodes, manualSearch=False):
|
|||
|
||||
try:
|
||||
curProvider.cache._clearCache()
|
||||
searchResults = curProvider.findSearchResults(show, episodes, search_mode, manualSearch)
|
||||
searchResults = curProvider.find_search_results(show, episodes, search_mode, manual_search)
|
||||
except exceptions.AuthException as e:
|
||||
logger.log(u"Authentication error: " + ex(e), logger.ERROR)
|
||||
break
|
||||
|
@ -703,7 +703,7 @@ def searchProviders(show, episodes, manualSearch=False):
|
|||
if bestResult.resultType == "torrent" and sickbeard.TORRENT_METHOD != "blackhole":
|
||||
bestResult.content = None
|
||||
if not bestResult.url.startswith('magnet'):
|
||||
bestResult.content = bestResult.provider.getURL(bestResult.url)
|
||||
bestResult.content = bestResult.provider.get_url(bestResult.url)
|
||||
if not bestResult.content:
|
||||
continue
|
||||
|
||||
|
|
|
@ -291,7 +291,7 @@ class RecentSearchQueueItem(generic_queue.QueueItem):
|
|||
|
||||
logger.log('Updating provider caches with recent upload data')
|
||||
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.isActive() and x.enable_recentsearch]
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.is_active() and x.enable_recentsearch]
|
||||
for curProvider in providers:
|
||||
# spawn separate threads for each provider so we don't need to wait for providers with slow network operation
|
||||
threads.append(threading.Thread(target=curProvider.cache.updateCache, name=origThreadName +
|
||||
|
|
|
@ -48,16 +48,16 @@ class TVCache:
|
|||
def __init__(self, provider):
|
||||
|
||||
self.provider = provider
|
||||
self.providerID = self.provider.getID()
|
||||
self.providerID = self.provider.get_id()
|
||||
self.providerDB = None
|
||||
self.minTime = 10
|
||||
|
||||
def _getDB(self):
|
||||
def get_db(self):
|
||||
return CacheDBConnection(self.providerID)
|
||||
|
||||
def _clearCache(self):
|
||||
if self.shouldClearCache():
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
myDB.action('DELETE FROM provider_cache WHERE provider = ?', [self.providerID])
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
|
@ -69,7 +69,7 @@ class TVCache:
|
|||
return data
|
||||
|
||||
def _checkAuth(self):
|
||||
return self.provider._checkAuth()
|
||||
return self.provider._check_auth()
|
||||
|
||||
def _checkItemAuth(self, title, url):
|
||||
return True
|
||||
|
@ -102,7 +102,7 @@ class TVCache:
|
|||
cl.append(ci)
|
||||
|
||||
if len(cl) > 0:
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
myDB.mass_action(cl)
|
||||
|
||||
return []
|
||||
|
@ -125,7 +125,7 @@ class TVCache:
|
|||
url = self._translateLinkURL(url)
|
||||
|
||||
logger.log(u'Attempting to add item to cache: ' + title, logger.DEBUG)
|
||||
return self._addCacheEntry(title, url)
|
||||
return self.add_cache_entry(title, url)
|
||||
|
||||
else:
|
||||
logger.log(
|
||||
|
@ -134,7 +134,7 @@ class TVCache:
|
|||
return None
|
||||
|
||||
def _getLastUpdate(self):
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
sqlResults = myDB.select('SELECT time FROM lastUpdate WHERE provider = ?', [self.providerID])
|
||||
|
||||
if sqlResults:
|
||||
|
@ -147,7 +147,7 @@ class TVCache:
|
|||
return datetime.datetime.fromtimestamp(lastTime)
|
||||
|
||||
def _getLastSearch(self):
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
sqlResults = myDB.select('SELECT time FROM lastSearch WHERE provider = ?', [self.providerID])
|
||||
|
||||
if sqlResults:
|
||||
|
@ -163,7 +163,7 @@ class TVCache:
|
|||
if not toDate:
|
||||
toDate = datetime.datetime.today()
|
||||
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
myDB.upsert('lastUpdate',
|
||||
{'time': int(time.mktime(toDate.timetuple()))},
|
||||
{'provider': self.providerID})
|
||||
|
@ -172,7 +172,7 @@ class TVCache:
|
|||
if not toDate:
|
||||
toDate = datetime.datetime.today()
|
||||
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
myDB.upsert('lastSearch',
|
||||
{'time': int(time.mktime(toDate.timetuple()))},
|
||||
{'provider': self.providerID})
|
||||
|
@ -196,7 +196,7 @@ class TVCache:
|
|||
|
||||
return True
|
||||
|
||||
def _addCacheEntry(self, name, url, parse_result=None, indexer_id=0):
|
||||
def add_cache_entry(self, name, url, parse_result=None, indexer_id=0):
|
||||
|
||||
# check if we passed in a parsed result or should we try and create one
|
||||
if not parse_result:
|
||||
|
@ -256,7 +256,7 @@ class TVCache:
|
|||
return []
|
||||
|
||||
def listPropers(self, date=None, delimiter='.'):
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
sql = "SELECT * FROM provider_cache WHERE name LIKE '%.PROPER.%' OR name LIKE '%.REPACK.%' AND provider = ?"
|
||||
|
||||
if date != None:
|
||||
|
@ -268,7 +268,7 @@ class TVCache:
|
|||
neededEps = {}
|
||||
cl = []
|
||||
|
||||
myDB = self._getDB()
|
||||
myDB = self.get_db()
|
||||
if type(episode) != list:
|
||||
sqlResults = myDB.select(
|
||||
'SELECT * FROM provider_cache WHERE provider = ? AND indexerid = ? AND season = ? AND episodes LIKE ?',
|
||||
|
@ -331,9 +331,10 @@ class TVCache:
|
|||
|
||||
logger.log(u'Found result ' + title + ' at ' + url)
|
||||
|
||||
result = self.provider.getResult([epObj])
|
||||
result = self.provider.get_result([epObj], url)
|
||||
if None is result:
|
||||
continue
|
||||
result.show = showObj
|
||||
result.url = url
|
||||
result.name = title
|
||||
result.quality = curQuality
|
||||
result.release_group = curReleaseGroup
|
||||
|
|
|
@ -3914,14 +3914,14 @@ class ConfigProviders(Config):
|
|||
if not name:
|
||||
return json.dumps({'error': 'No Provider Name specified'})
|
||||
|
||||
providerDict = dict(zip([x.getID() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
providerDict = dict(zip([x.get_id() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
|
||||
tempProvider = newznab.NewznabProvider(name, '')
|
||||
|
||||
if tempProvider.getID() in providerDict:
|
||||
return json.dumps({'error': 'Provider Name already exists as ' + providerDict[tempProvider.getID()].name})
|
||||
if tempProvider.get_id() in providerDict:
|
||||
return json.dumps({'error': 'Provider Name already exists as ' + providerDict[tempProvider.get_id()].name})
|
||||
else:
|
||||
return json.dumps({'success': tempProvider.getID()})
|
||||
return json.dumps({'success': tempProvider.get_id()})
|
||||
|
||||
def saveNewznabProvider(self, name, url, key=''):
|
||||
|
||||
|
@ -3942,12 +3942,12 @@ class ConfigProviders(Config):
|
|||
else:
|
||||
providerDict[name].needs_auth = True
|
||||
|
||||
return providerDict[name].getID() + '|' + providerDict[name].config_str()
|
||||
return providerDict[name].get_id() + '|' + providerDict[name].config_str()
|
||||
|
||||
else:
|
||||
newProvider = newznab.NewznabProvider(name, url, key=key)
|
||||
sickbeard.newznabProviderList.append(newProvider)
|
||||
return newProvider.getID() + '|' + newProvider.config_str()
|
||||
return newProvider.get_id() + '|' + newProvider.config_str()
|
||||
|
||||
def getNewznabCategories(self, name, url, key):
|
||||
'''
|
||||
|
@ -3969,7 +3969,7 @@ class ConfigProviders(Config):
|
|||
return json.dumps({'success' : False, 'error': error})
|
||||
|
||||
#Get list with Newznabproviders
|
||||
#providerDict = dict(zip([x.getID() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
#providerDict = dict(zip([x.get_id() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
|
||||
#Get newznabprovider obj with provided name
|
||||
tempProvider= newznab.NewznabProvider(name, url, key)
|
||||
|
@ -3980,7 +3980,7 @@ class ConfigProviders(Config):
|
|||
|
||||
def deleteNewznabProvider(self, nnid):
|
||||
|
||||
providerDict = dict(zip([x.getID() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
providerDict = dict(zip([x.get_id() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
|
||||
if nnid not in providerDict or providerDict[nnid].default:
|
||||
return '0'
|
||||
|
@ -3999,16 +3999,16 @@ class ConfigProviders(Config):
|
|||
return json.dumps({'error': 'Invalid name specified'})
|
||||
|
||||
providerDict = dict(
|
||||
zip([x.getID() for x in sickbeard.torrentRssProviderList], sickbeard.torrentRssProviderList))
|
||||
zip([x.get_id() for x in sickbeard.torrentRssProviderList], sickbeard.torrentRssProviderList))
|
||||
|
||||
tempProvider = rsstorrent.TorrentRssProvider(name, url, cookies)
|
||||
|
||||
if tempProvider.getID() in providerDict:
|
||||
return json.dumps({'error': 'Exists as ' + providerDict[tempProvider.getID()].name})
|
||||
if tempProvider.get_id() in providerDict:
|
||||
return json.dumps({'error': 'Exists as ' + providerDict[tempProvider.get_id()].name})
|
||||
else:
|
||||
(succ, errMsg) = tempProvider.validate_feed()
|
||||
if succ:
|
||||
return json.dumps({'success': tempProvider.getID()})
|
||||
return json.dumps({'success': tempProvider.get_id()})
|
||||
else:
|
||||
return json.dumps({'error': errMsg})
|
||||
|
||||
|
@ -4024,17 +4024,17 @@ class ConfigProviders(Config):
|
|||
providerDict[name].url = config.clean_url(url)
|
||||
providerDict[name].cookies = cookies
|
||||
|
||||
return providerDict[name].getID() + '|' + providerDict[name].config_str()
|
||||
return providerDict[name].get_id() + '|' + providerDict[name].config_str()
|
||||
|
||||
else:
|
||||
newProvider = rsstorrent.TorrentRssProvider(name, url, cookies)
|
||||
sickbeard.torrentRssProviderList.append(newProvider)
|
||||
return newProvider.getID() + '|' + newProvider.config_str()
|
||||
return newProvider.get_id() + '|' + newProvider.config_str()
|
||||
|
||||
def deleteTorrentRssProvider(self, id):
|
||||
|
||||
providerDict = dict(
|
||||
zip([x.getID() for x in sickbeard.torrentRssProviderList], sickbeard.torrentRssProviderList))
|
||||
zip([x.get_id() for x in sickbeard.torrentRssProviderList], sickbeard.torrentRssProviderList))
|
||||
|
||||
if id not in providerDict:
|
||||
return '0'
|
||||
|
@ -4055,7 +4055,7 @@ class ConfigProviders(Config):
|
|||
provider_list = []
|
||||
|
||||
newznabProviderDict = dict(
|
||||
zip([x.getID() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
zip([x.get_id() for x in sickbeard.newznabProviderList], sickbeard.newznabProviderList))
|
||||
|
||||
finishedNames = []
|
||||
|
||||
|
@ -4074,7 +4074,7 @@ class ConfigProviders(Config):
|
|||
|
||||
newProvider = newznab.NewznabProvider(cur_name, cur_url, key=cur_key)
|
||||
|
||||
cur_id = newProvider.getID()
|
||||
cur_id = newProvider.get_id()
|
||||
|
||||
# if it already exists then update it
|
||||
if cur_id in newznabProviderDict:
|
||||
|
@ -4118,11 +4118,11 @@ class ConfigProviders(Config):
|
|||
|
||||
# delete anything that is missing
|
||||
for curProvider in sickbeard.newznabProviderList:
|
||||
if curProvider.getID() not in finishedNames:
|
||||
if curProvider.get_id() not in finishedNames:
|
||||
sickbeard.newznabProviderList.remove(curProvider)
|
||||
|
||||
torrentRssProviderDict = dict(
|
||||
zip([x.getID() for x in sickbeard.torrentRssProviderList], sickbeard.torrentRssProviderList))
|
||||
zip([x.get_id() for x in sickbeard.torrentRssProviderList], sickbeard.torrentRssProviderList))
|
||||
finishedNames = []
|
||||
|
||||
if torrentrss_string:
|
||||
|
@ -4139,7 +4139,7 @@ class ConfigProviders(Config):
|
|||
|
||||
newProvider = rsstorrent.TorrentRssProvider(curName, curURL, curCookies)
|
||||
|
||||
curID = newProvider.getID()
|
||||
curID = newProvider.get_id()
|
||||
|
||||
# if it already exists then update it
|
||||
if curID in torrentRssProviderDict:
|
||||
|
@ -4154,7 +4154,7 @@ class ConfigProviders(Config):
|
|||
|
||||
# delete anything that is missing
|
||||
for curProvider in sickbeard.torrentRssProviderList:
|
||||
if curProvider.getID() not in finishedNames:
|
||||
if curProvider.get_id() not in finishedNames:
|
||||
sickbeard.torrentRssProviderList.remove(curProvider)
|
||||
|
||||
# do the enable/disable
|
||||
|
@ -4163,7 +4163,7 @@ class ConfigProviders(Config):
|
|||
curEnabled = config.to_int(curEnabled)
|
||||
|
||||
curProvObj = [x for x in sickbeard.providers.sortedProviderList() if
|
||||
x.getID() == curProvider and hasattr(x, 'enabled')]
|
||||
x.get_id() == curProvider and hasattr(x, 'enabled')]
|
||||
if curProvObj:
|
||||
curProvObj[0].enabled = bool(curEnabled)
|
||||
|
||||
|
@ -4179,31 +4179,31 @@ class ConfigProviders(Config):
|
|||
|
||||
if hasattr(curTorrentProvider, 'minseed'):
|
||||
try:
|
||||
curTorrentProvider.minseed = int(str(kwargs[curTorrentProvider.getID() + '_minseed']).strip())
|
||||
curTorrentProvider.minseed = int(str(kwargs[curTorrentProvider.get_id() + '_minseed']).strip())
|
||||
except:
|
||||
curTorrentProvider.minseed = 0
|
||||
|
||||
if hasattr(curTorrentProvider, 'minleech'):
|
||||
try:
|
||||
curTorrentProvider.minleech = int(str(kwargs[curTorrentProvider.getID() + '_minleech']).strip())
|
||||
curTorrentProvider.minleech = int(str(kwargs[curTorrentProvider.get_id() + '_minleech']).strip())
|
||||
except:
|
||||
curTorrentProvider.minleech = 0
|
||||
|
||||
if hasattr(curTorrentProvider, 'ratio'):
|
||||
try:
|
||||
curTorrentProvider.ratio = str(kwargs[curTorrentProvider.getID() + '_ratio']).strip()
|
||||
curTorrentProvider.ratio = str(kwargs[curTorrentProvider.get_id() + '_ratio']).strip()
|
||||
except:
|
||||
curTorrentProvider.ratio = None
|
||||
|
||||
if hasattr(curTorrentProvider, 'digest'):
|
||||
try:
|
||||
curTorrentProvider.digest = str(kwargs[curTorrentProvider.getID() + '_digest']).strip()
|
||||
curTorrentProvider.digest = str(kwargs[curTorrentProvider.get_id() + '_digest']).strip()
|
||||
except:
|
||||
curTorrentProvider.digest = None
|
||||
|
||||
if hasattr(curTorrentProvider, 'hash'):
|
||||
try:
|
||||
key = str(kwargs[curTorrentProvider.getID() + '_hash']).strip()
|
||||
key = str(kwargs[curTorrentProvider.get_id() + '_hash']).strip()
|
||||
if not starify(key, True):
|
||||
curTorrentProvider.hash = key
|
||||
except:
|
||||
|
@ -4211,7 +4211,7 @@ class ConfigProviders(Config):
|
|||
|
||||
if hasattr(curTorrentProvider, 'api_key'):
|
||||
try:
|
||||
key = str(kwargs[curTorrentProvider.getID() + '_api_key']).strip()
|
||||
key = str(kwargs[curTorrentProvider.get_id() + '_api_key']).strip()
|
||||
if not starify(key, True):
|
||||
curTorrentProvider.api_key = key
|
||||
except:
|
||||
|
@ -4219,13 +4219,13 @@ class ConfigProviders(Config):
|
|||
|
||||
if hasattr(curTorrentProvider, 'username'):
|
||||
try:
|
||||
curTorrentProvider.username = str(kwargs[curTorrentProvider.getID() + '_username']).strip()
|
||||
curTorrentProvider.username = str(kwargs[curTorrentProvider.get_id() + '_username']).strip()
|
||||
except:
|
||||
curTorrentProvider.username = None
|
||||
|
||||
if hasattr(curTorrentProvider, 'password'):
|
||||
try:
|
||||
key = str(kwargs[curTorrentProvider.getID() + '_password']).strip()
|
||||
key = str(kwargs[curTorrentProvider.get_id() + '_password']).strip()
|
||||
if set('*') != set(key):
|
||||
curTorrentProvider.password = key
|
||||
except:
|
||||
|
@ -4233,7 +4233,7 @@ class ConfigProviders(Config):
|
|||
|
||||
if hasattr(curTorrentProvider, 'passkey'):
|
||||
try:
|
||||
key = str(kwargs[curTorrentProvider.getID() + '_passkey']).strip()
|
||||
key = str(kwargs[curTorrentProvider.get_id() + '_passkey']).strip()
|
||||
if not starify(key, True):
|
||||
curTorrentProvider.passkey = key
|
||||
except:
|
||||
|
@ -4242,54 +4242,54 @@ class ConfigProviders(Config):
|
|||
if hasattr(curTorrentProvider, 'confirmed'):
|
||||
try:
|
||||
curTorrentProvider.confirmed = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_confirmed'])
|
||||
kwargs[curTorrentProvider.get_id() + '_confirmed'])
|
||||
except:
|
||||
curTorrentProvider.confirmed = 0
|
||||
|
||||
if hasattr(curTorrentProvider, 'proxy'):
|
||||
try:
|
||||
curTorrentProvider.proxy.enabled = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_proxy'])
|
||||
kwargs[curTorrentProvider.get_id() + '_proxy'])
|
||||
except:
|
||||
curTorrentProvider.proxy.enabled = 0
|
||||
|
||||
if hasattr(curTorrentProvider.proxy, 'url'):
|
||||
try:
|
||||
curTorrentProvider.proxy.url = str(kwargs[curTorrentProvider.getID() + '_proxy_url']).strip()
|
||||
curTorrentProvider.proxy.url = str(kwargs[curTorrentProvider.get_id() + '_proxy_url']).strip()
|
||||
except:
|
||||
curTorrentProvider.proxy.url = None
|
||||
|
||||
if hasattr(curTorrentProvider, 'freeleech'):
|
||||
try:
|
||||
curTorrentProvider.freeleech = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_freeleech'])
|
||||
kwargs[curTorrentProvider.get_id() + '_freeleech'])
|
||||
except:
|
||||
curTorrentProvider.freeleech = 0
|
||||
|
||||
if hasattr(curTorrentProvider, 'search_mode'):
|
||||
try:
|
||||
curTorrentProvider.search_mode = str(kwargs[curTorrentProvider.getID() + '_search_mode']).strip()
|
||||
curTorrentProvider.search_mode = str(kwargs[curTorrentProvider.get_id() + '_search_mode']).strip()
|
||||
except:
|
||||
curTorrentProvider.search_mode = 'eponly'
|
||||
|
||||
if hasattr(curTorrentProvider, 'search_fallback'):
|
||||
try:
|
||||
curTorrentProvider.search_fallback = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_search_fallback'])
|
||||
kwargs[curTorrentProvider.get_id() + '_search_fallback'])
|
||||
except:
|
||||
curTorrentProvider.search_fallback = 0 # these exceptions are catching unselected checkboxes
|
||||
|
||||
if hasattr(curTorrentProvider, 'enable_recentsearch'):
|
||||
try:
|
||||
curTorrentProvider.enable_recentsearch = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_enable_recentsearch'])
|
||||
kwargs[curTorrentProvider.get_id() + '_enable_recentsearch'])
|
||||
except:
|
||||
curTorrentProvider.enable_recentsearch = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
if hasattr(curTorrentProvider, 'enable_backlog'):
|
||||
try:
|
||||
curTorrentProvider.enable_backlog = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_enable_backlog'])
|
||||
kwargs[curTorrentProvider.get_id() + '_enable_backlog'])
|
||||
except:
|
||||
curTorrentProvider.enable_backlog = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
|
@ -4298,7 +4298,7 @@ class ConfigProviders(Config):
|
|||
|
||||
if hasattr(curNzbProvider, 'api_key'):
|
||||
try:
|
||||
key = str(kwargs[curNzbProvider.getID() + '_api_key']).strip()
|
||||
key = str(kwargs[curNzbProvider.get_id() + '_api_key']).strip()
|
||||
if not starify(key, True):
|
||||
curNzbProvider.api_key = key
|
||||
except:
|
||||
|
@ -4306,34 +4306,34 @@ class ConfigProviders(Config):
|
|||
|
||||
if hasattr(curNzbProvider, 'username'):
|
||||
try:
|
||||
curNzbProvider.username = str(kwargs[curNzbProvider.getID() + '_username']).strip()
|
||||
curNzbProvider.username = str(kwargs[curNzbProvider.get_id() + '_username']).strip()
|
||||
except:
|
||||
curNzbProvider.username = None
|
||||
|
||||
if hasattr(curNzbProvider, 'search_mode'):
|
||||
try:
|
||||
curNzbProvider.search_mode = str(kwargs[curNzbProvider.getID() + '_search_mode']).strip()
|
||||
curNzbProvider.search_mode = str(kwargs[curNzbProvider.get_id() + '_search_mode']).strip()
|
||||
except:
|
||||
curNzbProvider.search_mode = 'eponly'
|
||||
|
||||
if hasattr(curNzbProvider, 'search_fallback'):
|
||||
try:
|
||||
curNzbProvider.search_fallback = config.checkbox_to_value(
|
||||
kwargs[curNzbProvider.getID() + '_search_fallback'])
|
||||
kwargs[curNzbProvider.get_id() + '_search_fallback'])
|
||||
except:
|
||||
curNzbProvider.search_fallback = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
if hasattr(curNzbProvider, 'enable_recentsearch'):
|
||||
try:
|
||||
curNzbProvider.enable_recentsearch = config.checkbox_to_value(
|
||||
kwargs[curNzbProvider.getID() + '_enable_recentsearch'])
|
||||
kwargs[curNzbProvider.get_id() + '_enable_recentsearch'])
|
||||
except:
|
||||
curNzbProvider.enable_recentsearch = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
if hasattr(curNzbProvider, 'enable_backlog'):
|
||||
try:
|
||||
curNzbProvider.enable_backlog = config.checkbox_to_value(
|
||||
kwargs[curNzbProvider.getID() + '_enable_backlog'])
|
||||
kwargs[curNzbProvider.get_id() + '_enable_backlog'])
|
||||
except:
|
||||
curNzbProvider.enable_backlog = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
|
|
|
@ -59,7 +59,7 @@ class SearchTest(test.SickbeardTestDBCase):
|
|||
|
||||
def __init__(self, something):
|
||||
for provider in sickbeard.providers.sortedProviderList():
|
||||
provider.getURL = self._fake_getURL
|
||||
provider.get_url = self._fake_getURL
|
||||
#provider.isActive = self._fake_isActive
|
||||
|
||||
super(SearchTest, self).__init__(something)
|
||||
|
|
Loading…
Reference in a new issue