mirror of
https://github.com/SickGear/SickGear.git
synced 2024-12-01 08:53:37 +00:00
Merge pull request #129 from adam111316/feature/ChangeDailyToRecent
Change term "Daily" search to "Recent" search
This commit is contained in:
commit
f875e4c2fb
16 changed files with 148 additions and 130 deletions
|
@ -18,6 +18,8 @@
|
|||
* Add pull request checkout option to General Config/Advanced Settings
|
||||
* Add BTN api call parameter debug logging
|
||||
* Fix anime searches on BTN provider
|
||||
* Change replace "Daily-Search" with "Recent-Search"
|
||||
* Add daily search to recent search renaming to config migration code
|
||||
|
||||
[develop changelog]
|
||||
|
||||
|
|
|
@ -214,13 +214,13 @@
|
|||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curNewznabProvider, 'enable_daily'):
|
||||
#if $hasattr($curNewznabProvider, 'enable_recentsearch'):
|
||||
<div class="field-pair">
|
||||
<label for="${curNewznabProvider.getID()}_enable_daily">
|
||||
<span class="component-title">Enable daily searches</span>
|
||||
<label for="${curNewznabProvider.getID()}_enable_recentsearch">
|
||||
<span class="component-title">Enable recent searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNewznabProvider.getID()}_enable_daily" id="${curNewznabProvider.getID()}_enable_daily" <%= html_checked if curNewznabProvider.enable_daily else '' %>/>
|
||||
<p>perform daily searches at provider</p>
|
||||
<input type="checkbox" name="${curNewznabProvider.getID()}_enable_recentsearch" id="${curNewznabProvider.getID()}_enable_recentsearch" <%= html_checked if curNewznabProvider.enable_recentsearch else '' %>/>
|
||||
<p>perform recent searches at provider</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
@ -292,13 +292,13 @@
|
|||
#end if
|
||||
|
||||
|
||||
#if $hasattr($curNzbProvider, 'enable_daily'):
|
||||
#if $hasattr($curNzbProvider, 'enable_recentsearch'):
|
||||
<div class="field-pair">
|
||||
<label for="${curNzbProvider.getID()}_enable_daily">
|
||||
<span class="component-title">Enable daily searches</span>
|
||||
<label for="${curNzbProvider.getID()}_enable_recentsearch">
|
||||
<span class="component-title">Enable recent searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curNzbProvider.getID()}_enable_daily" id="${curNzbProvider.getID()}_enable_daily" <%= html_checked if curNzbProvider.enable_daily else '' %>/>
|
||||
<p>enable provider to perform daily searches.</p>
|
||||
<input type="checkbox" name="${curNzbProvider.getID()}_enable_recentsearch" id="${curNzbProvider.getID()}_enable_recentsearch" <%= html_checked if curNzbProvider.enable_recentsearch else '' %>/>
|
||||
<p>enable provider to perform recent searches.</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
@ -512,13 +512,13 @@
|
|||
</div>
|
||||
#end if
|
||||
|
||||
#if $hasattr($curTorrentProvider, 'enable_daily'):
|
||||
#if $hasattr($curTorrentProvider, 'enable_recentsearch'):
|
||||
<div class="field-pair">
|
||||
<label for="${curTorrentProvider.getID()}_enable_daily">
|
||||
<span class="component-title">Enable daily searches</span>
|
||||
<label for="${curTorrentProvider.getID()}_enable_recentsearch">
|
||||
<span class="component-title">Enable recent searches</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="${curTorrentProvider.getID()}_enable_daily" id="${curTorrentProvider.getID()}_enable_daily" <%= html_checked if curTorrentProvider.enable_daily else '' %>/>
|
||||
<p>enable provider to perform daily searches.</p>
|
||||
<input type="checkbox" name="${curTorrentProvider.getID()}_enable_recentsearch" id="${curTorrentProvider.getID()}_enable_recentsearch" <%= html_checked if curTorrentProvider.enable_recentsearch else '' %>/>
|
||||
<p>enable provider to perform recent searches.</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
|
|
@ -90,10 +90,10 @@
|
|||
|
||||
<div class="field-pair">
|
||||
<label>
|
||||
<span class="component-title">Daily search frequency</span>
|
||||
<span class="component-title">Recent search frequency</span>
|
||||
<span class="component-desc">
|
||||
<input type="text" name="dailysearch_frequency" value="$sickbeard.DAILYSEARCH_FREQUENCY" class="form-control input-sm input75" />
|
||||
<p>time in minutes between searches (min. $sickbeard.MIN_DAILYSEARCH_FREQUENCY)</p>
|
||||
<input type="text" name="recentsearch_frequency" value="$sickbeard.RECENTSEARCH_FREQUENCY" class="form-control input-sm input75" />
|
||||
<p>time in minutes between searches (min. $sickbeard.MIN_RECENTSEARCH_FREQUENCY)</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
@ -139,11 +139,11 @@
|
|||
</div>
|
||||
|
||||
<div class="field-pair">
|
||||
<label for="dailysearch_startup">
|
||||
<span class="component-title">Daily search on startup</span>
|
||||
<label for="recentsearch_startup">
|
||||
<span class="component-title">Recent search on startup</span>
|
||||
<span class="component-desc">
|
||||
<input type="checkbox" name="dailysearch_startup" id="dailysearch_startup" <%= html_checked if sickbeard.DAILYSEARCH_STARTUP == True else '' %>/>
|
||||
<p>start daily search on startup of SickGear</p>
|
||||
<input type="checkbox" name="recentsearch_startup" id="recentsearch_startup" <%= html_checked if sickbeard.RECENTSEARCH_STARTUP == True else '' %>/>
|
||||
<p>start recent search on startup of SickGear</p>
|
||||
</span>
|
||||
</label>
|
||||
</div>
|
||||
|
|
|
@ -62,7 +62,7 @@
|
|||
)[0 < ep_snatched]
|
||||
%>
|
||||
/ <span class="footerhighlight">$ep_total</span> episodes downloaded
|
||||
| daily search: <span class="footerhighlight"><%= str(sickbeard.dailySearchScheduler.timeLeft()).split('.')[0] %></span>
|
||||
| recent search: <span class="footerhighlight"><%= str(sickbeard.recentSearchScheduler.timeLeft()).split('.')[0] %></span>
|
||||
| backlog search: <span class="footerhighlight">$sbdatetime.sbdatetime.sbfdate($sickbeard.backlogSearchScheduler.nextRun())</span>
|
||||
|
||||
</div>
|
||||
|
|
|
@ -29,9 +29,9 @@ Currently running<br />
|
|||
#end if
|
||||
<br />
|
||||
|
||||
<h3>Daily Search:</h3>
|
||||
<h3>Recent Search:</h3>
|
||||
<a class="btn" href="$sbRoot/manage/manageSearches/forceSearch"><i class="icon-exclamation-sign"></i> Force</a>
|
||||
#if not $dailySearchStatus:
|
||||
#if not $recentSearchStatus:
|
||||
Not in progress<br />
|
||||
#else:
|
||||
In Progress<br />
|
||||
|
@ -53,7 +53,7 @@ In Progress<br />
|
|||
|
||||
<h3>Search Queue:</h3>
|
||||
Backlog: <i>$queueLength['backlog'] pending items</i></br>
|
||||
Daily: <i>$queueLength['daily'] pending items</i></br>
|
||||
Recent: <i>$queueLength['recent'] pending items</i></br>
|
||||
Manual: <i>$queueLength['manual'] pending items</i></br>
|
||||
Failed: <i>$queueLength['failed'] pending items</i></br>
|
||||
</div>
|
||||
|
|
|
@ -42,7 +42,7 @@ from sickbeard import searchBacklog, showUpdater, versionChecker, properFinder,
|
|||
from sickbeard import helpers, db, exceptions, show_queue, search_queue, scheduler, show_name_helpers
|
||||
from sickbeard import logger
|
||||
from sickbeard import naming
|
||||
from sickbeard import dailysearcher
|
||||
from sickbeard import searchRecent
|
||||
from sickbeard import scene_numbering, scene_exceptions, name_cache
|
||||
from indexers.indexer_api import indexerApi
|
||||
from indexers.indexer_exceptions import indexer_shownotfound, indexer_exception, indexer_error, indexer_episodenotfound, \
|
||||
|
@ -58,7 +58,7 @@ CFG = None
|
|||
CONFIG_FILE = None
|
||||
|
||||
# This is the version of the config we EXPECT to find
|
||||
CONFIG_VERSION = 5
|
||||
CONFIG_VERSION = 6
|
||||
|
||||
# Default encryption version (0 for None)
|
||||
ENCRYPTION_VERSION = 0
|
||||
|
@ -78,7 +78,7 @@ NO_RESIZE = False
|
|||
# system events
|
||||
events = None
|
||||
|
||||
dailySearchScheduler = None
|
||||
recentSearchScheduler = None
|
||||
backlogSearchScheduler = None
|
||||
showUpdateScheduler = None
|
||||
versionCheckScheduler = None
|
||||
|
@ -196,19 +196,19 @@ CHECK_PROPERS_INTERVAL = None
|
|||
ALLOW_HIGH_PRIORITY = False
|
||||
|
||||
AUTOPOSTPROCESSER_FREQUENCY = None
|
||||
DAILYSEARCH_FREQUENCY = None
|
||||
RECENTSEARCH_FREQUENCY = None
|
||||
UPDATE_FREQUENCY = None
|
||||
DAILYSEARCH_STARTUP = False
|
||||
RECENTSEARCH_STARTUP = False
|
||||
BACKLOG_FREQUENCY = None
|
||||
BACKLOG_STARTUP = False
|
||||
|
||||
DEFAULT_AUTOPOSTPROCESSER_FREQUENCY = 10
|
||||
DEFAULT_DAILYSEARCH_FREQUENCY = 40
|
||||
DEFAULT_RECENTSEARCH_FREQUENCY = 40
|
||||
DEFAULT_BACKLOG_FREQUENCY = 21
|
||||
DEFAULT_UPDATE_FREQUENCY = 1
|
||||
|
||||
MIN_AUTOPOSTPROCESSER_FREQUENCY = 1
|
||||
MIN_DAILYSEARCH_FREQUENCY = 10
|
||||
MIN_RECENTSEARCH_FREQUENCY = 10
|
||||
MIN_BACKLOG_FREQUENCY = 10
|
||||
MIN_UPDATE_FREQUENCY = 1
|
||||
|
||||
|
@ -459,7 +459,7 @@ TRAKT_API_KEY = 'abd806c54516240c76e4ebc9c5ccf394'
|
|||
__INITIALIZED__ = False
|
||||
|
||||
def get_backlog_cycle_time():
|
||||
cycletime = DAILYSEARCH_FREQUENCY * 2 + 7
|
||||
cycletime = RECENTSEARCH_FREQUENCY * 2 + 7
|
||||
return max([cycletime, 720])
|
||||
|
||||
def initialize(consoleLogging=True):
|
||||
|
@ -477,7 +477,7 @@ def initialize(consoleLogging=True):
|
|||
PLEX_SERVER_HOST, PLEX_HOST, PLEX_USERNAME, PLEX_PASSWORD, DEFAULT_BACKLOG_FREQUENCY, MIN_BACKLOG_FREQUENCY, BACKLOG_STARTUP, SKIP_REMOVED_FILES, \
|
||||
showUpdateScheduler, __INITIALIZED__, LAUNCH_BROWSER, UPDATE_SHOWS_ON_START, TRASH_REMOVE_SHOW, TRASH_ROTATE_LOGS, HOME_SEARCH_FOCUS, SORT_ARTICLE, showList, loadingShowList, \
|
||||
NEWZNAB_DATA, NZBS, NZBS_UID, NZBS_HASH, INDEXER_DEFAULT, INDEXER_TIMEOUT, USENET_RETENTION, TORRENT_DIR, \
|
||||
QUALITY_DEFAULT, FLATTEN_FOLDERS_DEFAULT, SUBTITLES_DEFAULT, STATUS_DEFAULT, DAILYSEARCH_STARTUP, \
|
||||
QUALITY_DEFAULT, FLATTEN_FOLDERS_DEFAULT, SUBTITLES_DEFAULT, STATUS_DEFAULT, RECENTSEARCH_STARTUP, \
|
||||
GROWL_NOTIFY_ONSNATCH, GROWL_NOTIFY_ONDOWNLOAD, GROWL_NOTIFY_ONSUBTITLEDOWNLOAD, TWITTER_NOTIFY_ONSNATCH, TWITTER_NOTIFY_ONDOWNLOAD, TWITTER_NOTIFY_ONSUBTITLEDOWNLOAD, \
|
||||
USE_GROWL, GROWL_HOST, GROWL_PASSWORD, USE_PROWL, PROWL_NOTIFY_ONSNATCH, PROWL_NOTIFY_ONDOWNLOAD, PROWL_NOTIFY_ONSUBTITLEDOWNLOAD, PROWL_API, PROWL_PRIORITY, PROG_DIR, \
|
||||
USE_PYTIVO, PYTIVO_NOTIFY_ONSNATCH, PYTIVO_NOTIFY_ONDOWNLOAD, PYTIVO_NOTIFY_ONSUBTITLEDOWNLOAD, PYTIVO_UPDATE_LIBRARY, PYTIVO_HOST, PYTIVO_SHARE_NAME, PYTIVO_TIVO_NAME, \
|
||||
|
@ -485,12 +485,12 @@ def initialize(consoleLogging=True):
|
|||
USE_PUSHALOT, PUSHALOT_NOTIFY_ONSNATCH, PUSHALOT_NOTIFY_ONDOWNLOAD, PUSHALOT_NOTIFY_ONSUBTITLEDOWNLOAD, PUSHALOT_AUTHORIZATIONTOKEN, \
|
||||
USE_PUSHBULLET, PUSHBULLET_NOTIFY_ONSNATCH, PUSHBULLET_NOTIFY_ONDOWNLOAD, PUSHBULLET_NOTIFY_ONSUBTITLEDOWNLOAD, PUSHBULLET_API, PUSHBULLET_DEVICE, \
|
||||
versionCheckScheduler, VERSION_NOTIFY, AUTO_UPDATE, NOTIFY_ON_UPDATE, PROCESS_AUTOMATICALLY, UNPACK, CPU_PRESET, \
|
||||
KEEP_PROCESSED_DIR, PROCESS_METHOD, TV_DOWNLOAD_DIR, MIN_DAILYSEARCH_FREQUENCY, DEFAULT_UPDATE_FREQUENCY, MIN_UPDATE_FREQUENCY, UPDATE_FREQUENCY, \
|
||||
KEEP_PROCESSED_DIR, PROCESS_METHOD, TV_DOWNLOAD_DIR, MIN_RECENTSEARCH_FREQUENCY, DEFAULT_UPDATE_FREQUENCY, MIN_UPDATE_FREQUENCY, UPDATE_FREQUENCY, \
|
||||
showQueueScheduler, searchQueueScheduler, ROOT_DIRS, CACHE_DIR, ACTUAL_CACHE_DIR, TIMEZONE_DISPLAY, \
|
||||
NAMING_PATTERN, NAMING_MULTI_EP, NAMING_ANIME_MULTI_EP, NAMING_FORCE_FOLDERS, NAMING_ABD_PATTERN, NAMING_CUSTOM_ABD, NAMING_SPORTS_PATTERN, NAMING_CUSTOM_SPORTS, NAMING_ANIME_PATTERN, NAMING_CUSTOM_ANIME, NAMING_STRIP_YEAR, \
|
||||
RENAME_EPISODES, AIRDATE_EPISODES, properFinderScheduler, PROVIDER_ORDER, autoPostProcesserScheduler, \
|
||||
WOMBLE, OMGWTFNZBS, OMGWTFNZBS_USERNAME, OMGWTFNZBS_APIKEY, providerList, newznabProviderList, torrentRssProviderList, \
|
||||
EXTRA_SCRIPTS, USE_TWITTER, TWITTER_USERNAME, TWITTER_PASSWORD, TWITTER_PREFIX, DAILYSEARCH_FREQUENCY, \
|
||||
EXTRA_SCRIPTS, USE_TWITTER, TWITTER_USERNAME, TWITTER_PASSWORD, TWITTER_PREFIX, RECENTSEARCH_FREQUENCY, \
|
||||
USE_BOXCAR, BOXCAR_USERNAME, BOXCAR_PASSWORD, BOXCAR_NOTIFY_ONDOWNLOAD, BOXCAR_NOTIFY_ONSUBTITLEDOWNLOAD, BOXCAR_NOTIFY_ONSNATCH, \
|
||||
USE_BOXCAR2, BOXCAR2_ACCESSTOKEN, BOXCAR2_NOTIFY_ONDOWNLOAD, BOXCAR2_NOTIFY_ONSUBTITLEDOWNLOAD, BOXCAR2_NOTIFY_ONSNATCH, \
|
||||
USE_PUSHOVER, PUSHOVER_USERKEY, PUSHOVER_APIKEY, PUSHOVER_NOTIFY_ONDOWNLOAD, PUSHOVER_NOTIFY_ONSUBTITLEDOWNLOAD, PUSHOVER_NOTIFY_ONSNATCH, \
|
||||
|
@ -498,7 +498,7 @@ def initialize(consoleLogging=True):
|
|||
USE_SYNOLOGYNOTIFIER, SYNOLOGYNOTIFIER_NOTIFY_ONSNATCH, SYNOLOGYNOTIFIER_NOTIFY_ONDOWNLOAD, SYNOLOGYNOTIFIER_NOTIFY_ONSUBTITLEDOWNLOAD, \
|
||||
USE_EMAIL, EMAIL_HOST, EMAIL_PORT, EMAIL_TLS, EMAIL_USER, EMAIL_PASSWORD, EMAIL_FROM, EMAIL_NOTIFY_ONSNATCH, EMAIL_NOTIFY_ONDOWNLOAD, EMAIL_NOTIFY_ONSUBTITLEDOWNLOAD, EMAIL_LIST, \
|
||||
USE_LISTVIEW, METADATA_XBMC, METADATA_XBMC_12PLUS, METADATA_MEDIABROWSER, METADATA_PS3, metadata_provider_dict, \
|
||||
NEWZBIN, NEWZBIN_USERNAME, NEWZBIN_PASSWORD, GIT_PATH, MOVE_ASSOCIATED_FILES, POSTPONE_IF_SYNC_FILES, dailySearchScheduler, NFO_RENAME, \
|
||||
NEWZBIN, NEWZBIN_USERNAME, NEWZBIN_PASSWORD, GIT_PATH, MOVE_ASSOCIATED_FILES, POSTPONE_IF_SYNC_FILES, recentSearchScheduler, NFO_RENAME, \
|
||||
GUI_NAME, HOME_LAYOUT, HISTORY_LAYOUT, DISPLAY_SHOW_SPECIALS, COMING_EPS_LAYOUT, COMING_EPS_SORT, COMING_EPS_DISPLAY_PAUSED, COMING_EPS_MISSED_RANGE, FUZZY_DATING, TRIM_ZERO, DATE_PRESET, TIME_PRESET, TIME_PRESET_W_SECONDS, THEME_NAME, \
|
||||
POSTER_SORTBY, POSTER_SORTDIR, \
|
||||
METADATA_WDTV, METADATA_TIVO, METADATA_MEDE8ER, IGNORE_WORDS, REQUIRE_WORDS, CALENDAR_UNPROTECTED, CREATE_MISSING_SHOW_DIRS, \
|
||||
|
@ -674,7 +674,7 @@ def initialize(consoleLogging=True):
|
|||
|
||||
ALLOW_HIGH_PRIORITY = bool(check_setting_int(CFG, 'General', 'allow_high_priority', 1))
|
||||
|
||||
DAILYSEARCH_STARTUP = bool(check_setting_int(CFG, 'General', 'dailysearch_startup', 1))
|
||||
RECENTSEARCH_STARTUP = bool(check_setting_int(CFG, 'General', 'recentsearch_startup', 1))
|
||||
BACKLOG_STARTUP = bool(check_setting_int(CFG, 'General', 'backlog_startup', 1))
|
||||
SKIP_REMOVED_FILES = bool(check_setting_int(CFG, 'General', 'skip_removed_files', 0))
|
||||
|
||||
|
@ -685,10 +685,10 @@ def initialize(consoleLogging=True):
|
|||
if AUTOPOSTPROCESSER_FREQUENCY < MIN_AUTOPOSTPROCESSER_FREQUENCY:
|
||||
AUTOPOSTPROCESSER_FREQUENCY = MIN_AUTOPOSTPROCESSER_FREQUENCY
|
||||
|
||||
DAILYSEARCH_FREQUENCY = check_setting_int(CFG, 'General', 'dailysearch_frequency',
|
||||
DEFAULT_DAILYSEARCH_FREQUENCY)
|
||||
if DAILYSEARCH_FREQUENCY < MIN_DAILYSEARCH_FREQUENCY:
|
||||
DAILYSEARCH_FREQUENCY = MIN_DAILYSEARCH_FREQUENCY
|
||||
RECENTSEARCH_FREQUENCY = check_setting_int(CFG, 'General', 'recentsearch_frequency',
|
||||
DEFAULT_RECENTSEARCH_FREQUENCY)
|
||||
if RECENTSEARCH_FREQUENCY < MIN_RECENTSEARCH_FREQUENCY:
|
||||
RECENTSEARCH_FREQUENCY = MIN_RECENTSEARCH_FREQUENCY
|
||||
|
||||
MIN_BACKLOG_FREQUENCY = get_backlog_cycle_time()
|
||||
BACKLOG_FREQUENCY = check_setting_int(CFG, 'General', 'backlog_frequency', DEFAULT_BACKLOG_FREQUENCY)
|
||||
|
@ -1022,11 +1022,10 @@ def initialize(consoleLogging=True):
|
|||
curTorrentProvider.getID() + '_search_fallback',
|
||||
0))
|
||||
|
||||
if hasattr(curTorrentProvider, 'enable_daily'):
|
||||
curTorrentProvider.enable_daily = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_enable_daily',
|
||||
1))
|
||||
|
||||
if hasattr(curTorrentProvider, 'enable_recentsearch'):
|
||||
curTorrentProvider.enable_recentsearch = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() +
|
||||
'_enable_recentsearch', 1))
|
||||
if hasattr(curTorrentProvider, 'enable_backlog'):
|
||||
curTorrentProvider.enable_backlog = bool(check_setting_int(CFG, curTorrentProvider.getID().upper(),
|
||||
curTorrentProvider.getID() + '_enable_backlog',
|
||||
|
@ -1050,11 +1049,10 @@ def initialize(consoleLogging=True):
|
|||
curNzbProvider.search_fallback = bool(check_setting_int(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_search_fallback',
|
||||
0))
|
||||
if hasattr(curNzbProvider, 'enable_daily'):
|
||||
curNzbProvider.enable_daily = bool(check_setting_int(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_enable_daily',
|
||||
1))
|
||||
|
||||
if hasattr(curNzbProvider, 'enable_recentsearch'):
|
||||
curNzbProvider.enable_recentsearch = bool(check_setting_int(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_enable_recentsearch',
|
||||
1))
|
||||
if hasattr(curNzbProvider, 'enable_backlog'):
|
||||
curNzbProvider.enable_backlog = bool(check_setting_int(CFG, curNzbProvider.getID().upper(),
|
||||
curNzbProvider.getID() + '_enable_backlog',
|
||||
|
@ -1124,11 +1122,11 @@ def initialize(consoleLogging=True):
|
|||
cycleTime=datetime.timedelta(seconds=3),
|
||||
threadName="SEARCHQUEUE")
|
||||
|
||||
update_interval = datetime.timedelta(minutes=DAILYSEARCH_FREQUENCY)
|
||||
dailySearchScheduler = scheduler.Scheduler(dailysearcher.DailySearcher(),
|
||||
update_interval = datetime.timedelta(minutes=RECENTSEARCH_FREQUENCY)
|
||||
recentSearchScheduler = scheduler.Scheduler(searchRecent.RecentSearcher(),
|
||||
cycleTime=update_interval,
|
||||
threadName="DAILYSEARCHER",
|
||||
run_delay=update_now if DAILYSEARCH_STARTUP
|
||||
threadName="RECENTSEARCHER",
|
||||
run_delay=update_now if RECENTSEARCH_STARTUP
|
||||
else update_interval)
|
||||
|
||||
update_interval = datetime.timedelta(minutes=BACKLOG_FREQUENCY)
|
||||
|
@ -1181,15 +1179,15 @@ def start():
|
|||
showUpdateScheduler, versionCheckScheduler, showQueueScheduler, \
|
||||
properFinderScheduler, autoPostProcesserScheduler, searchQueueScheduler, \
|
||||
subtitlesFinderScheduler, USE_SUBTITLES, traktCheckerScheduler, \
|
||||
dailySearchScheduler, events, started
|
||||
recentSearchScheduler, events, started
|
||||
|
||||
with INIT_LOCK:
|
||||
if __INITIALIZED__:
|
||||
# start sysetm events queue
|
||||
events.start()
|
||||
|
||||
# start the daily search scheduler
|
||||
dailySearchScheduler.start()
|
||||
# start the recent search scheduler
|
||||
recentSearchScheduler.start()
|
||||
|
||||
# start the backlog scheduler
|
||||
backlogSearchScheduler.start()
|
||||
|
@ -1230,7 +1228,7 @@ def halt():
|
|||
showUpdateScheduler, versionCheckScheduler, showQueueScheduler, \
|
||||
properFinderScheduler, autoPostProcesserScheduler, searchQueueScheduler, \
|
||||
subtitlesFinderScheduler, traktCheckerScheduler, \
|
||||
dailySearchScheduler, events, started
|
||||
recentSearchScheduler, events, started
|
||||
|
||||
with INIT_LOCK:
|
||||
|
||||
|
@ -1245,10 +1243,10 @@ def halt():
|
|||
except:
|
||||
pass
|
||||
|
||||
dailySearchScheduler.stop.set()
|
||||
logger.log(u"Waiting for the DAILYSEARCH thread to exit")
|
||||
recentSearchScheduler.stop.set()
|
||||
logger.log(u"Waiting for the RECENTSEARCH thread to exit")
|
||||
try:
|
||||
dailySearchScheduler.join(10)
|
||||
recentSearchScheduler.join(10)
|
||||
except:
|
||||
pass
|
||||
|
||||
|
@ -1397,13 +1395,13 @@ def save_config():
|
|||
new_config['General']['torrent_method'] = TORRENT_METHOD
|
||||
new_config['General']['usenet_retention'] = int(USENET_RETENTION)
|
||||
new_config['General']['autopostprocesser_frequency'] = int(AUTOPOSTPROCESSER_FREQUENCY)
|
||||
new_config['General']['dailysearch_frequency'] = int(DAILYSEARCH_FREQUENCY)
|
||||
new_config['General']['recentsearch_frequency'] = int(RECENTSEARCH_FREQUENCY)
|
||||
new_config['General']['backlog_frequency'] = int(BACKLOG_FREQUENCY)
|
||||
new_config['General']['update_frequency'] = int(UPDATE_FREQUENCY)
|
||||
new_config['General']['download_propers'] = int(DOWNLOAD_PROPERS)
|
||||
new_config['General']['check_propers_interval'] = CHECK_PROPERS_INTERVAL
|
||||
new_config['General']['allow_high_priority'] = int(ALLOW_HIGH_PRIORITY)
|
||||
new_config['General']['dailysearch_startup'] = int(DAILYSEARCH_STARTUP)
|
||||
new_config['General']['recentsearch_startup'] = int(RECENTSEARCH_STARTUP)
|
||||
new_config['General']['backlog_startup'] = int(BACKLOG_STARTUP)
|
||||
new_config['General']['skip_removed_files'] = int(SKIP_REMOVED_FILES)
|
||||
new_config['General']['quality_default'] = int(QUALITY_DEFAULT)
|
||||
|
@ -1526,9 +1524,9 @@ def save_config():
|
|||
if hasattr(curTorrentProvider, 'search_fallback'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_search_fallback'] = int(
|
||||
curTorrentProvider.search_fallback)
|
||||
if hasattr(curTorrentProvider, 'enable_daily'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_enable_daily'] = int(
|
||||
curTorrentProvider.enable_daily)
|
||||
if hasattr(curTorrentProvider, 'enable_recentsearch'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_enable_recentsearch'] = int(
|
||||
curTorrentProvider.enable_recentsearch)
|
||||
if hasattr(curTorrentProvider, 'enable_backlog'):
|
||||
new_config[curTorrentProvider.getID().upper()][curTorrentProvider.getID() + '_enable_backlog'] = int(
|
||||
curTorrentProvider.enable_backlog)
|
||||
|
@ -1550,9 +1548,9 @@ def save_config():
|
|||
if hasattr(curNzbProvider, 'search_fallback'):
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID() + '_search_fallback'] = int(
|
||||
curNzbProvider.search_fallback)
|
||||
if hasattr(curNzbProvider, 'enable_daily'):
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID() + '_enable_daily'] = int(
|
||||
curNzbProvider.enable_daily)
|
||||
if hasattr(curNzbProvider, 'enable_recentsearch'):
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID() + '_enable_recentsearch'] = int(
|
||||
curNzbProvider.enable_recentsearch)
|
||||
if hasattr(curNzbProvider, 'enable_backlog'):
|
||||
new_config[curNzbProvider.getID().upper()][curNzbProvider.getID() + '_enable_backlog'] = int(
|
||||
curNzbProvider.enable_backlog)
|
||||
|
|
|
@ -27,6 +27,8 @@ from sickbeard import helpers
|
|||
from sickbeard import logger
|
||||
from sickbeard import naming
|
||||
from sickbeard import db
|
||||
from sickbeard import providers
|
||||
from sickbeard.providers.generic import GenericProvider
|
||||
|
||||
naming_ep_type = ("%(seasonnumber)dx%(episodenumber)02d",
|
||||
"s%(seasonnumber)02de%(episodenumber)02d",
|
||||
|
@ -155,13 +157,13 @@ def change_AUTOPOSTPROCESSER_FREQUENCY(freq):
|
|||
|
||||
sickbeard.autoPostProcesserScheduler.cycleTime = datetime.timedelta(minutes=sickbeard.AUTOPOSTPROCESSER_FREQUENCY)
|
||||
|
||||
def change_DAILYSEARCH_FREQUENCY(freq):
|
||||
sickbeard.DAILYSEARCH_FREQUENCY = to_int(freq, default=sickbeard.DEFAULT_DAILYSEARCH_FREQUENCY)
|
||||
def change_RECENTSEARCH_FREQUENCY(freq):
|
||||
sickbeard.RECENTSEARCH_FREQUENCY = to_int(freq, default=sickbeard.DEFAULT_RECENTSEARCH_FREQUENCY)
|
||||
|
||||
if sickbeard.DAILYSEARCH_FREQUENCY < sickbeard.MIN_DAILYSEARCH_FREQUENCY:
|
||||
sickbeard.DAILYSEARCH_FREQUENCY = sickbeard.MIN_DAILYSEARCH_FREQUENCY
|
||||
if sickbeard.RECENTSEARCH_FREQUENCY < sickbeard.MIN_RECENTSEARCH_FREQUENCY:
|
||||
sickbeard.RECENTSEARCH_FREQUENCY = sickbeard.MIN_RECENTSEARCH_FREQUENCY
|
||||
|
||||
sickbeard.dailySearchScheduler.cycleTime = datetime.timedelta(minutes=sickbeard.DAILYSEARCH_FREQUENCY)
|
||||
sickbeard.recentSearchScheduler.cycleTime = datetime.timedelta(minutes=sickbeard.RECENTSEARCH_FREQUENCY)
|
||||
|
||||
def change_BACKLOG_FREQUENCY(freq):
|
||||
sickbeard.BACKLOG_FREQUENCY = to_int(freq, default=sickbeard.DEFAULT_BACKLOG_FREQUENCY)
|
||||
|
@ -445,7 +447,8 @@ class ConfigMigrator():
|
|||
2: 'Sync backup number with version number',
|
||||
3: 'Rename omgwtfnzb variables',
|
||||
4: 'Add newznab catIDs',
|
||||
5: 'Metadata update'
|
||||
5: 'Metadata update',
|
||||
6: 'Rename daily search to recent search'
|
||||
}
|
||||
|
||||
def migrate_config(self):
|
||||
|
@ -712,3 +715,18 @@ class ConfigMigrator():
|
|||
sickbeard.METADATA_WDTV = _migrate_metadata(metadata_wdtv, 'WDTV', use_banner)
|
||||
sickbeard.METADATA_TIVO = _migrate_metadata(metadata_tivo, 'TIVO', use_banner)
|
||||
sickbeard.METADATA_MEDE8ER = _migrate_metadata(metadata_mede8er, 'Mede8er', use_banner)
|
||||
|
||||
# Migration v6: Rename daily search to recent search
|
||||
def _migrate_v6(self):
|
||||
|
||||
sickbeard.RECENTSEARCH_FREQUENCY = check_setting_int(self.config_obj, 'General', 'dailysearch_frequency',
|
||||
sickbeard.DEFAULT_RECENTSEARCH_FREQUENCY)
|
||||
|
||||
sickbeard.RECENTSEARCH_STARTUP = bool(check_setting_int(self.config_obj, 'General', 'dailysearch_startup', 1))
|
||||
if sickbeard.RECENTSEARCH_FREQUENCY < sickbeard.MIN_RECENTSEARCH_FREQUENCY:
|
||||
sickbeard.RECENTSEARCH_FREQUENCY = sickbeard.MIN_RECENTSEARCH_FREQUENCY
|
||||
|
||||
for curProvider in providers.sortedProviderList():
|
||||
if hasattr(curProvider, 'enable_recentsearch'):
|
||||
curProvider.enable_recentsearch = bool(check_setting_int(self.config_obj, curProvider.getID().upper(),
|
||||
curProvider.getID() + '_enable_dailysearch', 1))
|
|
@ -97,7 +97,7 @@ def getNewznabProviderList(data):
|
|||
providerDict[curDefault.name].needs_auth = curDefault.needs_auth
|
||||
providerDict[curDefault.name].search_mode = curDefault.search_mode
|
||||
providerDict[curDefault.name].search_fallback = curDefault.search_fallback
|
||||
providerDict[curDefault.name].enable_daily = curDefault.enable_daily
|
||||
providerDict[curDefault.name].enable_recentsearch = curDefault.enable_recentsearch
|
||||
providerDict[curDefault.name].enable_backlog = curDefault.enable_backlog
|
||||
|
||||
return filter(lambda x: x, providerList)
|
||||
|
@ -109,13 +109,13 @@ def makeNewznabProvider(configString):
|
|||
|
||||
search_mode = 'eponly'
|
||||
search_fallback = 0
|
||||
enable_daily = 0
|
||||
enable_recentsearch = 0
|
||||
enable_backlog = 0
|
||||
|
||||
try:
|
||||
values = configString.split('|')
|
||||
if len(values) == 9:
|
||||
name, url, key, catIDs, enabled, search_mode, search_fallback, enable_daily, enable_backlog = values
|
||||
name, url, key, catIDs, enabled, search_mode, search_fallback, enable_recentsearch, enable_backlog = values
|
||||
else:
|
||||
name = values[0]
|
||||
url = values[1]
|
||||
|
@ -129,7 +129,7 @@ def makeNewznabProvider(configString):
|
|||
newznab = sys.modules['sickbeard.providers.newznab']
|
||||
|
||||
newProvider = newznab.NewznabProvider(name, url, key=key, catIDs=catIDs, search_mode=search_mode,
|
||||
search_fallback=search_fallback, enable_daily=enable_daily,
|
||||
search_fallback=search_fallback, enable_recentsearch=enable_recentsearch,
|
||||
enable_backlog=enable_backlog)
|
||||
newProvider.enabled = enabled == '1'
|
||||
|
||||
|
@ -157,13 +157,13 @@ def makeTorrentRssProvider(configString):
|
|||
cookies = None
|
||||
search_mode = 'eponly'
|
||||
search_fallback = 0
|
||||
enable_daily = 0
|
||||
enable_recentsearch = 0
|
||||
enable_backlog = 0
|
||||
|
||||
try:
|
||||
values = configString.split('|')
|
||||
if len(values) == 8:
|
||||
name, url, cookies, enabled, search_mode, search_fallback, enable_daily, enable_backlog = values
|
||||
name, url, cookies, enabled, search_mode, search_fallback, enable_recentsearch, enable_backlog = values
|
||||
else:
|
||||
name = values[0]
|
||||
url = values[1]
|
||||
|
@ -178,7 +178,7 @@ def makeTorrentRssProvider(configString):
|
|||
except:
|
||||
return
|
||||
|
||||
newProvider = torrentRss.TorrentRssProvider(name, url, cookies, search_mode, search_fallback, enable_daily,
|
||||
newProvider = torrentRss.TorrentRssProvider(name, url, cookies, search_mode, search_fallback, enable_recentsearch,
|
||||
enable_backlog)
|
||||
newProvider.enabled = enabled == '1'
|
||||
|
||||
|
|
|
@ -56,7 +56,7 @@ class GenericProvider:
|
|||
|
||||
self.search_mode = None
|
||||
self.search_fallback = False
|
||||
self.enable_daily = False
|
||||
self.enable_recentsearch = False
|
||||
self.enable_backlog = False
|
||||
|
||||
self.cache = tvcache.TVCache(self)
|
||||
|
|
|
@ -43,7 +43,7 @@ from lib.bencode import bdecode
|
|||
|
||||
class NewznabProvider(generic.NZBProvider):
|
||||
def __init__(self, name, url, key='', catIDs='5030,5040', search_mode='eponly', search_fallback=False,
|
||||
enable_daily=False, enable_backlog=False):
|
||||
enable_recentsearch=False, enable_backlog=False):
|
||||
|
||||
generic.NZBProvider.__init__(self, name)
|
||||
|
||||
|
@ -55,7 +55,7 @@ class NewznabProvider(generic.NZBProvider):
|
|||
|
||||
self.search_mode = search_mode
|
||||
self.search_fallback = search_fallback
|
||||
self.enable_daily = enable_daily
|
||||
self.enable_recentsearch = enable_recentsearch
|
||||
self.enable_backlog = enable_backlog
|
||||
|
||||
# a 0 in the key spot indicates that no key is needed
|
||||
|
@ -77,7 +77,7 @@ class NewznabProvider(generic.NZBProvider):
|
|||
def configStr(self):
|
||||
return self.name + '|' + self.url + '|' + self.key + '|' + self.catIDs + '|' + str(
|
||||
int(self.enabled)) + '|' + self.search_mode + '|' + str(int(self.search_fallback)) + '|' + str(
|
||||
int(self.enable_daily)) + '|' + str(int(self.enable_backlog))
|
||||
int(self.enable_recentsearch)) + '|' + str(int(self.enable_backlog))
|
||||
|
||||
def imageName(self):
|
||||
if ek.ek(os.path.isfile,
|
||||
|
|
|
@ -36,7 +36,7 @@ from lib.bencode import bdecode
|
|||
|
||||
|
||||
class TorrentRssProvider(generic.TorrentProvider):
|
||||
def __init__(self, name, url, cookies='', search_mode='eponly', search_fallback=False, enable_daily=False,
|
||||
def __init__(self, name, url, cookies='', search_mode='eponly', search_fallback=False, enable_recentsearch=False,
|
||||
enable_backlog=False):
|
||||
generic.TorrentProvider.__init__(self, name)
|
||||
self.cache = TorrentRssCache(self)
|
||||
|
@ -48,7 +48,7 @@ class TorrentRssProvider(generic.TorrentProvider):
|
|||
|
||||
self.search_mode = search_mode
|
||||
self.search_fallback = search_fallback
|
||||
self.enable_daily = enable_daily
|
||||
self.enable_recentsearch = enable_recentsearch
|
||||
self.enable_backlog = enable_backlog
|
||||
self.cookies = cookies
|
||||
|
||||
|
@ -59,7 +59,7 @@ class TorrentRssProvider(generic.TorrentProvider):
|
|||
self.enabled,
|
||||
self.search_mode or '',
|
||||
self.search_fallback,
|
||||
self.enable_daily,
|
||||
self.enable_recentsearch,
|
||||
self.enable_backlog)
|
||||
|
||||
def imageName(self):
|
||||
|
|
|
@ -375,7 +375,7 @@ def searchForNeededEpisodes():
|
|||
|
||||
episodes.extend(wantedEpisodes(curShow, fromDate))
|
||||
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.isActive() and x.enable_daily]
|
||||
providers = [x for x in sickbeard.providers.sortedProviderList() if x.isActive() and x.enable_recentsearch]
|
||||
for curProvider in providers:
|
||||
|
||||
# spawn separate threads for each provider so we don't need to wait for providers with slow network operation
|
||||
|
@ -435,7 +435,7 @@ def searchForNeededEpisodes():
|
|||
|
||||
if not didSearch:
|
||||
logger.log(
|
||||
u"No NZB/Torrent providers found or enabled in the SickGear config for daily searches. Please check your settings.",
|
||||
u"No NZB/Torrent providers found or enabled in the SickGear config for recent searches. Please check your settings.",
|
||||
logger.ERROR)
|
||||
|
||||
return foundResults.values()
|
||||
|
|
|
@ -32,7 +32,7 @@ from sickbeard import network_timezones
|
|||
from sickbeard.exceptions import ex
|
||||
|
||||
|
||||
class DailySearcher():
|
||||
class RecentSearcher():
|
||||
def __init__(self):
|
||||
self.lock = threading.Lock()
|
||||
self.amActive = False
|
||||
|
@ -97,8 +97,8 @@ class DailySearcher():
|
|||
myDB = db.DBConnection()
|
||||
myDB.mass_action(sql_l)
|
||||
|
||||
# queue episode for daily search
|
||||
dailysearch_queue_item = sickbeard.search_queue.DailySearchQueueItem()
|
||||
sickbeard.searchQueueScheduler.action.add_item(dailysearch_queue_item)
|
||||
# queue episode for recent search
|
||||
recentsearch_queue_item = sickbeard.search_queue.RecentSearchQueueItem()
|
||||
sickbeard.searchQueueScheduler.action.add_item(recentsearch_queue_item)
|
||||
|
||||
self.amActive = False
|
|
@ -33,7 +33,7 @@ from sickbeard.search import pickBestResult
|
|||
search_queue_lock = threading.Lock()
|
||||
|
||||
BACKLOG_SEARCH = 10
|
||||
DAILY_SEARCH = 20
|
||||
RECENT_SEARCH = 20
|
||||
FAILED_SEARCH = 30
|
||||
MANUAL_SEARCH = 40
|
||||
|
||||
|
@ -95,17 +95,17 @@ class SearchQueue(generic_queue.GenericQueue):
|
|||
return True
|
||||
return False
|
||||
|
||||
def is_dailysearch_in_progress(self):
|
||||
def is_recentsearch_in_progress(self):
|
||||
for cur_item in self.queue + [self.currentItem]:
|
||||
if isinstance(cur_item, DailySearchQueueItem):
|
||||
if isinstance(cur_item, RecentSearchQueueItem):
|
||||
return True
|
||||
return False
|
||||
|
||||
def queue_length(self):
|
||||
length = {'backlog': 0, 'daily': 0, 'manual': 0, 'failed': 0}
|
||||
length = {'backlog': 0, 'recent': 0, 'manual': 0, 'failed': 0}
|
||||
for cur_item in self.queue:
|
||||
if isinstance(cur_item, DailySearchQueueItem):
|
||||
length['daily'] += 1
|
||||
if isinstance(cur_item, RecentSearchQueueItem):
|
||||
length['recent'] += 1
|
||||
elif isinstance(cur_item, BacklogQueueItem):
|
||||
length['backlog'] += 1
|
||||
elif isinstance(cur_item, ManualSearchQueueItem):
|
||||
|
@ -116,8 +116,8 @@ class SearchQueue(generic_queue.GenericQueue):
|
|||
|
||||
|
||||
def add_item(self, item):
|
||||
if isinstance(item, DailySearchQueueItem):
|
||||
# daily searches
|
||||
if isinstance(item, RecentSearchQueueItem):
|
||||
# recent searches
|
||||
generic_queue.GenericQueue.add_item(self, item)
|
||||
elif isinstance(item, BacklogQueueItem) and not self.is_in_queue(item.show, item.segment):
|
||||
# backlog searches
|
||||
|
@ -128,16 +128,16 @@ class SearchQueue(generic_queue.GenericQueue):
|
|||
else:
|
||||
logger.log(u"Not adding item, it's already in the queue", logger.DEBUG)
|
||||
|
||||
class DailySearchQueueItem(generic_queue.QueueItem):
|
||||
class RecentSearchQueueItem(generic_queue.QueueItem):
|
||||
def __init__(self):
|
||||
self.success = None
|
||||
generic_queue.QueueItem.__init__(self, 'Daily Search', DAILY_SEARCH)
|
||||
generic_queue.QueueItem.__init__(self, 'Recent Search', RECENT_SEARCH)
|
||||
|
||||
def run(self):
|
||||
generic_queue.QueueItem.run(self)
|
||||
|
||||
try:
|
||||
logger.log("Beginning daily search for new episodes")
|
||||
logger.log("Beginning recent search for new episodes")
|
||||
foundResults = search.searchForNeededEpisodes()
|
||||
|
||||
if not len(foundResults):
|
||||
|
|
|
@ -97,7 +97,7 @@ class TVCache():
|
|||
myDB.action('DELETE FROM [' + self.providerID + '] WHERE 1')
|
||||
|
||||
def _get_title_and_url(self, item):
|
||||
# override this in the provider if daily search has a different data layout to backlog searches
|
||||
# override this in the provider if recent search has a different data layout to backlog searches
|
||||
return self.provider._get_title_and_url(item)
|
||||
|
||||
def _getRSSData(self):
|
||||
|
@ -222,7 +222,7 @@ class TVCache():
|
|||
return True
|
||||
|
||||
def shouldClearCache(self):
|
||||
# if daily search hasn't used our previous results yet then don't clear the cache
|
||||
# if recent search hasn't used our previous results yet then don't clear the cache
|
||||
if self.lastUpdate > self.lastSearch:
|
||||
return False
|
||||
|
||||
|
|
|
@ -630,7 +630,7 @@ class ManageSearches(MainHandler):
|
|||
# t.backlogPI = sickbeard.backlogSearchScheduler.action.getProgressIndicator()
|
||||
t.backlogPaused = sickbeard.searchQueueScheduler.action.is_backlog_paused() # @UndefinedVariable
|
||||
t.backlogRunning = sickbeard.searchQueueScheduler.action.is_backlog_in_progress() # @UndefinedVariable
|
||||
t.dailySearchStatus = sickbeard.dailySearchScheduler.action.amActive # @UndefinedVariable
|
||||
t.recentSearchStatus = sickbeard.recentSearchScheduler.action.amActive # @UndefinedVariable
|
||||
t.findPropersStatus = sickbeard.properFinderScheduler.action.amActive # @UndefinedVariable
|
||||
t.queueLength = sickbeard.searchQueueScheduler.action.queue_length()
|
||||
|
||||
|
@ -658,10 +658,10 @@ class ManageSearches(MainHandler):
|
|||
def forceSearch(self, *args, **kwargs):
|
||||
|
||||
# force it to run the next time it looks
|
||||
result = sickbeard.dailySearchScheduler.forceRun()
|
||||
result = sickbeard.recentSearchScheduler.forceRun()
|
||||
if result:
|
||||
logger.log(u"Daily search forced")
|
||||
ui.notifications.message('Daily search started')
|
||||
logger.log(u"Recent search forced")
|
||||
ui.notifications.message('Recent search started')
|
||||
|
||||
redirect("/manage/manageSearches/")
|
||||
|
||||
|
@ -1628,10 +1628,10 @@ class ConfigSearch(MainHandler):
|
|||
def saveSearch(self, use_nzbs=None, use_torrents=None, nzb_dir=None, sab_username=None, sab_password=None,
|
||||
sab_apikey=None, sab_category=None, sab_host=None, nzbget_username=None, nzbget_password=None,
|
||||
nzbget_category=None, nzbget_priority=None, nzbget_host=None, nzbget_use_https=None,
|
||||
backlog_days=None, backlog_frequency=None, dailysearch_frequency=None,
|
||||
backlog_days=None, backlog_frequency=None, recentsearch_frequency=None,
|
||||
nzb_method=None, torrent_method=None, usenet_retention=None,
|
||||
download_propers=None, check_propers_interval=None, allow_high_priority=None,
|
||||
backlog_startup=None, dailysearch_startup=None,
|
||||
backlog_startup=None, recentsearch_startup=None,
|
||||
torrent_dir=None, torrent_username=None, torrent_password=None, torrent_host=None,
|
||||
torrent_label=None, torrent_path=None, torrent_verify_cert=None,
|
||||
torrent_seed_time=None, torrent_paused=None, torrent_high_bandwidth=None, ignore_words=None, require_words=None):
|
||||
|
@ -1644,7 +1644,7 @@ class ConfigSearch(MainHandler):
|
|||
if not config.change_TORRENT_DIR(torrent_dir):
|
||||
results += ["Unable to create directory " + os.path.normpath(torrent_dir) + ", dir not changed."]
|
||||
|
||||
config.change_DAILYSEARCH_FREQUENCY(dailysearch_frequency)
|
||||
config.change_RECENTSEARCH_FREQUENCY(recentsearch_frequency)
|
||||
|
||||
config.change_BACKLOG_FREQUENCY(backlog_frequency)
|
||||
sickbeard.BACKLOG_DAYS = config.to_int(backlog_days, default=7)
|
||||
|
@ -1664,7 +1664,7 @@ class ConfigSearch(MainHandler):
|
|||
|
||||
sickbeard.ALLOW_HIGH_PRIORITY = config.checkbox_to_value(allow_high_priority)
|
||||
|
||||
sickbeard.DAILYSEARCH_STARTUP = config.checkbox_to_value(dailysearch_startup)
|
||||
sickbeard.RECENTSEARCH_STARTUP = config.checkbox_to_value(recentsearch_startup)
|
||||
sickbeard.BACKLOG_STARTUP = config.checkbox_to_value(backlog_startup)
|
||||
|
||||
sickbeard.SAB_USERNAME = sab_username
|
||||
|
@ -2092,10 +2092,10 @@ class ConfigProviders(MainHandler):
|
|||
newznabProviderDict[cur_id].search_fallback = 0
|
||||
|
||||
try:
|
||||
newznabProviderDict[cur_id].enable_daily = config.checkbox_to_value(
|
||||
kwargs[cur_id + '_enable_daily'])
|
||||
newznabProviderDict[cur_id].enable_recentsearch = config.checkbox_to_value(
|
||||
kwargs[cur_id + '_enable_recentsearch'])
|
||||
except:
|
||||
newznabProviderDict[cur_id].enable_daily = 0
|
||||
newznabProviderDict[cur_id].enable_recentsearch = 0
|
||||
|
||||
try:
|
||||
newznabProviderDict[cur_id].enable_backlog = config.checkbox_to_value(
|
||||
|
@ -2258,12 +2258,12 @@ class ConfigProviders(MainHandler):
|
|||
except:
|
||||
curTorrentProvider.search_fallback = 0 # these exceptions are catching unselected checkboxes
|
||||
|
||||
if hasattr(curTorrentProvider, 'enable_daily'):
|
||||
if hasattr(curTorrentProvider, 'enable_recentsearch'):
|
||||
try:
|
||||
curTorrentProvider.enable_daily = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_enable_daily'])
|
||||
curTorrentProvider.enable_recentsearch = config.checkbox_to_value(
|
||||
kwargs[curTorrentProvider.getID() + '_enable_recentsearch'])
|
||||
except:
|
||||
curTorrentProvider.enable_daily = 0 # these exceptions are actually catching unselected checkboxes
|
||||
curTorrentProvider.enable_recentsearch = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
if hasattr(curTorrentProvider, 'enable_backlog'):
|
||||
try:
|
||||
|
@ -2300,12 +2300,12 @@ class ConfigProviders(MainHandler):
|
|||
except:
|
||||
curNzbProvider.search_fallback = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
if hasattr(curNzbProvider, 'enable_daily'):
|
||||
if hasattr(curNzbProvider, 'enable_recentsearch'):
|
||||
try:
|
||||
curNzbProvider.enable_daily = config.checkbox_to_value(
|
||||
kwargs[curNzbProvider.getID() + '_enable_daily'])
|
||||
curNzbProvider.enable_recentsearch = config.checkbox_to_value(
|
||||
kwargs[curNzbProvider.getID() + '_enable_recentsearch'])
|
||||
except:
|
||||
curNzbProvider.enable_daily = 0 # these exceptions are actually catching unselected checkboxes
|
||||
curNzbProvider.enable_recentsearch = 0 # these exceptions are actually catching unselected checkboxes
|
||||
|
||||
if hasattr(curNzbProvider, 'enable_backlog'):
|
||||
try:
|
||||
|
|
Loading…
Reference in a new issue