mirror of
https://github.com/SickGear/SickGear.git
synced 2024-12-01 00:43:37 +00:00
d3a7f0ff5e
Add newznab smart logic to avoid missing releases when there are a great many recent releases. Change improve performance by using newznab server advertised capabilities. Change config/providers newznab to display only non-default categories. Change use scene season for wanted segment in backlog if show is scene numbering. Change combine Manage Searches / Backlog Search / Limited and Full to Force. Change consolidate limited and full backlog. Change config / Search / Backlog search frequency to instead spread backlog searches over a number of days. Change migrate minimum used value for search frequency into new minimum 7 for search spread. Change restrict nzb providers to 1 backlog batch run per day. Add to Config/Search/Unaired episodes/Allow episodes that are released early. Add to Config/Search/Unaired episodes/Use specific api requests to search for early episode releases. Add use related ids for newznab searches to increase search efficiency. Add periodic update of related show ids. Change terminology Edit Show/"Post processing" tab name to "Other". Add advanced feature "Related show IDs" to Edit Show/Other used for finding episodes and TV info. Add search info source image links to those that have zero id under Edit Show/Other/"Related show IDs". Add "set master" button to Edit Show/Other/"Related show IDs" for info source that can be changed. Change terminology displayShow "Indexers" to "Links" to cover internal and web links. Change add related show info sources on displayShow page. Change don't display "temporarily" defunct TVRage image link on displayShow pages unless it is master info source. Change if a defunct info source is the master of a show then present a link on displayShow to edit related show IDs. Change simplify the next backlog search run time display in the page footer. Change try ssl when fetching data thetvdb, imdb, trakt, scene exception. Change improve reliability to Trakt notifier by using show related id support. Change improve config/providers newznab categories layout. Change show loaded log message at start up and include info source. Change if episode has no airdate then set status to unaired (was skipped). Technical Change move scene_exceptions table from cache.db to sickbeard.db. Add related ids to show obj. Add use of mapped indexer ids for newznab. Add indexer to sql in wanted_eps. Add aired in (scene) season for wanted episodes. Add need_anime, need_sports, need_sd, need_hd, need_uhd to wanted episodes and added as parameter to update_providers. Add fix for lib lockfile/mkdirlockfile. Add set master TV info source logic. Change harden ui input validation. Add per action dialog confirmation. Change to reload page under more events. Change implement "Mark all added episodes Wanted to search for releases" when setting new info source.
427 lines
19 KiB
Python
427 lines
19 KiB
Python
#
|
|
# This file is part of SickGear.
|
|
#
|
|
# SickGear is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# SickGear is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
import datetime
|
|
import re
|
|
import traceback
|
|
|
|
import requests
|
|
import sickbeard
|
|
from collections import OrderedDict
|
|
from urllib import urlencode
|
|
from lib.dateutil.parser import parse
|
|
from lib.unidecode import unidecode
|
|
from libtrakt import TraktAPI
|
|
from libtrakt.exceptions import TraktAuthException, TraktException
|
|
from sickbeard import db, logger
|
|
from sickbeard.helpers import tryInt, getURL
|
|
from sickbeard.indexers.indexer_config import (INDEXER_TVDB, INDEXER_TVRAGE, INDEXER_TVMAZE,
|
|
INDEXER_IMDB, INDEXER_TRAKT, INDEXER_TMDB)
|
|
from lib.tmdb_api import TMDB
|
|
from lib.imdb import IMDb
|
|
|
|
defunct_indexer = []
|
|
indexer_list = []
|
|
tmdb_ids = {INDEXER_TVDB: 'tvdb_id', INDEXER_IMDB: 'imdb_id', INDEXER_TVRAGE: 'tvrage_id'}
|
|
|
|
|
|
class NewIdDict(dict):
|
|
def __init__(self, *args, **kwargs):
|
|
super(NewIdDict, self).__init__(*args, **kwargs)
|
|
|
|
@staticmethod
|
|
def set_value(value, old_value=None):
|
|
if old_value is MapStatus.MISMATCH or (0 < value and old_value not in [None, value] and 0 < old_value):
|
|
return MapStatus.MISMATCH
|
|
return value
|
|
|
|
@staticmethod
|
|
def get_value(value):
|
|
if value in [None, 0]:
|
|
return MapStatus.NOT_FOUND
|
|
return value
|
|
|
|
def __getitem__(self, key):
|
|
return self.get_value(super(NewIdDict, self).get(key))
|
|
|
|
def get(self, key, default=None):
|
|
return self.get_value(super(NewIdDict, self).get(key, default))
|
|
|
|
def __setitem__(self, key, value):
|
|
super(NewIdDict, self).__setitem__(key, self.set_value(value, self.get(key)))
|
|
|
|
def update(self, other=None, **kwargs):
|
|
if isinstance(other, dict):
|
|
other = {o: self.set_value(v, self.get(o)) for o, v in other.iteritems()}
|
|
super(NewIdDict, self).update(other, **kwargs)
|
|
|
|
|
|
class TvmazeDict(OrderedDict):
|
|
tvmaze_ids = {INDEXER_TVDB: 'thetvdb', INDEXER_IMDB: 'imdb', INDEXER_TVRAGE: 'tvrage'}
|
|
|
|
def __init__(self, *args, **kwds):
|
|
super(TvmazeDict, self).__init__(*args, **kwds)
|
|
|
|
def get_url(self, key):
|
|
if INDEXER_TVMAZE == key:
|
|
return '%sshows/%s' % (sickbeard.indexerApi(INDEXER_TVMAZE).config['base_url'], self.tvmaze_ids[key])
|
|
return '%slookup/shows?%s=%s%s' % (sickbeard.indexerApi(INDEXER_TVMAZE).config['base_url'],
|
|
self.tvmaze_ids[key], ('', 'tt')[key == INDEXER_IMDB],
|
|
(self[key], '%07d' % self[key])[key == INDEXER_IMDB])
|
|
|
|
|
|
class TraktDict(OrderedDict):
|
|
trakt_ids = {INDEXER_TVDB: 'tvdb', INDEXER_IMDB: 'imdb', INDEXER_TVRAGE: 'tvrage'}
|
|
|
|
def __init__(self, *args, **kwds):
|
|
super(TraktDict, self).__init__(*args, **kwds)
|
|
|
|
def get_url(self, key):
|
|
return 'search/%s/%s%s?type=show' % (self.trakt_ids[key], ('', 'tt')[key == INDEXER_IMDB],
|
|
(self[key], '%07d' % self[key])[key == INDEXER_IMDB])
|
|
|
|
|
|
def get_tvmaze_ids(url_tvmaze):
|
|
ids = {}
|
|
for url_key in url_tvmaze.iterkeys():
|
|
try:
|
|
res = getURL(url=url_tvmaze.get_url(url_key), json=True, raise_status_code=True, timeout=120)
|
|
if res and 'externals' in res:
|
|
ids[INDEXER_TVRAGE] = res['externals'].get('tvrage', 0)
|
|
ids[INDEXER_TVDB] = res['externals'].get('thetvdb', 0)
|
|
ids[INDEXER_IMDB] = tryInt(str(res['externals'].get('imdb')).replace('tt', ''))
|
|
ids[INDEXER_TVMAZE] = res.get('id', 0)
|
|
break
|
|
except (requests.HTTPError, Exception):
|
|
pass
|
|
return {k: v for k, v in ids.iteritems() if v not in (None, '', 0)}
|
|
|
|
|
|
def get_premieredate(show):
|
|
try:
|
|
first_ep = show.getEpisode(season=1, episode=1)
|
|
if first_ep and first_ep.airdate:
|
|
return first_ep.airdate
|
|
except (StandardError, Exception):
|
|
pass
|
|
return None
|
|
|
|
|
|
def clean_show_name(showname):
|
|
return re.sub(r'[(\s]*(?:19|20)\d\d[)\s]*$', '', isinstance(showname, unicode) and unidecode(showname) or showname)
|
|
|
|
|
|
def get_tvmaze_by_name(showname, premiere_date):
|
|
ids = {}
|
|
try:
|
|
url = '%ssearch/shows?%s' % (sickbeard.indexerApi(INDEXER_TVMAZE).config['base_url'],
|
|
urlencode({'q': clean_show_name(showname)}))
|
|
res = getURL(url=url, json=True, raise_status_code=True, timeout=120)
|
|
if res:
|
|
for r in res:
|
|
if 'show' in r and 'premiered' in r['show'] and 'externals' in r['show']:
|
|
premiered = parse(r['show']['premiered'], fuzzy=True)
|
|
if abs(premiere_date - premiered.date()) < datetime.timedelta(days=2):
|
|
ids[INDEXER_TVRAGE] = r['show']['externals'].get('tvrage', 0)
|
|
ids[INDEXER_TVDB] = r['show']['externals'].get('thetvdb', 0)
|
|
ids[INDEXER_IMDB] = tryInt(str(r['show']['externals'].get('imdb')).replace('tt', ''))
|
|
ids[INDEXER_TVMAZE] = r['show'].get('id', 0)
|
|
break
|
|
except (StandardError, Exception):
|
|
pass
|
|
return {k: v for k, v in ids.iteritems() if v not in (None, '', 0)}
|
|
|
|
|
|
def get_trakt_ids(url_trakt):
|
|
ids = {}
|
|
for url_key in url_trakt.iterkeys():
|
|
try:
|
|
res = TraktAPI().trakt_request(url_trakt.get_url(url_key))
|
|
if res:
|
|
found = False
|
|
for r in res:
|
|
if r.get('type', '') == 'show' and 'show' in r and 'ids' in r['show']:
|
|
ids[INDEXER_TVDB] = tryInt(r['show']['ids'].get('tvdb', 0))
|
|
ids[INDEXER_TVRAGE] = tryInt(r['show']['ids'].get('tvrage', 0))
|
|
ids[INDEXER_IMDB] = tryInt(str(r['show']['ids'].get('imdb')).replace('tt', ''))
|
|
ids[INDEXER_TRAKT] = tryInt(r['show']['ids'].get('trakt', 0))
|
|
ids[INDEXER_TMDB] = tryInt(r['show']['ids'].get('tmdb', 0))
|
|
found = True
|
|
break
|
|
if found:
|
|
break
|
|
except (TraktAuthException, TraktException, IndexError, KeyError):
|
|
pass
|
|
return {k: v for k, v in ids.iteritems() if v not in (None, '', 0)}
|
|
|
|
|
|
def get_imdbid_by_name(name, startyear):
|
|
ids = {}
|
|
try:
|
|
res = IMDb().search_movie(title=name)
|
|
for r in res:
|
|
if hasattr(r, 'movieID') and hasattr(r, 'data') and 'kind' in r.data and r.data['kind'] == 'tv series' \
|
|
and 'year' in r.data and r.data['year'] == startyear:
|
|
ids[INDEXER_IMDB] = tryInt(r.movieID)
|
|
except (StandardError, Exception):
|
|
pass
|
|
return {k: v for k, v in ids.iteritems() if v not in (None, '', 0)}
|
|
|
|
|
|
def map_indexers_to_show(show_obj, update=False, force=False, recheck=False):
|
|
"""
|
|
|
|
:return: mapped ids
|
|
:rtype: dict
|
|
:param show_obj: TVShow Object
|
|
:param update: add missing + previously not found ids
|
|
:param force: search for and replace all mapped/missing ids (excluding NO_AUTOMATIC_CHANGE flagged)
|
|
:param recheck: load all ids, don't remove existing
|
|
"""
|
|
mapped = {}
|
|
|
|
# init mapped indexers object
|
|
for indexer in indexer_list:
|
|
mapped[indexer] = {'id': (0, show_obj.indexerid)[int(indexer) == int(show_obj.indexer)],
|
|
'status': (MapStatus.NONE, MapStatus.SOURCE)[int(indexer) == int(show_obj.indexer)],
|
|
'date': datetime.date.fromordinal(1)}
|
|
|
|
my_db = db.DBConnection()
|
|
sql_results = my_db.select('SELECT' + ' * FROM indexer_mapping WHERE indexer_id = ? AND indexer = ?',
|
|
[show_obj.indexerid, show_obj.indexer])
|
|
|
|
# for each mapped entry
|
|
for curResult in sql_results:
|
|
date = tryInt(curResult['date'])
|
|
mapped[int(curResult['mindexer'])] = {'status': int(curResult['status']),
|
|
'id': int(curResult['mindexer_id']),
|
|
'date': datetime.date.fromordinal(date if 0 < date else 1)}
|
|
|
|
# get list of needed ids
|
|
mis_map = [k for k, v in mapped.iteritems() if (v['status'] not in [
|
|
MapStatus.NO_AUTOMATIC_CHANGE, MapStatus.SOURCE])
|
|
and ((0 == v['id'] and MapStatus.NONE == v['status'])
|
|
or force or recheck or (update and 0 == v['id'] and k not in defunct_indexer))]
|
|
if mis_map:
|
|
url_tvmaze = TvmazeDict()
|
|
url_trakt = TraktDict()
|
|
if show_obj.indexer == INDEXER_TVDB or show_obj.indexer == INDEXER_TVRAGE:
|
|
url_tvmaze[show_obj.indexer] = show_obj.indexerid
|
|
url_trakt[show_obj.indexer] = show_obj.indexerid
|
|
elif show_obj.indexer == INDEXER_TVMAZE:
|
|
url_tvmaze[INDEXER_TVMAZE] = show_obj.indexer
|
|
if show_obj.imdbid and re.search(r'\d+$', show_obj.imdbid):
|
|
url_tvmaze[INDEXER_IMDB] = tryInt(re.search(r'(?:tt)?(\d+)', show_obj.imdbid).group(1))
|
|
url_trakt[INDEXER_IMDB] = tryInt(re.search(r'(?:tt)?(\d+)', show_obj.imdbid).group(1))
|
|
for m, v in mapped.iteritems():
|
|
if m != show_obj.indexer and m in [INDEXER_TVDB, INDEXER_TVRAGE, INDEXER_TVRAGE, INDEXER_IMDB] and \
|
|
0 < v.get('id', 0):
|
|
url_tvmaze[m] = v['id']
|
|
|
|
new_ids = NewIdDict()
|
|
|
|
if isinstance(show_obj.imdbid, basestring) and re.search(r'\d+$', show_obj.imdbid):
|
|
new_ids[INDEXER_IMDB] = tryInt(re.search(r'(?:tt)?(\d+)', show_obj.imdbid))
|
|
|
|
if 0 < len(url_tvmaze):
|
|
new_ids.update(get_tvmaze_ids(url_tvmaze))
|
|
|
|
for m, v in new_ids.iteritems():
|
|
if m != show_obj.indexer and m in [INDEXER_TVDB, INDEXER_TVRAGE, INDEXER_TVRAGE, INDEXER_IMDB] and 0 < v:
|
|
url_trakt[m] = v
|
|
|
|
if url_trakt:
|
|
new_ids.update(get_trakt_ids(url_trakt))
|
|
|
|
if INDEXER_TVMAZE not in new_ids:
|
|
new_url_tvmaze = TvmazeDict()
|
|
for k, v in new_ids.iteritems():
|
|
if k != show_obj.indexer and k in [INDEXER_TVDB, INDEXER_TVRAGE, INDEXER_TVRAGE, INDEXER_IMDB] \
|
|
and 0 < v and k not in url_tvmaze:
|
|
new_url_tvmaze[k] = v
|
|
|
|
if 0 < len(new_url_tvmaze):
|
|
new_ids.update(get_tvmaze_ids(new_url_tvmaze))
|
|
|
|
if INDEXER_TVMAZE not in new_ids:
|
|
f_date = get_premieredate(show_obj)
|
|
if f_date and f_date is not datetime.date.fromordinal(1):
|
|
tvids = {k: v for k, v in get_tvmaze_by_name(show_obj.name, f_date).iteritems() if k == INDEXER_TVMAZE
|
|
or k not in new_ids or new_ids.get(k) in (None, 0, '', MapStatus.NOT_FOUND)}
|
|
new_ids.update(tvids)
|
|
|
|
if INDEXER_TRAKT not in new_ids:
|
|
new_url_trakt = TraktDict()
|
|
for k, v in new_ids.iteritems():
|
|
if k != show_obj.indexer and k in [INDEXER_TVDB, INDEXER_TVRAGE, INDEXER_IMDB] and 0 < v \
|
|
and k not in url_trakt:
|
|
new_url_trakt[k] = v
|
|
|
|
if 0 < len(new_url_trakt):
|
|
new_ids.update(get_trakt_ids(new_url_trakt))
|
|
|
|
if INDEXER_IMDB not in new_ids:
|
|
new_ids.update(get_imdbid_by_name(show_obj.name, show_obj.startyear))
|
|
|
|
if INDEXER_TMDB in mis_map \
|
|
and (None is new_ids.get(INDEXER_TMDB) or MapStatus.NOT_FOUND == new_ids.get(INDEXER_TMDB)) \
|
|
and (0 < mapped.get(INDEXER_TVDB, {'id': 0}).get('id', 0) or 0 < new_ids.get(INDEXER_TVDB, 0)
|
|
or 0 < mapped.get(INDEXER_IMDB, {'id': 0}).get('id', 0) or 0 < new_ids.get(INDEXER_TMDB, 0)
|
|
or 0 < mapped.get(INDEXER_TVRAGE, {'id': 0}).get('id', 0) or 0 < new_ids.get(INDEXER_TVRAGE, 0)):
|
|
try:
|
|
tmdb = TMDB(sickbeard.TMDB_API_KEY)
|
|
for d in [INDEXER_TVDB, INDEXER_IMDB, INDEXER_TVRAGE]:
|
|
c = (new_ids.get(d), mapped.get(d, {'id': 0}).get('id'))[0 < mapped.get(d, {'id': 0}).get('id', 0)]
|
|
if 0 >= c:
|
|
continue
|
|
if INDEXER_IMDB == d:
|
|
c = 'tt%07d' % c
|
|
if None is not c and 0 < c:
|
|
tmdb_data = tmdb.Find(c).info({'external_source': tmdb_ids[d]})
|
|
if isinstance(tmdb_data, dict) \
|
|
and 'tv_results' in tmdb_data and 0 < len(tmdb_data['tv_results']) \
|
|
and 'id' in tmdb_data['tv_results'][0] and 0 < tryInt(tmdb_data['tv_results'][0]['id']):
|
|
new_ids[INDEXER_TMDB] = tryInt(tmdb_data['tv_results'][0]['id'])
|
|
break
|
|
except (StandardError, Exception):
|
|
pass
|
|
|
|
if INDEXER_TMDB not in new_ids:
|
|
try:
|
|
tmdb = TMDB(sickbeard.TMDB_API_KEY)
|
|
tmdb_data = tmdb.Search().tv(params={'query': clean_show_name(show_obj.name),
|
|
'first_air_date_year': show_obj.startyear})
|
|
for s in tmdb_data.get('results'):
|
|
if clean_show_name(s['name']) == clean_show_name(show_obj.name):
|
|
new_ids[INDEXER_TMDB] = tryInt(s['id'])
|
|
break
|
|
except (StandardError, Exception):
|
|
pass
|
|
|
|
for i in indexer_list:
|
|
if i != show_obj.indexer and i in mis_map and 0 != new_ids.get(i, 0):
|
|
if 0 > new_ids[i]:
|
|
mapped[i] = {'status': new_ids[i], 'id': 0}
|
|
elif force or not recheck or 0 >= mapped.get(i, {'id': 0}).get('id', 0):
|
|
mapped[i] = {'status': MapStatus.NONE, 'id': new_ids[i]}
|
|
|
|
if [k for k in mis_map if 0 != mapped.get(k, {'id': 0, 'status': 0})['id'] or
|
|
mapped.get(k, {'id': 0, 'status': 0})['status'] not in [MapStatus.NONE, MapStatus.SOURCE]]:
|
|
sql_l = []
|
|
today = datetime.date.today()
|
|
date = today.toordinal()
|
|
for indexer in indexer_list:
|
|
|
|
if show_obj.indexer == indexer or indexer not in mis_map:
|
|
continue
|
|
|
|
if 0 != mapped[indexer]['id'] or MapStatus.NONE != mapped[indexer]['status']:
|
|
mapped[indexer]['date'] = today
|
|
sql_l.append([
|
|
'INSERT OR REPLACE INTO indexer_mapping (' +
|
|
'indexer_id, indexer, mindexer_id, mindexer, date, status) VALUES (?,?,?,?,?,?)',
|
|
[show_obj.indexerid, show_obj.indexer, mapped[indexer]['id'],
|
|
indexer, date, mapped[indexer]['status']]])
|
|
else:
|
|
sql_l.append([
|
|
'DELETE' + ' FROM indexer_mapping WHERE indexer_id = ? AND indexer = ? AND mindexer = ?',
|
|
[show_obj.indexerid, show_obj.indexer, indexer]])
|
|
|
|
if 0 < len(sql_l):
|
|
logger.log('Adding indexer mapping to DB for show: %s' % show_obj.name, logger.DEBUG)
|
|
my_db = db.DBConnection()
|
|
my_db.mass_action(sql_l)
|
|
|
|
show_obj.ids = mapped
|
|
return mapped
|
|
|
|
|
|
def save_mapping(show_obj, save_map=None):
|
|
sql_l = []
|
|
today = datetime.date.today()
|
|
date = today.toordinal()
|
|
for indexer in indexer_list:
|
|
|
|
if show_obj.indexer == indexer or (isinstance(save_map, list) and indexer not in save_map):
|
|
continue
|
|
|
|
if 0 != show_obj.ids[indexer]['id'] or MapStatus.NONE != show_obj.ids[indexer]['status']:
|
|
show_obj.ids[indexer]['date'] = today
|
|
sql_l.append([
|
|
'INSERT OR REPLACE INTO indexer_mapping (' +
|
|
'indexer_id, indexer, mindexer_id, mindexer, date, status) VALUES (?,?,?,?,?,?)',
|
|
[show_obj.indexerid, show_obj.indexer, show_obj.ids[indexer]['id'],
|
|
indexer, date, show_obj.ids[indexer]['status']]])
|
|
else:
|
|
sql_l.append([
|
|
'DELETE' + ' FROM indexer_mapping WHERE indexer_id = ? AND indexer = ? AND mindexer = ?',
|
|
[show_obj.indexerid, show_obj.indexer, indexer]])
|
|
|
|
if 0 < len(sql_l):
|
|
logger.log('Saving indexer mapping to DB for show: %s' % show_obj.name, logger.DEBUG)
|
|
my_db = db.DBConnection()
|
|
my_db.mass_action(sql_l)
|
|
|
|
|
|
def del_mapping(indexer, indexerid):
|
|
my_db = db.DBConnection()
|
|
my_db.action('DELETE' + ' FROM indexer_mapping WHERE indexer_id = ? AND indexer = ?', [indexerid, indexer])
|
|
|
|
|
|
def should_recheck_update_ids(show):
|
|
try:
|
|
today = datetime.date.today()
|
|
ids_updated = min([v.get('date') for k, v in show.ids.iteritems() if k != show.indexer and
|
|
k not in defunct_indexer] or [datetime.date.fromtimestamp(1)])
|
|
if today - ids_updated >= datetime.timedelta(days=365):
|
|
return True
|
|
first_ep = show.getEpisode(season=1, episode=1)
|
|
if first_ep and first_ep.airdate and first_ep.airdate > datetime.date.fromtimestamp(1):
|
|
show_age = (today - first_ep.airdate).days
|
|
for d in [365, 270, 180, 135, 90, 60, 30, 16, 9] + range(4, -4, -1):
|
|
if d <= show_age:
|
|
return ids_updated < (first_ep.airdate + datetime.timedelta(days=d))
|
|
except (StandardError, Exception):
|
|
pass
|
|
return False
|
|
|
|
|
|
def load_mapped_ids(**kwargs):
|
|
logger.log('Start loading Indexer mappings...')
|
|
for s in sickbeard.showList:
|
|
with s.lock:
|
|
n_kargs = kwargs.copy()
|
|
if 'update' in kwargs and should_recheck_update_ids(s):
|
|
n_kargs['recheck'] = True
|
|
try:
|
|
s.ids = sickbeard.indexermapper.map_indexers_to_show(s, **n_kargs)
|
|
except (StandardError, Exception):
|
|
logger.log('Error loading mapped id\'s for show: %s' % s.name, logger.ERROR)
|
|
logger.log('Traceback: %s' % traceback.format_exc(), logger.ERROR)
|
|
logger.log('Indexer mappings loaded')
|
|
|
|
|
|
class MapStatus:
|
|
def __init__(self):
|
|
pass
|
|
|
|
SOURCE = 1
|
|
NONE = 0
|
|
NOT_FOUND = -1
|
|
MISMATCH = -2
|
|
NO_AUTOMATIC_CHANGE = -100
|
|
|
|
allstatus = [SOURCE, NONE, NOT_FOUND, MISMATCH, NO_AUTOMATIC_CHANGE]
|