2023-01-12 01:04:47 +00:00
|
|
|
#
|
|
|
|
# This file is part of SickGear.
|
|
|
|
#
|
|
|
|
# SickGear is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU General Public License as published by
|
|
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
|
|
# (at your option) any later version.
|
|
|
|
#
|
|
|
|
# SickGear is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
# GNU General Public License for more details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU General Public License
|
|
|
|
# along with SickGear. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
import datetime
|
|
|
|
import re
|
|
|
|
import traceback
|
|
|
|
|
|
|
|
from . import classes, db, logger
|
|
|
|
from .helpers import try_int
|
|
|
|
from .indexers.indexer_config import TVINFO_IMDB, TVINFO_TMDB, TVINFO_TRAKT, TVINFO_TVDB, TVINFO_TVMAZE
|
|
|
|
|
|
|
|
import sickgear
|
|
|
|
|
|
|
|
from lib.dateutil.parser import parse
|
|
|
|
|
2023-02-11 18:02:58 +00:00
|
|
|
from six import iteritems, moves, string_types
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
# noinspection PyUnreachableCode
|
|
|
|
if False:
|
|
|
|
# noinspection PyUnresolvedReferences
|
|
|
|
from typing import Any, AnyStr, Dict, List, Optional, Tuple, Union
|
|
|
|
from six import integer_types
|
|
|
|
from sickgear.tv import TVShow
|
|
|
|
|
|
|
|
tv_maze_retry_wait = 10
|
|
|
|
defunct_indexer = []
|
|
|
|
indexer_list = []
|
|
|
|
|
|
|
|
|
|
|
|
class NewIdDict(dict):
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
|
|
tv_src = kwargs.pop('tv_src')
|
|
|
|
super(NewIdDict, self).__init__(*args, **kwargs)
|
|
|
|
self.verified = {s: (False, True)[s == tv_src] for s in indexer_list}
|
|
|
|
|
|
|
|
def set_value(self, value, old_value=None, tv_src=None, key=None):
|
|
|
|
# type: (Any, Any, int, int) -> Any
|
|
|
|
if (None is tv_src or tv_src != key) and old_value is MapStatus.MISMATCH or (
|
|
|
|
0 < value and old_value not in [None, value] and 0 < old_value):
|
|
|
|
return MapStatus.MISMATCH
|
|
|
|
if value and tv_src and tv_src == key:
|
|
|
|
self.verified[tv_src] = True
|
|
|
|
return value
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def get_value(value):
|
|
|
|
if value in [None, 0]:
|
|
|
|
return MapStatus.NOT_FOUND
|
|
|
|
return value
|
|
|
|
|
|
|
|
def __getitem__(self, key):
|
|
|
|
return self.get_value(super(NewIdDict, self).get(key))
|
|
|
|
|
|
|
|
def get(self, key, default=None):
|
|
|
|
return self.get_value(super(NewIdDict, self).get(key, default))
|
|
|
|
|
|
|
|
def __setitem__(self, key, value):
|
|
|
|
super(NewIdDict, self).__setitem__(key, self.set_value(value, self.get(key)))
|
|
|
|
|
|
|
|
def update(self, other=None, tv_src=None, **kwargs):
|
|
|
|
# type: (Dict[int, Any], int, Any) -> None
|
|
|
|
"""
|
|
|
|
updates dict with new ids
|
|
|
|
set MapStatus.MISMATCH if values mismatch, except if it's tv_src (this will be treated as verified source id)
|
|
|
|
|
|
|
|
:param other: new data dict
|
|
|
|
:param tv_src: verified tv src id
|
|
|
|
:param kwargs:
|
|
|
|
"""
|
|
|
|
if isinstance(other, dict):
|
|
|
|
other = {o: self.set_value(v, self.get(o), tv_src, o) for o, v in iteritems(other)}
|
|
|
|
super(NewIdDict, self).update(other, **kwargs)
|
|
|
|
|
|
|
|
|
|
|
|
def get_missing_ids(show_ids, show_obj, tv_src):
|
|
|
|
# type: (Dict[int, integer_types], TVShow, int) -> Dict[int, integer_types]
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param show_ids:
|
|
|
|
:param show_obj:
|
|
|
|
:param tv_src:
|
|
|
|
:return:
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
tvinfo_config = sickgear.TVInfoAPI(tv_src).api_params.copy()
|
|
|
|
tvinfo_config['cache_search'] = True
|
|
|
|
tvinfo_config['custom_ui'] = classes.AllShowInfosNoFilterListUI
|
|
|
|
t = sickgear.TVInfoAPI(tv_src).setup(**tvinfo_config)
|
|
|
|
show_name, f_date = None, None
|
|
|
|
if any(1 for k, v in iteritems(show_ids) if v and k in t.supported_id_searches):
|
|
|
|
try:
|
|
|
|
found_shows = t.search_show(ids=show_ids)
|
|
|
|
res_count = len(found_shows or [])
|
|
|
|
if 1 < res_count:
|
|
|
|
show_name, f_date = get_show_name_date(show_obj)
|
|
|
|
for show in found_shows or []:
|
|
|
|
if 1 == res_count or confirm_show(f_date, show['firstaired'], show_name,
|
|
|
|
clean_show_name(show['seriesname'])):
|
|
|
|
return combine_new_ids(show_ids, show['ids'], tv_src)
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
found_shows = t.search_show(name=clean_show_name(show_obj.name))
|
|
|
|
if not show_name:
|
|
|
|
show_name, f_date = get_show_name_date(show_obj)
|
|
|
|
for show in found_shows or []:
|
|
|
|
if confirm_show(f_date, show['firstaired'], show_name, clean_show_name(show['seriesname'])):
|
|
|
|
if any(v for k, v in iteritems(show['ids']) if tv_src != k and v):
|
|
|
|
f_show = [show]
|
|
|
|
else:
|
|
|
|
f_show = t.search_show(ids={tv_src: show['id']})
|
|
|
|
if f_show and 1 == len(f_show):
|
|
|
|
return combine_new_ids(show_ids, f_show[0]['ids'], tv_src)
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
return {}
|
|
|
|
|
|
|
|
|
|
|
|
def confirm_show(premiere_date, shows_premiere, expected_name, show_name):
|
|
|
|
# type: (Optional[datetime.date], Optional[Union[AnyStr, datetime.date]], AnyStr, AnyStr) -> bool
|
|
|
|
"""
|
|
|
|
confirm show possible confirmations:
|
2023-02-13 21:00:11 +00:00
|
|
|
1. premiere dates are less than 2 days apart
|
2023-01-12 01:04:47 +00:00
|
|
|
2. show name is the same and premiere year is 1 year or less apart
|
|
|
|
|
|
|
|
:param premiere_date: expected show premiere date
|
|
|
|
:param shows_premiere: compare date
|
|
|
|
:param expected_name:
|
|
|
|
:param show_name:
|
|
|
|
"""
|
|
|
|
if any(t is None for t in (premiere_date, shows_premiere)):
|
|
|
|
return False
|
|
|
|
if isinstance(shows_premiere, string_types):
|
|
|
|
try:
|
|
|
|
shows_premiere = parse(shows_premiere).date()
|
|
|
|
except (BaseException, Exception):
|
|
|
|
return False
|
|
|
|
start_year = (shows_premiere and shows_premiere.year) or 0
|
|
|
|
return abs(premiere_date - shows_premiere) < datetime.timedelta(days=2) or (
|
|
|
|
expected_name == show_name and abs(premiere_date.year - start_year) <= 1)
|
|
|
|
|
|
|
|
|
|
|
|
def get_premieredate(show_obj):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param show_obj: show object
|
|
|
|
:type show_obj: sickgear.tv.TVShow
|
|
|
|
:return:
|
|
|
|
:rtype: datetime.date or None
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
ep_obj = show_obj.first_aired_regular_episode
|
|
|
|
if ep_obj and ep_obj.airdate:
|
|
|
|
return ep_obj.airdate
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
return None
|
|
|
|
|
|
|
|
|
|
|
|
def clean_show_name(showname):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param showname: show name
|
|
|
|
:type showname: AnyStr
|
|
|
|
:return:
|
|
|
|
:rtype: AnyStr
|
|
|
|
"""
|
2023-02-11 18:02:58 +00:00
|
|
|
return re.sub(r'[(\s]*(?:19|20)\d\d[)\s]*$', '', showname)
|
2023-01-12 01:04:47 +00:00
|
|
|
|
|
|
|
|
|
|
|
def get_show_name_date(show_obj):
|
|
|
|
# type: (TVShow) -> Tuple[Optional[AnyStr], Optional[datetime.date]]
|
|
|
|
return clean_show_name(show_obj.name), get_premieredate(show_obj)
|
|
|
|
|
|
|
|
|
|
|
|
def combine_mapped_new_dict(mapped, new_ids):
|
|
|
|
# type: (Dict[int, Dict], Dict[int, integer_types]) -> Dict[int, integer_types]
|
|
|
|
return {n: m for d in ({k: v['id'] for k, v in iteritems(mapped) if v['id']}, new_ids) for n, m in iteritems(d)}
|
|
|
|
|
|
|
|
|
|
|
|
def combine_new_ids(cur_ids, new_ids, src_id):
|
|
|
|
# type: (Dict[int, integer_types], Dict[int, integer_types], int) -> Dict[int, integer_types]
|
|
|
|
"""
|
|
|
|
combine cur_ids with new_ids, priority has cur_ids with exception of src_id key
|
|
|
|
|
|
|
|
:param cur_ids:
|
|
|
|
:param new_ids:
|
|
|
|
:param src_id:
|
|
|
|
"""
|
|
|
|
return {k: v for d in (cur_ids, new_ids) for k, v in iteritems(d)
|
|
|
|
if v and (k == src_id or not cur_ids.get(k) or v == cur_ids.get(k, ''))}
|
|
|
|
|
|
|
|
|
|
|
|
def map_indexers_to_show(show_obj, update=False, force=False, recheck=False, im_sql_result=None):
|
|
|
|
# type: (sickgear.tv.TVShow, Optional[bool], Optional[bool], Optional[bool], Optional[list]) -> dict
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param show_obj: TVShow Object
|
|
|
|
:param update: add missing + previously not found ids
|
|
|
|
:param force: search for and replace all mapped/missing ids (excluding NO_AUTOMATIC_CHANGE flagged)
|
|
|
|
:param recheck: load all ids, don't remove existing
|
|
|
|
:param im_sql_result:
|
|
|
|
:return: mapped ids
|
|
|
|
"""
|
|
|
|
mapped = {}
|
|
|
|
|
|
|
|
# init mapped tvids object
|
|
|
|
for tvid in indexer_list:
|
|
|
|
mapped[tvid] = {'id': (0, show_obj.prodid)[int(tvid) == int(show_obj.tvid)],
|
|
|
|
'status': (MapStatus.NONE, MapStatus.SOURCE)[int(tvid) == int(show_obj.tvid)],
|
|
|
|
'date': datetime.date.fromordinal(1)}
|
|
|
|
|
|
|
|
sql_result = []
|
|
|
|
for cur_row in im_sql_result or []:
|
|
|
|
if show_obj.prodid == cur_row['indexer_id'] and show_obj.tvid == cur_row['indexer']:
|
|
|
|
sql_result.append(cur_row)
|
|
|
|
|
|
|
|
if not sql_result:
|
|
|
|
my_db = db.DBConnection()
|
|
|
|
sql_result = my_db.select(
|
|
|
|
'SELECT * FROM indexer_mapping WHERE indexer = ? AND indexer_id = ?', [show_obj.tvid, show_obj.prodid])
|
|
|
|
|
|
|
|
# for each mapped entry
|
|
|
|
for cur_row in sql_result or []:
|
|
|
|
date = try_int(cur_row['date'])
|
|
|
|
mapped[int(cur_row['mindexer'])] = {'status': int(cur_row['status']),
|
|
|
|
'id': int(cur_row['mindexer_id']),
|
|
|
|
'date': datetime.date.fromordinal(date if 0 < date else 1)}
|
|
|
|
|
|
|
|
# get list of needed ids
|
|
|
|
mis_map = [k for k, v in iteritems(mapped) if (v['status'] not in [
|
|
|
|
MapStatus.NO_AUTOMATIC_CHANGE, MapStatus.SOURCE])
|
|
|
|
and ((0 == v['id'] and MapStatus.NONE == v['status'])
|
|
|
|
or force or recheck or (update and 0 == v['id'] and k not in defunct_indexer))]
|
|
|
|
if mis_map:
|
|
|
|
src_tv_id = show_obj._tvid
|
|
|
|
new_ids = NewIdDict(tv_src=src_tv_id) # type: NewIdDict
|
|
|
|
if show_obj.imdbid and re.search(r'\d+$', show_obj.imdbid):
|
|
|
|
new_ids[TVINFO_IMDB] = try_int(re.search(r'(?:tt)?(\d+)', show_obj.imdbid).group(1))
|
|
|
|
all_ids_srcs = [src_tv_id] + [s for s in (TVINFO_TRAKT, TVINFO_TMDB, TVINFO_TVMAZE, TVINFO_TVDB, TVINFO_IMDB)
|
|
|
|
if s != src_tv_id]
|
|
|
|
searched, confirmed = {}, False
|
2023-02-13 21:00:11 +00:00
|
|
|
for _ in moves.range(len(all_ids_srcs)):
|
2023-01-12 01:04:47 +00:00
|
|
|
search_done = False
|
|
|
|
for i in all_ids_srcs:
|
|
|
|
if new_ids.verified.get(i):
|
|
|
|
continue
|
|
|
|
search_ids = {k: v for k, v in iteritems(combine_mapped_new_dict(mapped, new_ids))
|
|
|
|
if k not in searched.setdefault(i, {})}
|
|
|
|
if search_ids:
|
|
|
|
search_done = True
|
|
|
|
searched[i].update(search_ids)
|
|
|
|
new_ids.update(get_missing_ids(search_ids, show_obj, tv_src=i), tv_src=i)
|
|
|
|
if new_ids.get(i) and 0 < new_ids.get(i):
|
|
|
|
searched[i].update({i: new_ids[i]})
|
|
|
|
confirmed = all(v for k, v in iteritems(new_ids.verified) if k not in defunct_indexer)
|
|
|
|
if confirmed:
|
|
|
|
break
|
|
|
|
if confirmed or not search_done:
|
|
|
|
break
|
|
|
|
|
|
|
|
for i in indexer_list:
|
|
|
|
if i != show_obj.tvid and ((i in mis_map and 0 != new_ids.get(i, 0)) or
|
|
|
|
(new_ids.verified.get(i) and 0 < new_ids.get(i, 0))):
|
|
|
|
if i not in new_ids:
|
|
|
|
mapped[i] = {'status': MapStatus.NOT_FOUND, 'id': 0}
|
|
|
|
continue
|
|
|
|
if new_ids.verified.get(i) and 0 < new_ids[i] and mapped.get(i, {'id': 0})['id'] != new_ids[i]:
|
|
|
|
if i not in mis_map:
|
|
|
|
mis_map.append(i)
|
|
|
|
mapped[i] = {'status': MapStatus.NONE, 'id': new_ids[i]}
|
|
|
|
continue
|
|
|
|
if 0 > new_ids[i]:
|
|
|
|
mapped[i] = {'status': new_ids[i], 'id': 0}
|
|
|
|
elif force or not recheck or 0 >= mapped.get(i, {'id': 0}).get('id', 0):
|
|
|
|
mapped[i] = {'status': MapStatus.NONE, 'id': new_ids[i]}
|
|
|
|
|
|
|
|
if [k for k in mis_map if 0 != mapped.get(k, {'id': 0, 'status': 0})['id'] or
|
|
|
|
mapped.get(k, {'id': 0, 'status': 0})['status'] not in [MapStatus.NONE, MapStatus.SOURCE]]:
|
|
|
|
sql_l = []
|
|
|
|
today = datetime.date.today()
|
|
|
|
date = today.toordinal()
|
|
|
|
for tvid in indexer_list:
|
|
|
|
|
|
|
|
if show_obj.tvid == tvid or tvid not in mis_map:
|
|
|
|
continue
|
|
|
|
|
|
|
|
if 0 != mapped[tvid]['id'] or MapStatus.NONE != mapped[tvid]['status']:
|
|
|
|
mapped[tvid]['date'] = today
|
|
|
|
sql_l.append([
|
|
|
|
'REPLACE INTO indexer_mapping (indexer_id, indexer, mindexer_id, mindexer, date, status)'
|
|
|
|
' VALUES (?,?,?,?,?,?)',
|
|
|
|
[show_obj.prodid, show_obj.tvid, mapped[tvid]['id'], tvid, date, mapped[tvid]['status']]])
|
|
|
|
else:
|
|
|
|
sql_l.append([
|
|
|
|
'DELETE FROM indexer_mapping'
|
|
|
|
' WHERE indexer = ? AND indexer_id = ? AND mindexer = ?',
|
|
|
|
[show_obj.tvid, show_obj.prodid, tvid]])
|
|
|
|
|
|
|
|
if 0 < len(sql_l):
|
|
|
|
logger.debug('Adding TV info mapping to DB for show: %s' % show_obj.unique_name)
|
|
|
|
my_db = db.DBConnection()
|
|
|
|
my_db.mass_action(sql_l)
|
|
|
|
|
|
|
|
show_obj.ids = mapped
|
|
|
|
return mapped
|
|
|
|
|
|
|
|
|
|
|
|
def save_mapping(show_obj, save_map=None):
|
|
|
|
# type: (sickgear.tv.TVShow, Optional[List[int]]) -> None
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param show_obj: show object
|
|
|
|
:param save_map: list of tvid ints
|
|
|
|
"""
|
|
|
|
sql_l = []
|
|
|
|
today = datetime.date.today()
|
|
|
|
date = today.toordinal()
|
|
|
|
for tvid in indexer_list:
|
|
|
|
|
|
|
|
if show_obj.tvid == tvid or (isinstance(save_map, list) and tvid not in save_map):
|
|
|
|
continue
|
|
|
|
|
|
|
|
if 0 != show_obj.ids[tvid]['id'] or MapStatus.NONE != show_obj.ids[tvid]['status']:
|
|
|
|
show_obj.ids[tvid]['date'] = today
|
|
|
|
sql_l.append([
|
|
|
|
'REPLACE INTO indexer_mapping'
|
|
|
|
' (indexer_id, indexer, mindexer_id, mindexer, date, status) VALUES (?,?,?,?,?,?)',
|
|
|
|
[show_obj.prodid, show_obj.tvid, show_obj.ids[tvid]['id'],
|
|
|
|
tvid, date, show_obj.ids[tvid]['status']]])
|
|
|
|
else:
|
|
|
|
sql_l.append([
|
|
|
|
'DELETE FROM indexer_mapping WHERE indexer = ? AND indexer_id = ? AND mindexer = ?',
|
|
|
|
[show_obj.tvid, show_obj.prodid, tvid]])
|
|
|
|
|
|
|
|
if 0 < len(sql_l):
|
|
|
|
logger.debug('Saving TV info mapping to DB for show: %s' % show_obj.unique_name)
|
|
|
|
my_db = db.DBConnection()
|
|
|
|
my_db.mass_action(sql_l)
|
|
|
|
|
|
|
|
|
|
|
|
def del_mapping(tvid, prodid):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param tvid: tvid
|
|
|
|
:type tvid: int
|
|
|
|
:param prodid: prodid
|
|
|
|
:type prodid: int or long
|
|
|
|
"""
|
|
|
|
my_db = db.DBConnection()
|
|
|
|
my_db.action('DELETE FROM indexer_mapping WHERE indexer = ? AND indexer_id = ?', [tvid, prodid])
|
|
|
|
|
|
|
|
|
|
|
|
def should_recheck_update_ids(show_obj):
|
|
|
|
"""
|
|
|
|
|
|
|
|
:param show_obj: show object
|
|
|
|
:type show_obj: sickgear.tv.TVShow
|
|
|
|
:return:
|
|
|
|
:rtype: bool
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
today = datetime.date.today()
|
|
|
|
ids_updated = min([v.get('date') for k, v in iteritems(show_obj.ids) if k != show_obj.tvid and
|
|
|
|
k not in defunct_indexer] or [datetime.date.fromtimestamp(1)])
|
|
|
|
if today - ids_updated >= datetime.timedelta(days=365):
|
|
|
|
return True
|
|
|
|
ep_obj = show_obj.first_aired_regular_episode
|
|
|
|
if ep_obj and ep_obj.airdate and ep_obj.airdate > datetime.date.fromtimestamp(1):
|
|
|
|
show_age = (today - ep_obj.airdate).days
|
|
|
|
# noinspection PyTypeChecker
|
|
|
|
for d in [365, 270, 180, 135, 90, 60, 30, 16, 9] + range(4, -4, -1):
|
|
|
|
if d <= show_age:
|
|
|
|
return ids_updated < (ep_obj.airdate + datetime.timedelta(days=d))
|
|
|
|
except (BaseException, Exception):
|
|
|
|
pass
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
|
|
|
def load_mapped_ids(**kwargs):
|
|
|
|
logger.log('Start loading TV info mappings...')
|
|
|
|
if 'load_all' in kwargs:
|
|
|
|
del kwargs['load_all']
|
|
|
|
my_db = db.DBConnection()
|
|
|
|
sql_result = my_db.select('SELECT * FROM indexer_mapping ORDER BY indexer, indexer_id')
|
|
|
|
else:
|
|
|
|
sql_result = None
|
|
|
|
for cur_show_obj in sickgear.showList:
|
|
|
|
with cur_show_obj.lock:
|
|
|
|
n_kargs = kwargs.copy()
|
|
|
|
if 'update' in kwargs and should_recheck_update_ids(cur_show_obj):
|
|
|
|
n_kargs['recheck'] = True
|
|
|
|
if sql_result:
|
|
|
|
n_kargs['im_sql_result'] = sql_result
|
|
|
|
try:
|
|
|
|
cur_show_obj.ids = sickgear.indexermapper.map_indexers_to_show(cur_show_obj, **n_kargs)
|
|
|
|
except (BaseException, Exception):
|
|
|
|
logger.debug('Error loading mapped id\'s for show: %s' % cur_show_obj.unique_name)
|
2023-03-08 13:44:20 +00:00
|
|
|
logger.error('Traceback: %s' % traceback.format_exc())
|
2023-01-12 01:04:47 +00:00
|
|
|
logger.log('TV info mappings loaded')
|
|
|
|
|
|
|
|
|
|
|
|
class MapStatus(object):
|
|
|
|
def __init__(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
SOURCE = 1
|
|
|
|
NONE = 0
|
|
|
|
NOT_FOUND = -1
|
|
|
|
MISMATCH = -2
|
|
|
|
NO_AUTOMATIC_CHANGE = -100
|
|
|
|
|
|
|
|
allstatus = [SOURCE, NONE, NOT_FOUND, MISMATCH, NO_AUTOMATIC_CHANGE]
|