2023-01-12 01:04:47 +00:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2012 Mr_Orange <mr_orange@hotmail.it>
|
|
|
|
#
|
|
|
|
# This file is part of subliminal.
|
|
|
|
#
|
|
|
|
# subliminal is free software; you can redistribute it and/or modify it under
|
|
|
|
# the terms of the GNU Lesser General Public License as published by
|
|
|
|
# the Free Software Foundation; either version 3 of the License, or
|
|
|
|
# (at your option) any later version.
|
|
|
|
#
|
|
|
|
# subliminal is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
# GNU Lesser General Public License for more details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU Lesser General Public License
|
|
|
|
# along with subliminal. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
from . import ServiceBase
|
|
|
|
from ..exceptions import DownloadFailedError, ServiceError
|
|
|
|
from ..cache import cachedmethod
|
|
|
|
from ..language import language_set, Language
|
|
|
|
from ..subtitles import get_subtitle_path, ResultSubtitle, EXTENSIONS
|
|
|
|
from ..utils import get_keywords
|
|
|
|
from ..videos import Episode
|
|
|
|
from bs4 import BeautifulSoup
|
|
|
|
import logging
|
|
|
|
import re
|
|
|
|
import os
|
|
|
|
import requests
|
|
|
|
import zipfile
|
|
|
|
import guessit
|
|
|
|
from six import StringIO
|
|
|
|
|
|
|
|
from sickgear.common import Quality
|
|
|
|
|
|
|
|
logger = logging.getLogger("subliminal")
|
|
|
|
|
|
|
|
|
|
|
|
class Itasa(ServiceBase):
|
|
|
|
server_url = 'http://www.italiansubs.net/'
|
|
|
|
site_url = 'http://www.italiansubs.net/'
|
|
|
|
api_based = False
|
|
|
|
languages = language_set(['it'])
|
|
|
|
videos = [Episode]
|
|
|
|
require_video = False
|
|
|
|
required_features = ['permissive']
|
|
|
|
quality_dict = {Quality.SDTV : '',
|
|
|
|
Quality.SDDVD : 'dvdrip',
|
|
|
|
Quality.RAWHDTV : '1080i',
|
2023-02-13 21:00:11 +00:00
|
|
|
Quality.HDTV : '720p',
|
2023-01-12 01:04:47 +00:00
|
|
|
Quality.FULLHDTV : ('1080p','720p'),
|
|
|
|
Quality.HDWEBDL : 'web-dl',
|
|
|
|
Quality.FULLHDWEBDL : 'web-dl',
|
|
|
|
Quality.HDBLURAY : ('bdrip', 'bluray'),
|
|
|
|
Quality.FULLHDBLURAY : ('bdrip', 'bluray'),
|
|
|
|
Quality.UNKNOWN : 'unknown' #Any subtitle will be downloaded
|
|
|
|
}
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
def init(self):
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
super(Itasa, self).init()
|
|
|
|
login_pattern = '<input type="hidden" name="return" value="([^\n\r\t ]+?)" /><input type="hidden" name="([^\n\r\t ]+?)" value="([^\n\r\t ]+?)" />'
|
|
|
|
|
|
|
|
response = requests.get(self.server_url + 'index.php')
|
|
|
|
if response.status_code != 200:
|
|
|
|
raise ServiceError('Initiate failed')
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
match = re.search(login_pattern, response.content, re.IGNORECASE | re.DOTALL)
|
|
|
|
if not match:
|
|
|
|
raise ServiceError('Can not find unique id parameter on page')
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
login_parameter = {'username': 'sickbeard',
|
|
|
|
'passwd': 'subliminal',
|
|
|
|
'remember': 'yes',
|
|
|
|
'Submit': 'Login',
|
|
|
|
'remember': 'yes',
|
|
|
|
'option': 'com_user',
|
|
|
|
'task': 'login',
|
|
|
|
'silent': 'true',
|
2023-02-13 21:00:11 +00:00
|
|
|
'return': match.group(1),
|
2023-01-12 01:04:47 +00:00
|
|
|
match.group(2): match.group(3)
|
|
|
|
}
|
|
|
|
|
|
|
|
self.session = requests.session()
|
|
|
|
r = self.session.post(self.server_url + 'index.php', data=login_parameter)
|
|
|
|
if not re.search('logouticon.png', r.content, re.IGNORECASE | re.DOTALL):
|
|
|
|
raise ServiceError('Itasa Login Failed')
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
@cachedmethod
|
|
|
|
def get_series_id(self, name):
|
|
|
|
"""Get the show page and cache every show found in it"""
|
|
|
|
r = self.session.get(self.server_url + 'index.php?option=com_remository&Itemid=9')
|
|
|
|
soup = BeautifulSoup(r.content, self.required_features)
|
|
|
|
all_series = soup.find('div', attrs = {'id' : 'remositorycontainerlist'})
|
|
|
|
for tv_series in all_series.find_all(href=re.compile('func=select')):
|
|
|
|
series_name = tv_series.text.lower().strip().replace(':','')
|
|
|
|
match = re.search('&id=([0-9]+)', tv_series['href'])
|
|
|
|
if match is None:
|
|
|
|
continue
|
|
|
|
series_id = int(match.group(1))
|
|
|
|
self.cache_for(self.get_series_id, args=(series_name,), result=series_id)
|
|
|
|
return self.cached_value(self.get_series_id, args=(name,))
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
def get_episode_id(self, series, series_id, season, episode, quality):
|
|
|
|
"""Get the id subtitle for episode with the given quality"""
|
|
|
|
|
|
|
|
season_link = None
|
|
|
|
quality_link = None
|
|
|
|
episode_id = None
|
|
|
|
|
|
|
|
r = self.session.get(self.server_url + 'index.php?option=com_remository&Itemid=6&func=select&id=' + str(series_id))
|
|
|
|
soup = BeautifulSoup(r.content, self.required_features)
|
|
|
|
all_seasons = soup.find('div', attrs = {'id' : 'remositorycontainerlist'})
|
|
|
|
for seasons in all_seasons.find_all(href=re.compile('func=select')):
|
|
|
|
if seasons.text.lower().strip() == 'stagione %s' % str(season):
|
|
|
|
season_link = seasons['href']
|
|
|
|
break
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
if not season_link:
|
|
|
|
logger.debug(u'Could not find season %s for series %s' % (series, str(season)))
|
|
|
|
return None
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
r = self.session.get(season_link)
|
|
|
|
soup = BeautifulSoup(r.content, self.required_features)
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
all_qualities = soup.find('div', attrs = {'id' : 'remositorycontainerlist'})
|
|
|
|
for qualities in all_qualities.find_all(href=re.compile('func=select')):
|
|
|
|
if qualities.text.lower().strip() in self.quality_dict[quality]:
|
|
|
|
quality_link = qualities['href']
|
|
|
|
r = self.session.get(qualities['href'])
|
|
|
|
soup = BeautifulSoup(r.content, self.required_features)
|
|
|
|
break
|
|
|
|
|
2023-02-13 21:00:11 +00:00
|
|
|
#If we want SDTV we are just on the right page so quality link will be None
|
2023-01-12 01:04:47 +00:00
|
|
|
if not quality == Quality.SDTV and not quality_link:
|
|
|
|
logger.debug(u'Could not find a subtitle with required quality for series %s season %s' % (series, str(season)))
|
|
|
|
return None
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
all_episodes = soup.find('div', attrs = {'id' : 'remositoryfilelisting'})
|
|
|
|
for episodes in all_episodes.find_all(href=re.compile('func=fileinfo')):
|
|
|
|
ep_string = "%(seasonnumber)dx%(episodenumber)02d" % {'seasonnumber': season, 'episodenumber': episode}
|
|
|
|
if re.search(ep_string, episodes.text, re.I) or re.search('completa$', episodes.text, re.I):
|
|
|
|
match = re.search('&id=([0-9]+)', episodes['href'])
|
|
|
|
if match:
|
|
|
|
episode_id = match.group(1)
|
|
|
|
return episode_id
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
return episode_id
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
def list_checked(self, video, languages):
|
|
|
|
return self.query(video.path or video.release, languages, get_keywords(video.guess), video.series, video.season, video.episode)
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
def query(self, filepath, languages, keywords, series, season, episode):
|
|
|
|
|
|
|
|
logger.debug(u'Getting subtitles for %s season %d episode %d with languages %r' % (series, season, episode, languages))
|
|
|
|
self.init_cache()
|
|
|
|
try:
|
|
|
|
series = series.lower().replace('(','').replace(')','')
|
|
|
|
series_id = self.get_series_id(series)
|
|
|
|
except KeyError:
|
|
|
|
logger.debug(u'Could not find series id for %s' % series)
|
|
|
|
return []
|
2023-02-13 21:00:11 +00:00
|
|
|
|
|
|
|
episode_id = self.get_episode_id(series, series_id, season, episode, Quality.name_quality(filepath))
|
2023-01-12 01:04:47 +00:00
|
|
|
if not episode_id:
|
|
|
|
logger.debug(u'Could not find subtitle for series %s' % series)
|
|
|
|
return []
|
|
|
|
|
|
|
|
r = self.session.get(self.server_url + 'index.php?option=com_remository&Itemid=6&func=fileinfo&id=' + episode_id)
|
|
|
|
soup = BeautifulSoup(r.content)
|
|
|
|
|
|
|
|
sub_link = soup.find('div', attrs = {'id' : 'remositoryfileinfo'}).find(href=re.compile('func=download'))['href']
|
|
|
|
sub_language = self.get_language('it')
|
|
|
|
path = get_subtitle_path(filepath, sub_language, self.config.multi)
|
|
|
|
subtitle = ResultSubtitle(path, sub_language, self.__class__.__name__.lower(), sub_link)
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
return [subtitle]
|
|
|
|
|
|
|
|
def download(self, subtitle):
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
logger.info(u'Downloading %s in %s' % (subtitle.link, subtitle.path))
|
|
|
|
try:
|
|
|
|
r = self.session.get(subtitle.link, headers={'Referer': self.server_url, 'User-Agent': self.user_agent})
|
|
|
|
zipcontent = StringIO.StringIO(r.content)
|
|
|
|
zipsub = zipfile.ZipFile(zipcontent)
|
|
|
|
|
|
|
|
# if not zipsub.is_zipfile(zipcontent):
|
|
|
|
# raise DownloadFailedError('Downloaded file is not a zip file')
|
|
|
|
|
|
|
|
subfile = ''
|
|
|
|
if len(zipsub.namelist()) == 1:
|
|
|
|
subfile = zipsub.namelist()[0]
|
|
|
|
else:
|
|
|
|
#Season Zip Retrive Season and episode Numbers from path
|
|
|
|
guess = guessit.guess_file_info(subtitle.path, 'episode')
|
|
|
|
ep_string = "s%(seasonnumber)02de%(episodenumber)02d" % {'seasonnumber': guess['season'], 'episodenumber': guess['episodeNumber']}
|
|
|
|
for file in zipsub.namelist():
|
|
|
|
if re.search(ep_string, file, re.I):
|
|
|
|
subfile = file
|
|
|
|
break
|
|
|
|
if os.path.splitext(subfile)[1] in EXTENSIONS:
|
|
|
|
with open(subtitle.path, 'wb') as f:
|
|
|
|
f.write(zipsub.open(subfile).read())
|
|
|
|
else:
|
|
|
|
zipsub.close()
|
|
|
|
raise DownloadFailedError('No subtitles found in zip file')
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
zipsub.close()
|
|
|
|
except Exception as e:
|
|
|
|
if os.path.exists(subtitle.path):
|
|
|
|
os.remove(subtitle.path)
|
|
|
|
raise DownloadFailedError(str(e))
|
2023-02-13 21:00:11 +00:00
|
|
|
|
2023-01-12 01:04:47 +00:00
|
|
|
logger.debug(u'Download finished')
|
2023-02-13 21:00:11 +00:00
|
|
|
|
|
|
|
Service = Itasa
|