2014-03-10 05:18:05 +00:00
# Author: Mr_Orange <mr_orange@hotmail.it>
# URL: http://code.google.com/p/sickbeard/
#
# This file is part of Sick Beard.
#
# Sick Beard is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Sick Beard is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Sick Beard. If not, see <http://www.gnu.org/licenses/>.
from __future__ import with_statement
import sys
import os
import traceback
import urllib , urllib2
import re
import datetime
import urlparse
import sickbeard
import generic
from sickbeard . common import Quality , Overview
from sickbeard . name_parser . parser import NameParser , InvalidNameException
from sickbeard import logger
from sickbeard import tvcache
from sickbeard import helpers
from sickbeard import db
from sickbeard import classes
from sickbeard . show_name_helpers import allPossibleShowNames , sanitizeSceneName
from sickbeard . exceptions import ex
from sickbeard import encodingKludge as ek
from sickbeard import clients
from sickbeard import tv
from lib import requests
2014-03-20 08:15:22 +00:00
from lib . bs4 import BeautifulSoup
2014-03-10 05:18:05 +00:00
from lib . unidecode import unidecode
class KATProvider ( generic . TorrentProvider ) :
def __init__ ( self ) :
generic . TorrentProvider . __init__ ( self , " KickAssTorrents " )
self . supportsBacklog = True
self . cache = KATCache ( self )
self . url = ' http://kickass.to/ '
self . searchurl = self . url + ' usearch/ %s /?field=seeders&sorder=desc ' #order by seed
def isEnabled ( self ) :
return sickbeard . KAT
def imageName ( self ) :
return ' kat.png '
def getQuality ( self , item ) :
quality = Quality . sceneQuality ( item [ 0 ] )
return quality
def _reverseQuality ( self , quality ) :
quality_string = ' '
if quality == Quality . SDTV :
quality_string = ' HDTV x264 '
if quality == Quality . SDDVD :
quality_string = ' DVDRIP '
elif quality == Quality . HDTV :
quality_string = ' 720p HDTV x264 '
elif quality == Quality . FULLHDTV :
quality_string = ' 1080p HDTV x264 '
elif quality == Quality . RAWHDTV :
quality_string = ' 1080i HDTV mpeg2 '
elif quality == Quality . HDWEBDL :
quality_string = ' 720p WEB-DL h264 '
elif quality == Quality . FULLHDWEBDL :
quality_string = ' 1080p WEB-DL h264 '
elif quality == Quality . HDBLURAY :
quality_string = ' 720p Bluray x264 '
elif quality == Quality . FULLHDBLURAY :
quality_string = ' 1080p Bluray x264 '
return quality_string
def _find_season_quality ( self , title , torrent_link , ep_number ) :
""" Return the modified title of a Season Torrent with the quality found inspecting torrent file list """
mediaExtensions = [ ' avi ' , ' mkv ' , ' wmv ' , ' divx ' ,
' vob ' , ' dvr-ms ' , ' wtv ' , ' ts '
' ogv ' , ' rar ' , ' zip ' , ' mp4 ' ]
quality = Quality . UNKNOWN
fileName = None
data = self . getURL ( torrent_link )
if not data :
return None
try :
soup = BeautifulSoup ( data , features = [ " html5lib " , " permissive " ] )
file_table = soup . find ( ' table ' , attrs = { ' class ' : ' torrentFileList ' } )
if not file_table :
return None
files = [ x . text for x in file_table . find_all ( ' td ' , attrs = { ' class ' : ' torFileName ' } ) ]
videoFiles = filter ( lambda x : x . rpartition ( " . " ) [ 2 ] . lower ( ) in mediaExtensions , files )
#Filtering SingleEpisode/MultiSeason Torrent
if len ( videoFiles ) < ep_number or len ( videoFiles ) > float ( ep_number * 1.1 ) :
logger . log ( u " Result " + title + " have " + str ( ep_number ) + " episode and episodes retrived in torrent are " + str ( len ( videoFiles ) ) , logger . DEBUG )
logger . log ( u " Result " + title + " Seem to be a Single Episode or MultiSeason torrent, skipping result... " , logger . DEBUG )
return None
if Quality . sceneQuality ( title ) != Quality . UNKNOWN :
return title
for fileName in videoFiles :
quality = Quality . sceneQuality ( os . path . basename ( fileName ) )
if quality != Quality . UNKNOWN : break
if fileName != None and quality == Quality . UNKNOWN :
quality = Quality . assumeQuality ( os . path . basename ( fileName ) )
if quality == Quality . UNKNOWN :
logger . log ( u " Unable to obtain a Season Quality for " + title , logger . DEBUG )
return None
try :
myParser = NameParser ( )
parse_result = myParser . parse ( fileName , True )
except InvalidNameException :
return None
logger . log ( u " Season quality for " + title + " is " + Quality . qualityStrings [ quality ] , logger . DEBUG )
if parse_result . series_name and parse_result . season_number :
title = parse_result . series_name + ' S %02d ' % int ( parse_result . season_number ) + ' ' + self . _reverseQuality ( quality )
return title
except Exception , e :
logger . log ( u " Failed parsing " + self . name + " Traceback: " + traceback . format_exc ( ) , logger . ERROR )
def _get_season_search_strings ( self , show , season , wantedEp , searchSeason = False ) :
search_string = { ' Episode ' : [ ] }
if not show :
return [ ]
self . show = show
if searchSeason :
search_string = { ' Season ' : [ ] , ' Episode ' : [ ] }
for show_name in set ( allPossibleShowNames ( show ) ) :
ep_string = show_name + ' S %02d ' % int ( season ) + ' -S %02d ' % int ( season ) + ' E ' + ' category:tv ' #1) ShowName SXX -SXXE
search_string [ ' Season ' ] . append ( ep_string )
ep_string = show_name + ' Season ' + str ( season ) + ' -Ep* ' + ' category:tv ' #2) ShowName Season X
search_string [ ' Season ' ] . append ( ep_string )
for ep_obj in wantedEp :
search_string [ ' Episode ' ] + = self . _get_episode_search_strings ( ep_obj ) [ 0 ] [ ' Episode ' ]
if not search_string [ ' Episode ' ] :
return [ ]
return [ search_string ]
def _get_episode_search_strings ( self , ep_obj , add_string = ' ' ) :
search_string = { ' Episode ' : [ ] }
if not ep_obj :
return [ ]
self . show = ep_obj . show
if ep_obj . show . air_by_date :
for show_name in set ( allPossibleShowNames ( ep_obj . show ) ) :
ep_string = sanitizeSceneName ( show_name ) + ' ' + \
str ( ep_obj . airdate ) + ' | ' + \
helpers . custom_strftime ( ' % Y % b {S} ' , ep_obj . airdate )
search_string [ ' Episode ' ] . append ( ep_string )
else :
for show_name in set ( allPossibleShowNames ( ep_obj . show ) ) :
2014-03-20 08:15:22 +00:00
ep_string = sanitizeSceneName ( show_name ) + ' ' + ' season: ' + str ( ep_obj . season ) + ' episode: ' + str ( ep_obj . episode )
2014-03-10 05:18:05 +00:00
search_string [ ' Episode ' ] . append ( re . sub ( ' \ s+ ' , ' ' , ep_string ) )
return [ search_string ]
def _doSearch ( self , search_params , show = None ) :
results = [ ]
items = { ' Season ' : [ ] , ' Episode ' : [ ] , ' RSS ' : [ ] }
for mode in search_params . keys ( ) :
for search_string in search_params [ mode ] :
if mode != ' RSS ' :
searchURL = self . searchurl % ( urllib . quote ( unidecode ( search_string ) ) )
logger . log ( u " Search string: " + searchURL , logger . DEBUG )
else :
searchURL = self . url + ' tv/?field=time_add&sorder=desc '
logger . log ( u " KAT cache update URL: " + searchURL , logger . DEBUG )
html = self . getURL ( searchURL )
if not html :
continue
try :
soup = BeautifulSoup ( html , features = [ " html5lib " , " permissive " ] )
torrent_table = soup . find ( ' table ' , attrs = { ' class ' : ' data ' } )
torrent_rows = torrent_table . find_all ( ' tr ' ) if torrent_table else [ ]
#Continue only if one Release is found
if len ( torrent_rows ) < 2 :
logger . log ( u " The Data returned from " + self . name + " do not contains any torrent " , logger . WARNING )
continue
for tr in torrent_rows [ 1 : ] :
try :
link = self . url + ( tr . find ( ' div ' , { ' class ' : ' torrentname ' } ) . find_all ( ' a ' ) [ 1 ] ) [ ' href ' ]
id = tr . get ( ' id ' ) [ - 7 : ]
title = ( tr . find ( ' div ' , { ' class ' : ' torrentname ' } ) . find_all ( ' a ' ) [ 1 ] ) . text
url = tr . find ( ' a ' , ' imagnet ' ) [ ' href ' ]
verified = True if tr . find ( ' a ' , ' iverify ' ) else False
trusted = True if tr . find ( ' img ' , { ' alt ' : ' verified ' } ) else False
seeders = int ( tr . find_all ( ' td ' ) [ - 2 ] . text )
leechers = int ( tr . find_all ( ' td ' ) [ - 1 ] . text )
except ( AttributeError , TypeError ) :
continue
if mode != ' RSS ' and seeders == 0 :
continue
if sickbeard . KAT_VERIFIED and not verified :
logger . log ( u " KAT Provider found result " + title + " but that doesn ' t seem like a verified result so I ' m ignoring it " , logger . DEBUG )
continue
#Check number video files = episode in season and find the real Quality for full season torrent analyzing files in torrent
if mode == ' Season ' :
ep_number = int ( len ( search_params [ ' Episode ' ] ) / len ( set ( allPossibleShowNames ( self . show ) ) ) )
title = self . _find_season_quality ( title , link , ep_number )
if not title or not url :
continue
item = title , url , id , seeders , leechers
items [ mode ] . append ( item )
except Exception , e :
logger . log ( u " Failed to parsing " + self . name + " Traceback: " + traceback . format_exc ( ) , logger . ERROR )
#For each search mode sort all the items by seeders
items [ mode ] . sort ( key = lambda tup : tup [ 3 ] , reverse = True )
results + = items [ mode ]
return results
def _get_title_and_url ( self , item ) :
title , url , id , seeders , leechers = item
if url :
url = url . replace ( ' & ' , ' & ' )
return ( title , url )
def getURL ( self , url , headers = None ) :
try :
# Remove double-slashes from url
parsed = list ( urlparse . urlparse ( url ) )
parsed [ 2 ] = re . sub ( " / { 2,} " , " / " , parsed [ 2 ] ) # replace two or more / with one
url = urlparse . urlunparse ( parsed )
r = requests . get ( url )
except ( requests . exceptions . ConnectionError , requests . exceptions . HTTPError ) , e :
logger . log ( u " Error loading " + self . name + " URL: " + str ( sys . exc_info ( ) ) + " - " + ex ( e ) , logger . ERROR )
return None
if r . status_code != 200 :
logger . log ( self . name + u " page requested with url " + url + " returned status code is " + str ( r . status_code ) + ' : ' + clients . http_error_code [ r . status_code ] , logger . WARNING )
return None
return r . content
def downloadResult ( self , result ) :
"""
Save the result to disk .
"""
torrent_hash = re . findall ( ' urn:btih:([ \ w] { 32,40}) ' , result . url ) [ 0 ] . upper ( )
if not torrent_hash :
logger . log ( " Unable to extract torrent hash from link: " + ex ( result . url ) , logger . ERROR )
return False
try :
r = requests . get ( ' http://torcache.net/torrent/ ' + torrent_hash + ' .torrent ' )
except Exception , e :
logger . log ( " Unable to connect to Torcache: " + ex ( e ) , logger . ERROR )
return False
if not r . status_code == 200 :
return False
magnetFileName = ek . ek ( os . path . join , sickbeard . TORRENT_DIR , helpers . sanitizeFileName ( result . name ) + ' . ' + self . providerType )
magnetFileContent = r . content
try :
with open ( magnetFileName , ' wb ' ) as fileOut :
fileOut . write ( magnetFileContent )
helpers . chmodAsParent ( magnetFileName )
2014-03-20 08:15:22 +00:00
except EnvironmentError , e :
2014-03-10 05:18:05 +00:00
logger . log ( " Unable to save the file: " + ex ( e ) , logger . ERROR )
return False
logger . log ( u " Saved magnet link to " + magnetFileName + " " , logger . MESSAGE )
return True
def findPropers ( self , search_date = datetime . datetime . today ( ) ) :
results = [ ]
sqlResults = db . DBConnection ( ) . select ( ' SELECT s.show_name, e.showid, e.season, e.episode, e.status, e.airdate, s.indexer FROM tv_episodes AS e ' +
' INNER JOIN tv_shows AS s ON (e.showid = s.indexer_id) ' +
' WHERE e.airdate >= ' + str ( search_date . toordinal ( ) ) +
' AND (e.status IN ( ' + ' , ' . join ( [ str ( x ) for x in Quality . DOWNLOADED ] ) + ' ) ' +
' OR (e.status IN ( ' + ' , ' . join ( [ str ( x ) for x in Quality . SNATCHED ] ) + ' ))) '
)
if not sqlResults :
return [ ]
for sqlShow in sqlResults :
curShow = helpers . findCertainShow ( sickbeard . showList , int ( sqlShow [ " showid " ] ) )
2014-03-20 08:15:22 +00:00
curEp = curShow . getEpisode ( int ( sqlShow [ " season " ] ) , int ( sqlShow [ " episode " ] ) )
2014-03-10 05:18:05 +00:00
searchString = self . _get_episode_search_strings ( curEp , add_string = ' PROPER|REPACK ' )
for item in self . _doSearch ( searchString [ 0 ] ) :
title , url = self . _get_title_and_url ( item )
results . append ( classes . Proper ( title , url , datetime . datetime . today ( ) ) )
return results
class KATCache ( tvcache . TVCache ) :
def __init__ ( self , provider ) :
tvcache . TVCache . __init__ ( self , provider )
# only poll ThePirateBay every 10 minutes max
self . minTime = 20
def updateCache ( self ) :
if not self . shouldUpdate ( ) :
return
search_params = { ' RSS ' : [ ' rss ' ] }
rss_results = self . provider . _doSearch ( search_params )
if rss_results :
self . setLastUpdate ( )
else :
return [ ]
logger . log ( u " Clearing " + self . provider . name + " cache and updating with new information " )
self . _clearCache ( )
cl = [ ]
for result in rss_results :
item = ( result [ 0 ] , result [ 1 ] )
ci = self . _parseItem ( item )
if ci is not None :
cl . append ( ci )
if len ( cl ) > 0 :
myDB = self . _getDB ( )
myDB . mass_action ( cl )
def _parseItem ( self , item ) :
( title , url ) = item
if not title or not url :
return None
logger . log ( u " Adding item to cache: " + title , logger . DEBUG )
return self . _addCacheEntry ( title , url )
provider = KATProvider ( )