SickGear/_cleaner.py

186 lines
9 KiB
Python
Raw Permalink Normal View History

# remove this file when no longer needed
import os
import importlib.util
import io
import shutil
magic_number = importlib.util.MAGIC_NUMBER.hex()
parent_dir = os.path.abspath(os.path.dirname(__file__))
magic_number_file = os.path.join(parent_dir, '.python_magic.tmp')
old_magic = ''
try:
if os.path.isfile(magic_number_file):
with io.open(magic_number_file, 'r', encoding='utf-8') as mf:
old_magic = mf.read()
except (BaseException, Exception):
pass
if old_magic != magic_number:
# print('Python magic changed: removing all .pyc, .pyo files')
for pc in ['sickgear', 'lib']:
search_dir = os.path.join(parent_dir, pc)
for dpath, dnames, fnames in os.walk(search_dir):
for filename in [fn for fn in fnames if os.path.splitext(fn)[-1].lower() in ('.pyc', '.pyo')]:
try:
os.remove(os.path.abspath(os.path.join(dpath, filename)))
except (BaseException, Exception):
pass
try:
with io.open(magic_number_file, 'w+') as mf:
mf.write(magic_number)
except (BaseException, Exception):
pass
# print('finished')
# skip cleaned005 as used during dev by testers
cleanups = [
['.cleaned008.tmp', r'lib\tornado_py3', [
r'lib\bs4_py2\builder\__pycache__', r'lib\bs4_py2\builder', r'lib\bs4_py2',
r'lib\bs4_py3\builder\__pycache__', r'lib\bs4_py3\builder', r'lib\bs4_py3',
r'lib\diskcache_py2\__pycache__', r'lib\diskcache_py2',
r'lib\diskcache_py3\__pycache__', r'lib\diskcache_py3',
r'lib\feedparser_py2\datetimes\__pycache__', r'lib\feedparser_py2\datetimes',
r'lib\feedparser_py2\namespaces\__pycache__', r'lib\feedparser_py2\namespaces',
r'lib\feedparser_py2\parsers\__pycache__', r'lib\feedparser_py2\parsers',
r'lib\feedparser_py2\__pycache__', r'lib\feedparser_py2',
r'lib\feedparser_py3\datetimes\__pycache__', r'lib\feedparser_py3\datetimes',
r'lib\feedparser_py3\namespaces\__pycache__', r'lib\feedparser_py3\namespaces',
r'lib\feedparser_py3\parsers\__pycache__', r'lib\feedparser_py3\parsers',
r'lib\feedparser_py3\__pycache__', r'lib\feedparser_py3',
r'lib\hachoir_py2\core\__pycache__', r'lib\hachoir_py2\core',
r'lib\hachoir_py2\field\__pycache__', r'lib\hachoir_py2\field',
r'lib\hachoir_py2\metadata\__pycache__', r'lib\hachoir_py2\metadata',
r'lib\hachoir_py2\parser\__pycache__', r'lib\hachoir_py2\parser',
r'lib\hachoir_py2\stream\__pycache__', r'lib\hachoir_py2\stream',
r'lib\hachoir_py2\__pycache__', r'lib\hachoir_py2',
r'lib\hachoir_py3\core\__pycache__', r'lib\hachoir_py3\core',
r'lib\hachoir_py3\field\__pycache__', r'lib\hachoir_py3\field',
r'lib\hachoir_py3\metadata\__pycache__', r'lib\hachoir_py3\metadata',
r'lib\hachoir_py3\parser\__pycache__', r'lib\hachoir_py3\parser',
r'lib\hachoir_py3\stream\__pycache__', r'lib\hachoir_py3\stream',
r'lib\hachoir_py3\__pycache__', r'lib\hachoir_py3',
r'lib\idna_py2\__pycache__', r'lib\idna_py2',
r'lib\idna_py3\__pycache__', r'lib\idna_py3',
r'lib\rarfile_py2\__pycache__', r'lib\rarfile_py2',
r'lib\rarfile_py3\__pycache__', r'lib\rarfile_py3',
r'lib\requests_py2\__pycache__', r'lib\requests_py2',
r'lib\requests_py3\__pycache__', r'lib\requests_py3',
r'lib\soupsieve_py2\__pycache__', r'lib\soupsieve_py2',
r'lib\soupsieve_py3\__pycache__', r'lib\soupsieve_py3',
r'lib\tornado_py2\platform\__pycache__', r'lib\tornado_py2\platform',
r'lib\tornado_py2\__pycache__', r'lib\tornado_py2',
r'lib\tornado_py3\platform\__pycache__', r'lib\tornado_py3\platform',
r'lib\tornado_py3\__pycache__', r'lib\tornado_py3',
r'lib\urllib3\packages\ssl_match_hostname\__pycache__', r'lib\urllib3\packages\ssl_match_hostname',
r'sickbeard\clients\__pycache__', r'sickbeard\clients',
r'sickbeard\databases\__pycache__', r'sickbeard\databases',
r'sickbeard\indexers\__pycache__', r'sickbeard\indexers',
r'sickbeard\metadata\__pycache__', r'sickbeard\metadata',
r'sickbeard\name_parser\__pycache__', r'sickbeard\name_parser',
r'sickbeard\notifiers\__pycache__', r'sickbeard\notifiers',
r'sickbeard\providers\__pycache__', r'sickbeard\providers',
r'sickbeard\__pycache__', r'sickbeard',
]],
['.cleaned007.tmp', r'lib\tvmaze_api', [
r'lib\imdb_api\__pycache__', r'lib\imdb_api',
r'lib\libtrakt\__pycache__', r'lib\libtrakt',
r'lib\tvdb_api\__pycache__', r'lib\tvdb_api',
r'lib\tvmaze_api\__pycache__', r'lib\tvmaze_api']],
['.cleaned006.tmp', r'lib\boto', [
r'lib\boto', r'lib\growl',
r'lib\hachoir\core', r'lib\hachoir\field', r'lib\hachoir\metadata',
r'lib\hachoir\parser\archive', r'lib\hachoir\parser\audio',
r'lib\hachoir\parser\common', r'lib\hachoir\parser\container',
r'lib\hachoir\parser\image', r'lib\hachoir\parser\misc',
r'lib\hachoir\parser\network', r'lib\hachoir\parser\program',
r'lib\hachoir\parser\video', r'lib\hachoir\parser', r'lib\hachoir\stream',
r'lib\httplib2\lib\oauth2\lib\pythontwitter\lib\tmdb_api']],
['.cleaned004.tmp', r'lib\requests\packages', [
r'lib\requests\packages', r'lib\pynma']],
['.cleaned003.tmp', r'lib\imdb', [
r'lib\imdb']],
['.cleaned002.tmp', r'lib\hachoir_core', [
'.cleaned.tmp', 'tornado',
r'lib\feedcache', r'lib\hachoir_core', r'lib\hachoir_metadata', r'lib\hachoir_parser',
r'lib\jsonrpclib', r'lib\shove', r'lib\trakt', r'lib\tvrage_api', r'lib\unrar2']],
]
for cleaned_path, test_path, dir_list in cleanups:
cleaned_file = os.path.abspath(os.path.join(parent_dir, cleaned_path))
test = os.path.abspath(os.path.join(parent_dir, *test_path.split('\\')))
if not os.path.isfile(cleaned_file) or os.path.exists(test):
dead_dirs = [os.path.abspath(os.path.join(parent_dir, *d.split('\\'))) for d in dir_list]
for dpath, dnames, fnames in os.walk(parent_dir):
for dead_dir in filter(lambda x: x in dead_dirs, [os.path.abspath(os.path.join(dpath, d)) for d in dnames]):
try:
shutil.rmtree(dead_dir)
except (BaseException, Exception):
pass
for filename in [fn for fn in fnames if os.path.splitext(fn)[-1].lower() in ('.pyc', '.pyo')]:
try:
os.remove(os.path.abspath(os.path.join(dpath, filename)))
except (BaseException, Exception):
pass
with io.open(cleaned_file, 'w+', encoding='utf-8') as fp:
fp.write(u'This file exists to prevent a rerun delete of *.pyc, *.pyo files')
fp.flush()
os.fsync(fp.fileno())
cleaned_file = os.path.abspath(os.path.join(parent_dir, '.cleaned_html5lib.tmp'))
test = os.path.abspath(os.path.join(parent_dir, 'lib', 'html5lib', 'treebuilders', '_base.pyc'))
danger_output = os.path.abspath(os.path.join(parent_dir, '__README-DANGER.txt'))
bad_files = []
if not os.path.isfile(cleaned_file) or os.path.exists(test):
for dead_path in [os.path.abspath(os.path.join(parent_dir, *d)) for d in [
('lib', 'html5lib', 'trie'),
('lib', 'html5lib', 'serializer')
]]:
try:
shutil.rmtree(dead_path)
except (BaseException, Exception):
pass
for dead_file in [os.path.abspath(os.path.join(parent_dir, *d)) for d in [
('lib', 'html5lib', 'ihatexml.py'),
('lib', 'html5lib', 'inputstream.py'),
('lib', 'html5lib', 'tokenizer.py'),
('lib', 'html5lib', 'utils.py'),
('lib', 'html5lib', 'filters', '_base.py'),
('lib', 'html5lib', 'sanitizer.py'),
('lib', 'html5lib', 'treebuilders', '_base.py'),
('lib', 'html5lib', 'treewalkers', '_base.py'),
('lib', 'html5lib', 'treewalkers', 'lxmletree.py'),
('lib', 'html5lib', 'treewalkers', 'genshistream.py'),
]]:
for ext in ['', 'c', 'o']:
name = '%s.py%s' % (os.path.splitext(dead_file)[:-1][0], ext)
if os.path.exists(name):
try:
os.remove(name)
except (BaseException, Exception):
bad_files += [name]
if any(bad_files):
swap_name = cleaned_file
cleaned_file = danger_output
danger_output = swap_name
msg = u'Failed (permissions?) to delete file(s). You must manually delete:\r\n%s' % '\r\n'.join(bad_files)
print(msg)
else:
msg = u'This file exists to prevent a rerun delete of dead lib/html5lib files'
with io.open(cleaned_file, 'w+', encoding='utf-8') as fp:
fp.write(msg)
fp.flush()
os.fsync(fp.fileno())
try:
os.remove(danger_output)
except (BaseException, Exception):
pass