mirror of
https://github.com/SickGear/SickGear.git
synced 2024-12-21 02:03:38 +00:00
388 lines
15 KiB
Python
388 lines
15 KiB
Python
# Use of this source code is governed by the MIT license.
|
|
__license__ = "MIT"
|
|
|
|
__all__ = [
|
|
'LXMLTreeBuilderForXML',
|
|
'LXMLTreeBuilder',
|
|
]
|
|
|
|
try:
|
|
from collections.abc import Callable # Python 3.6
|
|
except ImportError as e:
|
|
from collections import Callable
|
|
|
|
from io import BytesIO
|
|
from io import StringIO
|
|
from lxml import etree
|
|
from ..element import (
|
|
Comment,
|
|
Doctype,
|
|
NamespacedAttribute,
|
|
ProcessingInstruction,
|
|
XMLProcessingInstruction,
|
|
)
|
|
from ..builder import (
|
|
DetectsXMLParsedAsHTML,
|
|
FAST,
|
|
HTML,
|
|
HTMLTreeBuilder,
|
|
PERMISSIVE,
|
|
ParserRejectedMarkup,
|
|
TreeBuilder,
|
|
XML)
|
|
from ..dammit import EncodingDetector
|
|
|
|
LXML = 'lxml'
|
|
|
|
def _invert(d):
|
|
"Invert a dictionary."
|
|
return dict((v,k) for k, v in list(d.items()))
|
|
|
|
class LXMLTreeBuilderForXML(TreeBuilder):
|
|
DEFAULT_PARSER_CLASS = etree.XMLParser
|
|
|
|
is_xml = True
|
|
processing_instruction_class = XMLProcessingInstruction
|
|
|
|
NAME = "lxml-xml"
|
|
ALTERNATE_NAMES = ["xml"]
|
|
|
|
# Well, it's permissive by XML parser standards.
|
|
features = [NAME, LXML, XML, FAST, PERMISSIVE]
|
|
|
|
CHUNK_SIZE = 512
|
|
|
|
# This namespace mapping is specified in the XML Namespace
|
|
# standard.
|
|
DEFAULT_NSMAPS = dict(xml='http://www.w3.org/XML/1998/namespace')
|
|
|
|
DEFAULT_NSMAPS_INVERTED = _invert(DEFAULT_NSMAPS)
|
|
|
|
# NOTE: If we parsed Element objects and looked at .sourceline,
|
|
# we'd be able to see the line numbers from the original document.
|
|
# But instead we build an XMLParser or HTMLParser object to serve
|
|
# as the target of parse messages, and those messages don't include
|
|
# line numbers.
|
|
# See: https://bugs.launchpad.net/lxml/+bug/1846906
|
|
|
|
def initialize_soup(self, soup):
|
|
"""Let the BeautifulSoup object know about the standard namespace
|
|
mapping.
|
|
|
|
:param soup: A `BeautifulSoup`.
|
|
"""
|
|
super(LXMLTreeBuilderForXML, self).initialize_soup(soup)
|
|
self._register_namespaces(self.DEFAULT_NSMAPS)
|
|
|
|
def _register_namespaces(self, mapping):
|
|
"""Let the BeautifulSoup object know about namespaces encountered
|
|
while parsing the document.
|
|
|
|
This might be useful later on when creating CSS selectors.
|
|
|
|
This will track (almost) all namespaces, even ones that were
|
|
only in scope for part of the document. If two namespaces have
|
|
the same prefix, only the first one encountered will be
|
|
tracked. Un-prefixed namespaces are not tracked.
|
|
|
|
:param mapping: A dictionary mapping namespace prefixes to URIs.
|
|
"""
|
|
for key, value in list(mapping.items()):
|
|
# This is 'if key' and not 'if key is not None' because we
|
|
# don't track un-prefixed namespaces. Soupselect will
|
|
# treat an un-prefixed namespace as the default, which
|
|
# causes confusion in some cases.
|
|
if key and key not in self.soup._namespaces:
|
|
# Let the BeautifulSoup object know about a new namespace.
|
|
# If there are multiple namespaces defined with the same
|
|
# prefix, the first one in the document takes precedence.
|
|
self.soup._namespaces[key] = value
|
|
|
|
def default_parser(self, encoding):
|
|
"""Find the default parser for the given encoding.
|
|
|
|
:param encoding: A string.
|
|
:return: Either a parser object or a class, which
|
|
will be instantiated with default arguments.
|
|
"""
|
|
if self._default_parser is not None:
|
|
return self._default_parser
|
|
return etree.XMLParser(
|
|
target=self, strip_cdata=False, recover=True, encoding=encoding)
|
|
|
|
def parser_for(self, encoding):
|
|
"""Instantiate an appropriate parser for the given encoding.
|
|
|
|
:param encoding: A string.
|
|
:return: A parser object such as an `etree.XMLParser`.
|
|
"""
|
|
# Use the default parser.
|
|
parser = self.default_parser(encoding)
|
|
|
|
if isinstance(parser, Callable):
|
|
# Instantiate the parser with default arguments
|
|
parser = parser(
|
|
target=self, strip_cdata=False, recover=True, encoding=encoding
|
|
)
|
|
return parser
|
|
|
|
def __init__(self, parser=None, empty_element_tags=None, **kwargs):
|
|
# TODO: Issue a warning if parser is present but not a
|
|
# callable, since that means there's no way to create new
|
|
# parsers for different encodings.
|
|
self._default_parser = parser
|
|
if empty_element_tags is not None:
|
|
self.empty_element_tags = set(empty_element_tags)
|
|
self.soup = None
|
|
self.nsmaps = [self.DEFAULT_NSMAPS_INVERTED]
|
|
self.active_namespace_prefixes = [dict(self.DEFAULT_NSMAPS)]
|
|
super(LXMLTreeBuilderForXML, self).__init__(**kwargs)
|
|
|
|
def _getNsTag(self, tag):
|
|
# Split the namespace URL out of a fully-qualified lxml tag
|
|
# name. Copied from lxml's src/lxml/sax.py.
|
|
if tag[0] == '{':
|
|
return tuple(tag[1:].split('}', 1))
|
|
else:
|
|
return (None, tag)
|
|
|
|
def prepare_markup(self, markup, user_specified_encoding=None,
|
|
exclude_encodings=None,
|
|
document_declared_encoding=None):
|
|
"""Run any preliminary steps necessary to make incoming markup
|
|
acceptable to the parser.
|
|
|
|
lxml really wants to get a bytestring and convert it to
|
|
Unicode itself. So instead of using UnicodeDammit to convert
|
|
the bytestring to Unicode using different encodings, this
|
|
implementation uses EncodingDetector to iterate over the
|
|
encodings, and tell lxml to try to parse the document as each
|
|
one in turn.
|
|
|
|
:param markup: Some markup -- hopefully a bytestring.
|
|
:param user_specified_encoding: The user asked to try this encoding.
|
|
:param document_declared_encoding: The markup itself claims to be
|
|
in this encoding.
|
|
:param exclude_encodings: The user asked _not_ to try any of
|
|
these encodings.
|
|
|
|
:yield: A series of 4-tuples:
|
|
(markup, encoding, declared encoding,
|
|
has undergone character replacement)
|
|
|
|
Each 4-tuple represents a strategy for converting the
|
|
document to Unicode and parsing it. Each strategy will be tried
|
|
in turn.
|
|
"""
|
|
is_html = not self.is_xml
|
|
if is_html:
|
|
self.processing_instruction_class = ProcessingInstruction
|
|
# We're in HTML mode, so if we're given XML, that's worth
|
|
# noting.
|
|
DetectsXMLParsedAsHTML.warn_if_markup_looks_like_xml(
|
|
markup, stacklevel=3
|
|
)
|
|
else:
|
|
self.processing_instruction_class = XMLProcessingInstruction
|
|
|
|
if isinstance(markup, str):
|
|
# We were given Unicode. Maybe lxml can parse Unicode on
|
|
# this system?
|
|
|
|
# TODO: This is a workaround for
|
|
# https://bugs.launchpad.net/lxml/+bug/1948551.
|
|
# We can remove it once the upstream issue is fixed.
|
|
if len(markup) > 0 and markup[0] == u'\N{BYTE ORDER MARK}':
|
|
markup = markup[1:]
|
|
yield markup, None, document_declared_encoding, False
|
|
|
|
if isinstance(markup, str):
|
|
# No, apparently not. Convert the Unicode to UTF-8 and
|
|
# tell lxml to parse it as UTF-8.
|
|
yield (markup.encode("utf8"), "utf8",
|
|
document_declared_encoding, False)
|
|
|
|
# This was provided by the end-user; treat it as a known
|
|
# definite encoding per the algorithm laid out in the HTML5
|
|
# spec. (See the EncodingDetector class for details.)
|
|
known_definite_encodings = [user_specified_encoding]
|
|
|
|
# This was found in the document; treat it as a slightly lower-priority
|
|
# user encoding.
|
|
user_encodings = [document_declared_encoding]
|
|
detector = EncodingDetector(
|
|
markup, known_definite_encodings=known_definite_encodings,
|
|
user_encodings=user_encodings, is_html=is_html,
|
|
exclude_encodings=exclude_encodings
|
|
)
|
|
for encoding in detector.encodings:
|
|
yield (detector.markup, encoding, document_declared_encoding, False)
|
|
|
|
def feed(self, markup):
|
|
if isinstance(markup, bytes):
|
|
markup = BytesIO(markup)
|
|
elif isinstance(markup, str):
|
|
markup = StringIO(markup)
|
|
|
|
# Call feed() at least once, even if the markup is empty,
|
|
# or the parser won't be initialized.
|
|
data = markup.read(self.CHUNK_SIZE)
|
|
try:
|
|
self.parser = self.parser_for(self.soup.original_encoding)
|
|
self.parser.feed(data)
|
|
while len(data) != 0:
|
|
# Now call feed() on the rest of the data, chunk by chunk.
|
|
data = markup.read(self.CHUNK_SIZE)
|
|
if len(data) != 0:
|
|
self.parser.feed(data)
|
|
self.parser.close()
|
|
except (UnicodeDecodeError, LookupError, etree.ParserError) as e:
|
|
raise ParserRejectedMarkup(e)
|
|
|
|
def close(self):
|
|
self.nsmaps = [self.DEFAULT_NSMAPS_INVERTED]
|
|
|
|
def start(self, name, attrs, nsmap={}):
|
|
# Make sure attrs is a mutable dict--lxml may send an immutable dictproxy.
|
|
attrs = dict(attrs)
|
|
nsprefix = None
|
|
# Invert each namespace map as it comes in.
|
|
if len(nsmap) == 0 and len(self.nsmaps) > 1:
|
|
# There are no new namespaces for this tag, but
|
|
# non-default namespaces are in play, so we need a
|
|
# separate tag stack to know when they end.
|
|
self.nsmaps.append(None)
|
|
elif len(nsmap) > 0:
|
|
# A new namespace mapping has come into play.
|
|
|
|
# First, Let the BeautifulSoup object know about it.
|
|
self._register_namespaces(nsmap)
|
|
|
|
# Then, add it to our running list of inverted namespace
|
|
# mappings.
|
|
self.nsmaps.append(_invert(nsmap))
|
|
|
|
# The currently active namespace prefixes have
|
|
# changed. Calculate the new mapping so it can be stored
|
|
# with all Tag objects created while these prefixes are in
|
|
# scope.
|
|
current_mapping = dict(self.active_namespace_prefixes[-1])
|
|
current_mapping.update(nsmap)
|
|
|
|
# We should not track un-prefixed namespaces as we can only hold one
|
|
# and it will be recognized as the default namespace by soupsieve,
|
|
# which may be confusing in some situations.
|
|
if '' in current_mapping:
|
|
del current_mapping['']
|
|
self.active_namespace_prefixes.append(current_mapping)
|
|
|
|
# Also treat the namespace mapping as a set of attributes on the
|
|
# tag, so we can recreate it later.
|
|
attrs = attrs.copy()
|
|
for prefix, namespace in list(nsmap.items()):
|
|
attribute = NamespacedAttribute(
|
|
"xmlns", prefix, "http://www.w3.org/2000/xmlns/")
|
|
attrs[attribute] = namespace
|
|
|
|
# Namespaces are in play. Find any attributes that came in
|
|
# from lxml with namespaces attached to their names, and
|
|
# turn then into NamespacedAttribute objects.
|
|
new_attrs = {}
|
|
for attr, value in list(attrs.items()):
|
|
namespace, attr = self._getNsTag(attr)
|
|
if namespace is None:
|
|
new_attrs[attr] = value
|
|
else:
|
|
nsprefix = self._prefix_for_namespace(namespace)
|
|
attr = NamespacedAttribute(nsprefix, attr, namespace)
|
|
new_attrs[attr] = value
|
|
attrs = new_attrs
|
|
|
|
namespace, name = self._getNsTag(name)
|
|
nsprefix = self._prefix_for_namespace(namespace)
|
|
self.soup.handle_starttag(
|
|
name, namespace, nsprefix, attrs,
|
|
namespaces=self.active_namespace_prefixes[-1]
|
|
)
|
|
|
|
def _prefix_for_namespace(self, namespace):
|
|
"""Find the currently active prefix for the given namespace."""
|
|
if namespace is None:
|
|
return None
|
|
for inverted_nsmap in reversed(self.nsmaps):
|
|
if inverted_nsmap is not None and namespace in inverted_nsmap:
|
|
return inverted_nsmap[namespace]
|
|
return None
|
|
|
|
def end(self, name):
|
|
self.soup.endData()
|
|
completed_tag = self.soup.tagStack[-1]
|
|
namespace, name = self._getNsTag(name)
|
|
nsprefix = None
|
|
if namespace is not None:
|
|
for inverted_nsmap in reversed(self.nsmaps):
|
|
if inverted_nsmap is not None and namespace in inverted_nsmap:
|
|
nsprefix = inverted_nsmap[namespace]
|
|
break
|
|
self.soup.handle_endtag(name, nsprefix)
|
|
if len(self.nsmaps) > 1:
|
|
# This tag, or one of its parents, introduced a namespace
|
|
# mapping, so pop it off the stack.
|
|
out_of_scope_nsmap = self.nsmaps.pop()
|
|
|
|
if out_of_scope_nsmap is not None:
|
|
# This tag introduced a namespace mapping which is no
|
|
# longer in scope. Recalculate the currently active
|
|
# namespace prefixes.
|
|
self.active_namespace_prefixes.pop()
|
|
|
|
def pi(self, target, data):
|
|
self.soup.endData()
|
|
data = target + ' ' + data
|
|
self.soup.handle_data(data)
|
|
self.soup.endData(self.processing_instruction_class)
|
|
|
|
def data(self, content):
|
|
self.soup.handle_data(content)
|
|
|
|
def doctype(self, name, pubid, system):
|
|
self.soup.endData()
|
|
doctype = Doctype.for_name_and_ids(name, pubid, system)
|
|
self.soup.object_was_parsed(doctype)
|
|
|
|
def comment(self, content):
|
|
"Handle comments as Comment objects."
|
|
self.soup.endData()
|
|
self.soup.handle_data(content)
|
|
self.soup.endData(Comment)
|
|
|
|
def test_fragment_to_document(self, fragment):
|
|
"""See `TreeBuilder`."""
|
|
return '<?xml version="1.0" encoding="utf-8"?>\n%s' % fragment
|
|
|
|
|
|
class LXMLTreeBuilder(HTMLTreeBuilder, LXMLTreeBuilderForXML):
|
|
|
|
NAME = LXML
|
|
ALTERNATE_NAMES = ["lxml-html"]
|
|
|
|
features = ALTERNATE_NAMES + [NAME, HTML, FAST, PERMISSIVE]
|
|
is_xml = False
|
|
processing_instruction_class = ProcessingInstruction
|
|
|
|
def default_parser(self, encoding):
|
|
return etree.HTMLParser
|
|
|
|
def feed(self, markup):
|
|
encoding = self.soup.original_encoding
|
|
try:
|
|
self.parser = self.parser_for(encoding)
|
|
self.parser.feed(markup)
|
|
self.parser.close()
|
|
except (UnicodeDecodeError, LookupError, etree.ParserError) as e:
|
|
raise ParserRejectedMarkup(e)
|
|
|
|
|
|
def test_fragment_to_document(self, fragment):
|
|
"""See `TreeBuilder`."""
|
|
return '<html><body>%s</body></html>' % fragment
|