2018-09-17 08:27:00 +08:00
|
|
|
"""Diagnostic functions, mainly for use when doing tech support."""
|
|
|
|
|
2019-09-28 12:22:17 +08:00
|
|
|
# Use of this source code is governed by the MIT license.
|
2018-09-17 08:27:00 +08:00
|
|
|
__license__ = "MIT"
|
|
|
|
|
|
|
|
import cProfile
|
2019-09-28 12:22:17 +08:00
|
|
|
from io import StringIO
|
|
|
|
from html.parser import HTMLParser
|
2018-09-17 08:27:00 +08:00
|
|
|
import bs4
|
|
|
|
from bs4 import BeautifulSoup, __version__
|
|
|
|
from bs4.builder import builder_registry
|
|
|
|
|
|
|
|
import os
|
|
|
|
import pstats
|
|
|
|
import random
|
|
|
|
import tempfile
|
|
|
|
import time
|
|
|
|
import traceback
|
|
|
|
import sys
|
|
|
|
import cProfile
|
|
|
|
|
|
|
|
def diagnose(data):
|
|
|
|
"""Diagnostic suite for isolating common problems."""
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Diagnostic running on Beautiful Soup %s" % __version__)
|
|
|
|
print("Python version %s" % sys.version)
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
basic_parsers = ["html.parser", "html5lib", "lxml"]
|
|
|
|
for name in basic_parsers:
|
|
|
|
for builder in builder_registry.builders:
|
|
|
|
if name in builder.features:
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
basic_parsers.remove(name)
|
2019-09-24 18:23:11 +08:00
|
|
|
print((
|
2018-09-17 08:27:00 +08:00
|
|
|
"I noticed that %s is not installed. Installing it may help." %
|
2019-09-24 18:23:11 +08:00
|
|
|
name))
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
if 'lxml' in basic_parsers:
|
2019-09-28 12:22:17 +08:00
|
|
|
basic_parsers.append("lxml-xml")
|
2018-09-17 08:27:00 +08:00
|
|
|
try:
|
|
|
|
from lxml import etree
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Found lxml version %s" % ".".join(map(str,etree.LXML_VERSION)))
|
|
|
|
except ImportError as e:
|
2018-09-17 08:27:00 +08:00
|
|
|
print (
|
|
|
|
"lxml is not installed or couldn't be imported.")
|
|
|
|
|
|
|
|
|
|
|
|
if 'html5lib' in basic_parsers:
|
|
|
|
try:
|
|
|
|
import html5lib
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Found html5lib version %s" % html5lib.__version__)
|
|
|
|
except ImportError as e:
|
2018-09-17 08:27:00 +08:00
|
|
|
print (
|
|
|
|
"html5lib is not installed or couldn't be imported.")
|
|
|
|
|
|
|
|
if hasattr(data, 'read'):
|
|
|
|
data = data.read()
|
|
|
|
elif data.startswith("http:") or data.startswith("https:"):
|
2019-09-24 18:23:11 +08:00
|
|
|
print('"%s" looks like a URL. Beautiful Soup is not an HTTP client.' % data)
|
|
|
|
print("You need to use some other library to get the document behind the URL, and feed that document to Beautiful Soup.")
|
2018-09-17 08:27:00 +08:00
|
|
|
return
|
2019-09-28 12:22:17 +08:00
|
|
|
else:
|
|
|
|
try:
|
|
|
|
if os.path.exists(data):
|
|
|
|
print('"%s" looks like a filename. Reading data from the file.' % data)
|
|
|
|
with open(data) as fp:
|
|
|
|
data = fp.read()
|
|
|
|
except ValueError:
|
|
|
|
# This can happen on some platforms when the 'filename' is
|
|
|
|
# too long. Assume it's data and not a filename.
|
|
|
|
pass
|
|
|
|
print()
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
for parser in basic_parsers:
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Trying to parse your markup with %s" % parser)
|
2018-09-17 08:27:00 +08:00
|
|
|
success = False
|
|
|
|
try:
|
2019-09-28 12:22:17 +08:00
|
|
|
soup = BeautifulSoup(data, features=parser)
|
2018-09-17 08:27:00 +08:00
|
|
|
success = True
|
2019-09-24 18:23:11 +08:00
|
|
|
except Exception as e:
|
|
|
|
print("%s could not parse the markup." % parser)
|
2018-09-17 08:27:00 +08:00
|
|
|
traceback.print_exc()
|
|
|
|
if success:
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Here's what %s did with the markup:" % parser)
|
|
|
|
print(soup.prettify())
|
2018-09-17 08:27:00 +08:00
|
|
|
|
2019-09-24 18:23:11 +08:00
|
|
|
print("-" * 80)
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
def lxml_trace(data, html=True, **kwargs):
|
|
|
|
"""Print out the lxml events that occur during parsing.
|
|
|
|
|
|
|
|
This lets you see how lxml parses a document when no Beautiful
|
|
|
|
Soup code is running.
|
|
|
|
"""
|
|
|
|
from lxml import etree
|
|
|
|
for event, element in etree.iterparse(StringIO(data), html=html, **kwargs):
|
2019-09-24 18:23:11 +08:00
|
|
|
print(("%s, %4s, %s" % (event, element.tag, element.text)))
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
class AnnouncingParser(HTMLParser):
|
|
|
|
"""Announces HTMLParser parse events, without doing anything else."""
|
|
|
|
|
|
|
|
def _p(self, s):
|
|
|
|
print(s)
|
|
|
|
|
|
|
|
def handle_starttag(self, name, attrs):
|
|
|
|
self._p("%s START" % name)
|
|
|
|
|
|
|
|
def handle_endtag(self, name):
|
|
|
|
self._p("%s END" % name)
|
|
|
|
|
|
|
|
def handle_data(self, data):
|
|
|
|
self._p("%s DATA" % data)
|
|
|
|
|
|
|
|
def handle_charref(self, name):
|
|
|
|
self._p("%s CHARREF" % name)
|
|
|
|
|
|
|
|
def handle_entityref(self, name):
|
|
|
|
self._p("%s ENTITYREF" % name)
|
|
|
|
|
|
|
|
def handle_comment(self, data):
|
|
|
|
self._p("%s COMMENT" % data)
|
|
|
|
|
|
|
|
def handle_decl(self, data):
|
|
|
|
self._p("%s DECL" % data)
|
|
|
|
|
|
|
|
def unknown_decl(self, data):
|
|
|
|
self._p("%s UNKNOWN-DECL" % data)
|
|
|
|
|
|
|
|
def handle_pi(self, data):
|
|
|
|
self._p("%s PI" % data)
|
|
|
|
|
|
|
|
def htmlparser_trace(data):
|
|
|
|
"""Print out the HTMLParser events that occur during parsing.
|
|
|
|
|
|
|
|
This lets you see how HTMLParser parses a document when no
|
|
|
|
Beautiful Soup code is running.
|
|
|
|
"""
|
|
|
|
parser = AnnouncingParser()
|
|
|
|
parser.feed(data)
|
|
|
|
|
|
|
|
_vowels = "aeiou"
|
|
|
|
_consonants = "bcdfghjklmnpqrstvwxyz"
|
|
|
|
|
|
|
|
def rword(length=5):
|
|
|
|
"Generate a random word-like string."
|
|
|
|
s = ''
|
|
|
|
for i in range(length):
|
|
|
|
if i % 2 == 0:
|
|
|
|
t = _consonants
|
|
|
|
else:
|
|
|
|
t = _vowels
|
|
|
|
s += random.choice(t)
|
|
|
|
return s
|
|
|
|
|
|
|
|
def rsentence(length=4):
|
|
|
|
"Generate a random sentence-like string."
|
2019-09-28 12:22:17 +08:00
|
|
|
return " ".join(rword(random.randint(4,9)) for i in list(range(length)))
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
def rdoc(num_elements=1000):
|
|
|
|
"""Randomly generate an invalid HTML document."""
|
|
|
|
tag_names = ['p', 'div', 'span', 'i', 'b', 'script', 'table']
|
|
|
|
elements = []
|
|
|
|
for i in range(num_elements):
|
|
|
|
choice = random.randint(0,3)
|
|
|
|
if choice == 0:
|
|
|
|
# New tag.
|
|
|
|
tag_name = random.choice(tag_names)
|
|
|
|
elements.append("<%s>" % tag_name)
|
|
|
|
elif choice == 1:
|
|
|
|
elements.append(rsentence(random.randint(1,4)))
|
|
|
|
elif choice == 2:
|
|
|
|
# Close a tag.
|
|
|
|
tag_name = random.choice(tag_names)
|
|
|
|
elements.append("</%s>" % tag_name)
|
|
|
|
return "<html>" + "\n".join(elements) + "</html>"
|
|
|
|
|
|
|
|
def benchmark_parsers(num_elements=100000):
|
|
|
|
"""Very basic head-to-head performance benchmark."""
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Comparative parser benchmark on Beautiful Soup %s" % __version__)
|
2018-09-17 08:27:00 +08:00
|
|
|
data = rdoc(num_elements)
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Generated a large invalid HTML document (%d bytes)." % len(data))
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
for parser in ["lxml", ["lxml", "html"], "html5lib", "html.parser"]:
|
|
|
|
success = False
|
|
|
|
try:
|
|
|
|
a = time.time()
|
|
|
|
soup = BeautifulSoup(data, parser)
|
|
|
|
b = time.time()
|
|
|
|
success = True
|
2019-09-24 18:23:11 +08:00
|
|
|
except Exception as e:
|
|
|
|
print("%s could not parse the markup." % parser)
|
2018-09-17 08:27:00 +08:00
|
|
|
traceback.print_exc()
|
|
|
|
if success:
|
2019-09-24 18:23:11 +08:00
|
|
|
print("BS4+%s parsed the markup in %.2fs." % (parser, b-a))
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
from lxml import etree
|
|
|
|
a = time.time()
|
|
|
|
etree.HTML(data)
|
|
|
|
b = time.time()
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Raw lxml parsed the markup in %.2fs." % (b-a))
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
import html5lib
|
|
|
|
parser = html5lib.HTMLParser()
|
|
|
|
a = time.time()
|
|
|
|
parser.parse(data)
|
|
|
|
b = time.time()
|
2019-09-24 18:23:11 +08:00
|
|
|
print("Raw html5lib parsed the markup in %.2fs." % (b-a))
|
2018-09-17 08:27:00 +08:00
|
|
|
|
|
|
|
def profile(num_elements=100000, parser="lxml"):
|
|
|
|
|
|
|
|
filehandle = tempfile.NamedTemporaryFile()
|
|
|
|
filename = filehandle.name
|
|
|
|
|
|
|
|
data = rdoc(num_elements)
|
|
|
|
vars = dict(bs4=bs4, data=data, parser=parser)
|
|
|
|
cProfile.runctx('bs4.BeautifulSoup(data, parser)' , vars, vars, filename)
|
|
|
|
|
|
|
|
stats = pstats.Stats(filename)
|
|
|
|
# stats.strip_dirs()
|
|
|
|
stats.sort_stats("cumulative")
|
|
|
|
stats.print_stats('_html5lib|bs4', 50)
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
diagnose(sys.stdin.read())
|