-
Notifications
You must be signed in to change notification settings - Fork 17
/
fetcher.py
executable file
·489 lines (424 loc) · 18 KB
/
fetcher.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
"""
Greynir: Natural language processing for Icelandic
Fetcher module
Copyright (C) 2023 Miðeind ehf.
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see http://www.gnu.org/licenses/.
This module contains utility classes for web page fetching and tokenization.
"""
from typing import Any, Dict, Iterator, List, Optional, Set, Tuple, cast
from types import ModuleType
import re
import importlib
import logging
import requests
import urllib.parse as urlparse
from urllib.error import HTTPError
from bs4 import BeautifulSoup, NavigableString, Tag # type: ignore
from reynir import tokenize, Tok
from nertokenizer import recognize_entities
from db import SessionContext, Session
from db.models import Root, Article as ArticleRow
# The HTML parser to use with BeautifulSoup
# _HTML_PARSER = "html5lib"
_HTML_PARSER = "html.parser"
# The user agent string to send in headers when fetching URLs
_USER_AGENT = "GreynirScraper (https://greynir.is)"
class Fetcher:
"""The worker class that scrapes the known roots"""
# HTML tags that we explicitly don't want to look at
_EXCLUDE_TAGS = frozenset(["script", "audio", "video", "style"])
# HTML tags that typically denote blocks (DIV-like),
# not inline constructs (SPAN-like)
_BLOCK_TAGS = frozenset(
[
"p",
"h1",
"h2",
"h3",
"h4",
"div",
"main",
"article",
"header",
"section",
"table",
"thead",
"tbody",
"tr",
"td",
"ul",
"li",
"form",
"option",
"input",
"label",
"figure",
"figcaption",
"footer",
]
)
_INLINE_BLOCK_TAGS = frozenset(["span"]) # Inserted with whitespace
_WHITESPACE_TAGS = frozenset(["img"]) # Inserted as whitespace
_BREAK_TAGS = frozenset(["br", "hr"]) # Cause paragraph breaks at outermost level
# Cache of instantiated scrape helpers
_helpers: Dict[str, ModuleType] = dict()
def __init__(self):
"""No instances are supposed to be created of this class"""
assert False
class TextList:
"""Accumulates raw text blocks and eliminates
unnecessary nesting indicators"""
def __init__(self) -> None:
self._result: List[str] = []
self._nesting = 0
self._white = False
def append(self, w: str) -> None:
if self._nesting > 0:
if w.isspace():
# Whitespace is not reason to emit nesting markers
return
self._result.append("[[" * self._nesting)
self._nesting = 0
self._result.append(w)
self._white = False
def append_whitespace(self) -> None:
if self._nesting == 0:
# No need to append whitespace if we're just inside a begin-block
if not self._white:
self._result.append(" ")
self._white = True
def begin(self) -> None:
self._nesting += 1
def end(self) -> None:
if self._nesting > 0:
self._nesting -= 1
else:
self._result.append("]]")
self._white = True
def insert_break(self) -> None:
"""Used to cut paragraphs at <br> and <hr> tags"""
if self._nesting == 0:
self._result.append("]][[")
self._white = True
def result(self) -> str:
"""Return the accumulated result as a string"""
assert self._nesting == 0
text = "".join(self._result).strip()
# Eliminate soft hyphen and zero width space characters
text = re.sub("\u00AD|\u200B", "", text)
# Eliminate space before and after paragraph begin markers
text = re.sub(r"\s*\[\[\s*", "[[", text)
# Eliminate space before and after paragraph end markers
text = re.sub(r"\s*\]\]\s*", "]]", text)
# Eliminate consecutive whitespace
text = re.sub(r"\s+", " ", text)
return text
@staticmethod
def extract_text(soup: Optional[Tag], result: "Fetcher.TextList") -> None:
"""Append the human-readable text found in an HTML soup
to the result TextList"""
if soup is None:
return
for t in soup.children:
if type(t) is NavigableString:
# Text content node
result.append(cast(str, t))
elif isinstance(t, NavigableString):
# Comment, CDATA or other text data: ignore
pass
elif not isinstance(t, Tag):
pass
elif t.name in Fetcher._BREAK_TAGS:
result.insert_break()
# html.parser (erroneously) nests content inside
# <br> and <hr> tags
Fetcher.extract_text(t, result)
elif t.name in Fetcher._WHITESPACE_TAGS:
# Tags that we interpret as whitespace, such as <img>
result.append_whitespace()
# html.parser nests content inside <img> tags if
# they are not explicitly closed
Fetcher.extract_text(t, result)
elif t.name in Fetcher._BLOCK_TAGS:
# Nested block tag
result.begin() # Begin block
Fetcher.extract_text(t, result)
result.end() # End block
elif t.name in Fetcher._INLINE_BLOCK_TAGS:
# Put whitespace around the inline block
# so that words don't run together
result.append_whitespace()
Fetcher.extract_text(t, result)
result.append_whitespace()
elif t.name not in Fetcher._EXCLUDE_TAGS:
# Non-block tag
Fetcher.extract_text(t, result)
@staticmethod
def to_tokens(
soup: Tag, enclosing_session: Optional[Session] = None
) -> Iterator[Tok]:
"""Convert an HTML soup root into a parsable token stream"""
# Extract the text content of the HTML into a list
tlist = Fetcher.TextList()
Fetcher.extract_text(soup, tlist)
text = tlist.result()
# Tokenize the resulting text, returning a generator
token_stream = tokenize(text)
return recognize_entities(token_stream, enclosing_session=enclosing_session)
@classmethod
def raw_fetch_url(cls, url: str) -> Optional[str]:
"""Low-level fetch of an URL, returning a decoded string"""
html_doc = None
try:
# Normal external HTTP/HTTPS fetch
headers = {"User-Agent": _USER_AGENT}
r = requests.get(url, timeout=10, headers=headers)
# pylint: disable=no-member
if r.status_code == requests.codes.ok:
html_doc = r.text
else:
logging.warning(f"HTTP status {r.status_code} for URL {url}")
except requests.exceptions.ConnectionError as e:
logging.error(f"ConnectionError: {e} for URL {url}")
html_doc = None
except requests.exceptions.ChunkedEncodingError as e:
logging.error(f"ChunkedEncodingError: {e} for URL {url}")
html_doc = None
except HTTPError as e:
logging.error(f"HTTPError: {e} for URL {url}")
html_doc = None
except UnicodeEncodeError as e:
logging.error(f"Exception when opening URL {url}: {e}")
html_doc = None
except UnicodeDecodeError as e:
logging.error(f"Exception when decoding HTML of {url}: {e}")
html_doc = None
return html_doc
@classmethod
def _get_helper(cls, root: Root) -> Optional[ModuleType]:
"""Return a scrape helper instance for the given root"""
# Obtain an instance of a scraper helper class for this root
assert root.scr_module is not None
assert root.scr_class is not None
helper_id = root.scr_module + "." + root.scr_class
if helper_id in Fetcher._helpers:
# Already instantiated a helper: get it
helper = Fetcher._helpers[helper_id]
else:
# Dynamically instantiate a new helper class instance
mod = importlib.import_module(root.scr_module)
helper_class = getattr(mod, root.scr_class, None) if mod else None
helper = helper_class(root) if helper_class else None
if not helper:
logging.error(f"Unable to instantiate helper {helper_id}")
else:
Fetcher._helpers[helper_id] = helper
return helper
@staticmethod
def make_soup(doc: str, helper: Optional[ModuleType] = None) -> Optional[Tag]:
"""Convert a document to a soup, using the helper if available"""
if helper is None:
soup = BeautifulSoup(doc, _HTML_PARSER) if doc else None
if soup is None or soup.html is None:
return None
else:
soup = helper.make_soup(doc) if doc else None
return soup
@classmethod
def tokenize_html(
cls, url: str, html: str, enclosing_session: Optional[Session] = None
) -> Optional[Iterator[Tok]]:
"""Convert HTML into a token iterable (generator)"""
with SessionContext(enclosing_session) as session:
helper = cls.helper_for(session, url)
soup = Fetcher.make_soup(html, helper)
content: Optional[Tag]
if soup is None or soup.html is None:
content = None
elif helper is None:
content = soup.html.body
else:
content = cast(Any, helper).get_content(soup)
# Convert the content soup to a token iterable (generator)
return (
Fetcher.to_tokens(content, enclosing_session=session)
if content
else None
)
@staticmethod
def children(root: Root, soup: Tag) -> Set[str]:
"""Return a set of child URLs within a HTML soup,
relative to the given root"""
# Establish the root URL base parameters
root_s = urlparse.urlsplit(root.url)
root_url = urlparse.urlunsplit(root_s)
root_url_slash = urlparse.urlunsplit(
(root_s.scheme, root_s.netloc, "/", root_s.query, "")
)
# Collect all interesting <a> tags from the soup and obtain their href-s:
fetch: Set[str] = set()
for link in soup.find_all("a"):
href: Optional[str] = link.get("href")
if not href:
continue
# Split the href into its components
s = urlparse.urlsplit(href)
if s.scheme and s.scheme not in {"http", "https"}:
# Not HTTP
continue
if s.netloc and not (
s.netloc == root.domain or s.netloc.endswith("." + root.domain)
):
# External domain - we're not interested
continue
# Seems to be a bug in urllib: fragments are put into the
# path if there is no canonical path
newpath = s.path
if newpath.startswith("#") or newpath.startswith("/#"):
newpath = ""
if not newpath and not s.query:
# No meaningful path info present
continue
# Make sure the newpath is properly urlencoded
if newpath:
newpath = urlparse.quote(newpath)
# Fill in missing stuff from the root URL base parameters
newurl = (
s.scheme or root_s.scheme,
s.netloc or root_s.netloc,
newpath,
s.query,
"",
)
# Make a complete new URL to fetch
url = urlparse.urlunsplit(newurl)
if url in {root_url, root_url_slash}:
# Exclude the root URL
continue
# Looks legit: add to the fetch set
fetch.add(url)
return fetch
@classmethod
def helper_for(cls, session: Session, url: str) -> Optional[ModuleType]:
"""Return a scrape helper for the root of the given url"""
s = urlparse.urlsplit(url)
root: Optional[Root] = None
# Find which root this URL belongs to, if any
r: Root
for r in session.query(Root).all():
root_s = urlparse.urlsplit(r.url)
# Find the root of the domain, i.e. www.ruv.is -> ruv.is
root_domain = ".".join(root_s.netloc.split(".")[-2:])
# This URL belongs to a root if the domain (netloc) part
# ends with the root domain
if s.netloc == root_domain or s.netloc.endswith("." + root_domain):
root = r
break
# Obtain a scrape helper for the root, if any
return cls._get_helper(root) if root else None
# noinspection PyComparisonWithNone
@classmethod
def find_article(
cls, url: str, enclosing_session: Optional[Session] = None
) -> Optional[ArticleRow]:
"""Return a scraped article object, if found, else None"""
article = None
with SessionContext(enclosing_session, commit=True) as session:
article = (
session.query(ArticleRow)
.filter_by(url=url)
.filter(ArticleRow.scraped != None)
.one_or_none()
)
return article
# noinspection PyComparisonWithNone
@classmethod
def is_known_url(cls, url: str, session: Optional[Session] = None) -> bool:
"""Return True if the URL has already been scraped"""
return cls.find_article(url, session) is not None
@classmethod
def fetch_article(
cls, url: str, enclosing_session: Optional[Session] = None
) -> Tuple[Optional[ArticleRow], Any, Optional[Tag]]:
"""Fetch a previously scraped article, returning
a tuple (article, metadata, content) or None if error"""
with SessionContext(enclosing_session) as session:
article = cls.find_article(url, session)
if article is None:
return (None, None, None)
html_doc = article.html
if not html_doc:
return (None, None, None)
helper = cls.helper_for(session, url)
# Parse the HTML
soup = Fetcher.make_soup(html_doc, helper)
if soup is None or soup.html is None:
logging.warning(f"Fetcher.fetch_article({url}): No soup")
return (None, None, None)
# Obtain the metadata and the content from the resulting soup
metadata = cast(Any, helper).get_metadata(soup) if helper else None
content = cast(Any, helper).get_content(soup) if helper else soup.html.body
return (article, metadata, content)
@classmethod
def fetch_url(
cls, url: str, enclosing_session: Optional[Session] = None
) -> Optional[Tuple[Any, Optional[Tag]]]:
"""Fetch a URL using the scraping mechanism, returning
a tuple (metadata, content) or None if error"""
with SessionContext(enclosing_session) as session:
helper = cls.helper_for(session, url)
html_doc: Optional[str] = None
if helper is None or not hasattr(helper, "fetch_url"):
# Do a straight HTTP fetch
html_doc = cls.raw_fetch_url(url)
else:
# Hand off to the helper
html_doc = cast(Any, helper).fetch_url(url)
if not html_doc:
return None
# Parse the HTML
soup = Fetcher.make_soup(html_doc, helper)
if soup is None or soup.html is None:
logging.warning(f"Fetcher.fetch_url({url}): No soup or no soup.html")
return None
# Obtain the metadata and the content from the resulting soup
metadata = cast(Any, helper).get_metadata(soup) if helper else None
content: Optional[Tag] = (
cast(Any, helper).get_content(soup) if helper else soup.html.body
)
return (metadata, content)
@classmethod
def fetch_url_html(
cls, url: str, enclosing_session: Optional[Session] = None
) -> Tuple[Optional[str], Any, Optional[ModuleType]]:
"""Fetch a URL using the scraping mechanism, returning
a tuple (html, metadata, helper) or None if error"""
html_doc, metadata, helper = None, None, None
with SessionContext(enclosing_session) as session:
html_doc: Optional[str] = None
helper = cls.helper_for(session, url)
if helper is None or not hasattr(helper, "fetch_url"):
# Do a straight HTTP fetch
html_doc = cls.raw_fetch_url(url)
else:
# Hand off to the helper
html_doc = cast(Any, helper).fetch_url(url)
if not html_doc:
return (None, None, None)
# Parse the HTML
soup = Fetcher.make_soup(html_doc, helper)
if soup is None:
logging.warning(f"Fetcher.fetch_url_html({url}): No soup")
return (None, None, None)
# Obtain the metadata from the resulting soup
metadata = cast(Any, helper).get_metadata(soup) if helper else None
return (html_doc, metadata, helper)