| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153 |
- __package__ = 'archivebox.extractors'
- import re
- from html.parser import HTMLParser
- from pathlib import Path
- from typing import Optional
- from ..index.schema import Link, ArchiveResult, ArchiveOutput, ArchiveError
- from ..util import (
- enforce_types,
- download_url,
- htmldecode,
- )
- from ..config import (
- TIMEOUT,
- CHECK_SSL_VALIDITY,
- SAVE_TITLE,
- CURL_BINARY,
- CURL_ARGS,
- CURL_VERSION,
- CURL_USER_AGENT,
- )
- from ..logging_util import TimedProgress
- HTML_TITLE_REGEX = re.compile(
- r'<title.*?>' # start matching text after <title> tag
- r'([^<>]+)', # get everything up to these symbols
- re.IGNORECASE | re.MULTILINE | re.DOTALL | re.UNICODE,
- )
- class TitleParser(HTMLParser):
- def __init__(self, *args, **kwargs):
- super().__init__(*args, **kwargs)
- self.title_tag = ""
- self.title_og = ""
- self.inside_title_tag = False
- @property
- def title(self):
- return self.title_tag or self.title_og or None
- def handle_starttag(self, tag, attrs):
- if tag.lower() == "title" and not self.title_tag:
- self.inside_title_tag = True
- elif tag.lower() == "meta" and not self.title_og:
- attrs = dict(attrs)
- if attrs.get("property") == "og:title" and attrs.get("content"):
- self.title_og = attrs.get("content")
- def handle_data(self, data):
- if self.inside_title_tag and data:
- self.title_tag += data.strip()
-
- def handle_endtag(self, tag):
- if tag.lower() == "title":
- self.inside_title_tag = False
- @enforce_types
- def get_html(link: Link, path: Path, timeout: int=TIMEOUT) -> str:
- """
- Try to find wget, singlefile and then dom files.
- If none is found, download the url again.
- """
- canonical = link.canonical_outputs()
- abs_path = path.absolute()
- # prefer chrome-generated DOM dump to singlefile as singlefile output often includes HUGE url(data:image/...base64) strings that crash parsers
- sources = [canonical["dom_path"], canonical["singlefile_path"], canonical["wget_path"]]
- document = None
- for source in sources:
- try:
- with open(abs_path / source, "r", encoding="utf-8") as f:
- document = f.read()
- break
- except (FileNotFoundError, TypeError):
- continue
- if document is None:
- return download_url(link.url, timeout=timeout)
- else:
- return document
- @enforce_types
- def should_save_title(link: Link, out_dir: Optional[str]=None, overwrite: Optional[bool]=False) -> bool:
- # if link already has valid title, skip it
- if not overwrite and link.title and not link.title.lower().startswith('http'):
- return False
- return SAVE_TITLE
- def extract_title_with_regex(html):
- match = re.search(HTML_TITLE_REGEX, html)
- output = htmldecode(match.group(1).strip()) if match else None
- return output
- @enforce_types
- def save_title(link: Link, out_dir: Optional[Path]=None, timeout: int=TIMEOUT) -> ArchiveResult:
- """try to guess the page's title from its content"""
- from core.models import Snapshot
- output: ArchiveOutput = None
- cmd = [
- CURL_BINARY,
- *CURL_ARGS,
- '--max-time', str(timeout),
- *(['--user-agent', '{}'.format(CURL_USER_AGENT)] if CURL_USER_AGENT else []),
- *([] if CHECK_SSL_VALIDITY else ['--insecure']),
- link.url,
- ]
- status = 'succeeded'
- timer = TimedProgress(timeout, prefix=' ')
- try:
- html = get_html(link, out_dir, timeout=timeout)
- try:
- # try using relatively strict html parser first
- parser = TitleParser()
- parser.feed(html)
- output = parser.title
- if output is None:
- raise
- except Exception:
- # fallback to regex that can handle broken/malformed html
- output = extract_title_with_regex(html)
-
- # if title is better than the one in the db, update db with new title
- if isinstance(output, str) and output:
- if not link.title or len(output) >= len(link.title):
- Snapshot.objects.filter(url=link.url,
- timestamp=link.timestamp)\
- .update(title=output)
- else:
- # if no content was returned, dont save a title (because it might be a temporary error)
- if not html:
- raise ArchiveError('Unable to detect page title')
- # output = html[:128] # use first bit of content as the title
- output = link.base_url # use the filename as the title (better UX)
- except Exception as err:
- status = 'failed'
- output = err
- finally:
- timer.end()
- return ArchiveResult(
- cmd=cmd,
- pwd=str(out_dir),
- cmd_version=CURL_VERSION,
- output=output,
- status=status,
- **timer.stats,
- )
|