|
|
@@ -9,8 +9,9 @@ __package__ = 'archivebox.parsers'
|
|
|
|
|
|
import re
|
|
|
import os
|
|
|
+from io import StringIO
|
|
|
|
|
|
-from typing import Tuple, List
|
|
|
+from typing import IO, Tuple, List
|
|
|
from datetime import datetime
|
|
|
|
|
|
from ..system import atomic_write
|
|
|
@@ -37,15 +38,7 @@ from .generic_rss import parse_generic_rss_export
|
|
|
from .generic_json import parse_generic_json_export
|
|
|
from .generic_txt import parse_generic_txt_export
|
|
|
|
|
|
-
|
|
|
-@enforce_types
|
|
|
-def parse_links(source_file: str) -> Tuple[List[Link], str]:
|
|
|
- """parse a list of URLs with their metadata from an
|
|
|
- RSS feed, bookmarks export, or text file
|
|
|
- """
|
|
|
-
|
|
|
- check_url_parsing_invariants()
|
|
|
- PARSERS = (
|
|
|
+PARSERS = (
|
|
|
# Specialized parsers
|
|
|
('Pocket HTML', parse_pocket_html_export),
|
|
|
('Pinboard RSS', parse_pinboard_rss_export),
|
|
|
@@ -60,30 +53,66 @@ def parse_links(source_file: str) -> Tuple[List[Link], str]:
|
|
|
# Fallback parser
|
|
|
('Plain Text', parse_generic_txt_export),
|
|
|
)
|
|
|
+
|
|
|
+@enforce_types
|
|
|
+def parse_links_memory(urls: List[str]):
|
|
|
+ """
|
|
|
+ parse a list of URLS without touching the filesystem
|
|
|
+ """
|
|
|
+ check_url_parsing_invariants()
|
|
|
+
|
|
|
+ timer = TimedProgress(TIMEOUT * 4)
|
|
|
+ #urls = list(map(lambda x: x + "\n", urls))
|
|
|
+ file = StringIO()
|
|
|
+ file.writelines(urls)
|
|
|
+ file.name = "io_string"
|
|
|
+ output = _parse(file, timer)
|
|
|
+
|
|
|
+ if output is not None:
|
|
|
+ return output
|
|
|
+
|
|
|
+ timer.end()
|
|
|
+ return [], 'Failed to parse'
|
|
|
+
|
|
|
+
|
|
|
+@enforce_types
|
|
|
+def parse_links(source_file: str) -> Tuple[List[Link], str]:
|
|
|
+ """parse a list of URLs with their metadata from an
|
|
|
+ RSS feed, bookmarks export, or text file
|
|
|
+ """
|
|
|
+
|
|
|
+ check_url_parsing_invariants()
|
|
|
+
|
|
|
timer = TimedProgress(TIMEOUT * 4)
|
|
|
with open(source_file, 'r', encoding='utf-8') as file:
|
|
|
- for parser_name, parser_func in PARSERS:
|
|
|
- try:
|
|
|
- links = list(parser_func(file))
|
|
|
- if links:
|
|
|
- timer.end()
|
|
|
- return links, parser_name
|
|
|
- except Exception as err: # noqa
|
|
|
- pass
|
|
|
- # Parsers are tried one by one down the list, and the first one
|
|
|
- # that succeeds is used. To see why a certain parser was not used
|
|
|
- # due to error or format incompatibility, uncomment this line:
|
|
|
- # print('[!] Parser {} failed: {} {}'.format(parser_name, err.__class__.__name__, err))
|
|
|
- # raise
|
|
|
+ output = _parse(file, timer)
|
|
|
+
|
|
|
+ if output is not None:
|
|
|
+ return output
|
|
|
|
|
|
timer.end()
|
|
|
return [], 'Failed to parse'
|
|
|
|
|
|
+def _parse(to_parse: IO[str], timer) -> Tuple[List[Link], str]:
|
|
|
+ for parser_name, parser_func in PARSERS:
|
|
|
+ try:
|
|
|
+ links = list(parser_func(to_parse))
|
|
|
+ if links:
|
|
|
+ timer.end()
|
|
|
+ return links, parser_name
|
|
|
+ except Exception as err: # noqa
|
|
|
+ pass
|
|
|
+ # Parsers are tried one by one down the list, and the first one
|
|
|
+ # that succeeds is used. To see why a certain parser was not used
|
|
|
+ # due to error or format incompatibility, uncomment this line:
|
|
|
+ # print('[!] Parser {} failed: {} {}'.format(parser_name, err.__class__.__name__, err))
|
|
|
+ # raise
|
|
|
+
|
|
|
|
|
|
@enforce_types
|
|
|
def save_text_as_source(raw_text: str, filename: str='{ts}-stdin.txt', out_dir: str=OUTPUT_DIR) -> str:
|
|
|
ts = str(datetime.now().timestamp()).split('.', 1)[0]
|
|
|
- source_path = os.path.join(OUTPUT_DIR, SOURCES_DIR_NAME, filename.format(ts=ts))
|
|
|
+ source_path = os.path.join(out_dir, SOURCES_DIR_NAME, filename.format(ts=ts))
|
|
|
atomic_write(source_path, raw_text)
|
|
|
log_source_saved(source_file=source_path)
|
|
|
return source_path
|