aboutsummaryrefslogtreecommitdiffstats
path: root/proposals
diff options
context:
space:
mode:
authorBryan Newbold <bnewbold@archive.org>2020-11-03 11:29:22 -0800
committerBryan Newbold <bnewbold@archive.org>2020-11-03 11:29:22 -0800
commit40e2e20378fb06e43cc93f67427f865a0de0a692 (patch)
treeaee7738016ee3afa7dc744942376caefce876bfe /proposals
parentbd9075adef2733df046621ef799c3b29e00fac57 (diff)
downloadsandcrawler-40e2e20378fb06e43cc93f67427f865a0de0a692.tar.gz
sandcrawler-40e2e20378fb06e43cc93f67427f865a0de0a692.zip
commit WIP HTML ingest proposal
Diffstat (limited to 'proposals')
-rw-r--r--proposals/20201026_html_ingest.md97
1 files changed, 97 insertions, 0 deletions
diff --git a/proposals/20201026_html_ingest.md b/proposals/20201026_html_ingest.md
new file mode 100644
index 0000000..90bc6e5
--- /dev/null
+++ b/proposals/20201026_html_ingest.md
@@ -0,0 +1,97 @@
+
+status: wip
+
+HTML Ingest Pipeline
+========================
+
+Basic goal: given an ingest request of type 'html', output an object (JSON)
+which could be imported into fatcat.
+
+Should work with things like (scholarly) blog posts, micropubs, registrations,
+protocols. Doesn't need to work with everything to start. "Platform" sites
+(like youtube, figshare, etc) will probably be a different ingest worker.
+
+A current unknown is what the expected size of this metadata is. Both in number
+of documents and amount of metadata per document.
+
+Example HTML articles to start testing:
+
+- complex distill article: <https://distill.pub/2020/bayesian-optimization/>
+- old HTML journal: <http://web.archive.org/web/20081120141926fw_/http://www.mundanebehavior.org/issues/v5n1/rosen.htm>
+- NIH pub: <https://www.nlm.nih.gov/pubs/techbull/ja02/ja02_locatorplus_merge.html>
+- first mondays (OJS): <https://firstmonday.org/ojs/index.php/fm/article/view/10274/9729>
+- d-lib: <http://www.dlib.org/dlib/july17/williams/07williams.html>
+
+## Ingest Process
+
+Follow base URL to terminal document, which is assumed to be a status=200 HTML document.
+
+Verify that terminal document is fulltext. Extract both metadata and fulltext.
+
+Extract list of sub-resources. Filter out unwanted (eg favicon, analytics,
+unnecessary), apply a sanity limit. Convert to fully qualified URLs. For each
+sub-resource, fetch down to the terminal resource, and compute hashes/metadata.
+
+TODO:
+- will probably want to parallelize sub-resource fetching. async?
+- behavior when failure fetching sub-resources
+
+
+## Ingest Result Schema
+
+JSON should
+
+The minimum that could be persisted for later table lookup are:
+
+- (url, datetime): CDX table
+- sha1hex: `file_meta` table
+
+Probably makes most sense to have all this end up in a large JSON object though.
+
+
+## New SQL Tables
+
+`html_meta`
+ surt,
+ timestamp (str?)
+ primary key: (surt, timestamp)
+ sha1hex (indexed)
+ updated
+ status
+ has_teixml
+ biblio (JSON)
+ resources (JSON)
+
+Also writes to `ingest_file_result`, `file_meta`, and `cdx`, all only for the base HTML document.
+
+## Fatcat API Wants
+
+Would be nice to have lookup by SURT+timestamp, and/or by sha1hex of terminal base file.
+
+`hide` option for cdx rows; also for fileset equivalent.
+
+## New Workers
+
+Could reuse existing worker, have code branch depending on type of ingest.
+
+ingest file worker
+ => same as existing worker, because could be calling SPN
+
+persist result
+ => same as existing worker
+
+persist html text
+ => talks to seaweedfs
+
+
+## New Kafka Topics
+
+HTML ingest result topic (webcapture-ish)
+
+sandcrawler-ENV.html-teixml
+ JSON
+ same as other fulltext topics
+
+## TODO
+
+- refactor ingest worker to be more general