aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorBryan Newbold <bnewbold@archive.org>2020-04-15 12:39:40 -0700
committerBryan Newbold <bnewbold@archive.org>2020-04-15 12:39:40 -0700
commit3ca6f8cf8f99739af5a830af0ddc021bb69a7706 (patch)
treececaed09063ffb1a48fd7cec716e691e57ff956f
parentb1a68eefd96025ffe81a4e8e05b23dd71b9a6d92 (diff)
downloadsandcrawler-3ca6f8cf8f99739af5a830af0ddc021bb69a7706.tar.gz
sandcrawler-3ca6f8cf8f99739af5a830af0ddc021bb69a7706.zip
2020-04 datacite ingest (in progress)
-rw-r--r--notes/ingest/2020-04-07_datacite.md18
1 files changed, 18 insertions, 0 deletions
diff --git a/notes/ingest/2020-04-07_datacite.md b/notes/ingest/2020-04-07_datacite.md
new file mode 100644
index 0000000..b0217f0
--- /dev/null
+++ b/notes/ingest/2020-04-07_datacite.md
@@ -0,0 +1,18 @@
+
+After the broad datacite crawl, want to ingest paper PDFs into fatcat. But many
+of the DOIs are for, eg, datasets, and don't want to waste time on those.
+
+Instead of using full ingest request file from the crawl, will generate a new
+ingest request file using `fatcat_ingest.py` and set that up for bulk crawling.
+
+## Generate Requests
+
+ ./fatcat_ingest.py --allow-non-oa --release-types article-journal,paper-conference,article,report,thesis,book,chapter query "doi_registrar:datacite" | pv -l > /srv/fatcat/snapshots/datacite_papers_20200407.ingest_request.json
+ => Expecting 8905453 release objects in search queries
+ => 8.91M 11:49:50 [ 209 /s]
+ => Counter({'elasticsearch_release': 8905453, 'ingest_request': 8905453, 'estimate': 8905453})
+
+## Bulk Ingest
+
+ cat /srv/fatcat/snapshots/datacite_papers_20200407.ingest_request.json | rg -v "\\\\" | jq . -c | kafkacat -P -b wbgrp-svc263.us.archive.org -t sandcrawler-prod.ingest-file-requests-bulk -p -1
+