diff options
author | Bryan Newbold <bnewbold@robocracy.org> | 2020-07-01 16:33:45 -0700 |
---|---|---|
committer | Bryan Newbold <bnewbold@robocracy.org> | 2020-07-01 16:33:45 -0700 |
commit | 7ed9ae436813e1b8e275a1d43ab2ea6f05a429f4 (patch) | |
tree | 49a5b63fabb239514b15ac5ce3b625c2af7b1973 | |
parent | ca8fa64c1590a43b1e92fd8898275625d083451a (diff) | |
download | fatcat-7ed9ae436813e1b8e275a1d43ab2ea6f05a429f4.tar.gz fatcat-7ed9ae436813e1b8e275a1d43ab2ea6f05a429f4.zip |
commit old README about bulk downloads
-rw-r--r-- | extra/bulk_download/README.md | 40 |
1 files changed, 40 insertions, 0 deletions
diff --git a/extra/bulk_download/README.md b/extra/bulk_download/README.md new file mode 100644 index 00000000..83b92fd9 --- /dev/null +++ b/extra/bulk_download/README.md @@ -0,0 +1,40 @@ + +## Download Fatcat Fulltext from web.archive.org in Bulk + +These quick-and-dirty directions use UNIX utilities to download from the +Internet Archive (either in the wayback machine or archive.org). To make a +proper mirror (eg, for research or preservation use), you would want to verify +hashes (fixity), handle additional retries, and handle files which are not +preserved in Internet Archive, retain linkage between files and fatcat +identifiers, etc. + +You can download a file entity dump from the most recent "Bulk Metadata Export" +item from the [snapshots and exports collection](https://archive.org/details/fatcat_snapshots_and_exports?sort=-publicdate). + +Create a TSV file containing the SHA1 and a single URL for each file +entity: + + zcat file_export.json.gz \ + | grep '"application/pdf"' + | jq -cr '.sha1 as $sha1 | .urls | map(select((.url | startswith("https://web.archive.org/web/")) or (.url | startswith("https://archive.org/download/")))) | select(. != []) | [$sha1, .[0].url] | @tsv' \ + > fatcat_files_sha1_iaurl.tsv + +Then use the GNU `parallel` command to call `curl` in parallel to fetch files. +The `-j` argument controls parallelism. Please don't create exessive load on +Internet Archive infrastructure by downloading with too many threads. 10 +parallel threads is a decent amount of load. + + cat fatcat_files_sha1_iaurl.tsv \ + | awk '{print "curl -Lfs --write-out \"%{http_code}\\t" $1 "\\t%{url_effective}\\n\" \"" $2 "\" -o ", $1 ".pdf"}' \ + | parallel --bar -j4 {} \ + > fetch_status.log + +This will write out a status log containing the HTTP status code, expected file +SHA1, and attempted URL. You can check for errors (and potentially try) with: + + grep -v "^200" fetch_status.log + +Or, count status codes: + + cut -f1 fetch_status.log | sort | uniq -c | sort -nr + |