aboutsummaryrefslogtreecommitdiffstats
path: root/python/persist_tool.py
diff options
context:
space:
mode:
authorBryan Newbold <bnewbold@archive.org>2020-03-19 16:10:40 -0700
committerBryan Newbold <bnewbold@archive.org>2020-03-19 16:10:42 -0700
commit88f337f2cc40824ed3eaf32b1fec17c3b053bfdf (patch)
treeae7ae1a02906adf663098dc4e7762279d5ac2ac8 /python/persist_tool.py
parente21fac21cc5a4267357a499f75f048ee5fd38ddb (diff)
downloadsandcrawler-88f337f2cc40824ed3eaf32b1fec17c3b053bfdf.tar.gz
sandcrawler-88f337f2cc40824ed3eaf32b1fec17c3b053bfdf.zip
persist grobid: add option to skip S3 upload
Motivation for this is that current S3 target (minio) is overloaded, with too many files on a single partition (80 million+). Going to look in to seaweedfs and other options, but for now stopping minio persist. Data is all stored in kafka anyways.
Diffstat (limited to 'python/persist_tool.py')
-rwxr-xr-xpython/persist_tool.py4
1 files changed, 4 insertions, 0 deletions
diff --git a/python/persist_tool.py b/python/persist_tool.py
index 19e6dd7..869af06 100755
--- a/python/persist_tool.py
+++ b/python/persist_tool.py
@@ -41,6 +41,7 @@ def run_grobid(args):
s3_access_key=args.s3_access_key,
s3_secret_key=args.s3_secret_key,
s3_only=args.s3_only,
+ db_only=args.db_only,
)
pusher = JsonLinePusher(
worker,
@@ -135,6 +136,9 @@ def main():
sub_grobid.add_argument('--s3-only',
action='store_true',
help="only upload TEI-XML to S3 (don't write to database)")
+ sub_grobid.add_argument('--db-only',
+ action='store_true',
+ help="only write status to sandcrawler-db (don't save TEI-XML to S3)")
sub_grobid_disk = subparsers.add_parser('grobid-disk',
help="dump GRBOID output to (local) files on disk")