aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--python/sandcrawler/__init__.py1
-rwxr-xr-xpython/sandcrawler_worker.py44
2 files changed, 37 insertions, 8 deletions
diff --git a/python/sandcrawler/__init__.py b/python/sandcrawler/__init__.py
index c9cc0c9..236570e 100644
--- a/python/sandcrawler/__init__.py
+++ b/python/sandcrawler/__init__.py
@@ -4,4 +4,5 @@ from .misc import gen_file_metadata, b32_hex, parse_cdx_line, parse_cdx_datetime
from .workers import KafkaSink, KafkaGrobidSink, JsonLinePusher, CdxLinePusher, CdxLinePusher, KafkaJsonPusher, BlackholeSink, ZipfilePusher, MultiprocessWrapper
from .ia import WaybackClient, WaybackError, CdxApiClient, CdxApiError, SavePageNowClient, SavePageNowError, SavePageNowRemoteError
from .ingest import IngestFileWorker
+from .persist import PersistCdxWorker, PersistIngestFileResultWorker, PersistGrobidWorker, PersistGrobidDiskWorker
diff --git a/python/sandcrawler_worker.py b/python/sandcrawler_worker.py
index 895a5b9..c926911 100755
--- a/python/sandcrawler_worker.py
+++ b/python/sandcrawler_worker.py
@@ -6,6 +6,7 @@ Outputs might either be pushed back into Kafka, or directly into sandcrawler-db
or minio.
"""
+import os
import sys
import argparse
import datetime
@@ -28,9 +29,16 @@ def run_grobid_extract(args):
consume_topic=consume_topic, group="grobid-extract")
pusher.run()
-def run_grobid_persist(args):
+def run_persist_grobid(args):
consume_topic = "sandcrawler-{}.grobid-output-pg".format(args.env)
- worker = PersistGrobidWorker()
+ worker = PersistGrobidWorker(
+ db_url=args.db_url,
+ s3_url=args.s3_url,
+ s3_bucket=args.s3_bucket,
+ s3_access_key=args.s3_access_key,
+ s3_secret_key=args.s3_secret_key,
+ s3_only=args.s3_only,
+ )
pusher = KafkaJsonPusher(worker=worker, kafka_hosts=args.kafka_hosts,
consume_topic=consume_topic, group="grobid-persist")
pusher.run()
@@ -45,9 +53,11 @@ def run_ingest_file(args):
consume_topic=consume_topic, group="ingest-file", batch_size=1)
pusher.run()
-def run_ingest_file_persist(args):
+def run_persist_ingest_file(args):
consume_topic = "sandcrawler-{}.ingest-file-results".format(args.env)
- worker = PersistIngestFileResultWorker()
+ worker = PersistIngestFileResultWorker(
+ db_url=args.db_url,
+ )
pusher = KafkaJsonPusher(worker=worker, kafka_hosts=args.kafka_hosts,
consume_topic=consume_topic, group="ingest-persist")
pusher.run()
@@ -64,23 +74,41 @@ def main():
parser.add_argument('--grobid-host',
default="http://grobid.qa.fatcat.wiki",
help="GROBID API host/port")
+ parser.add_argument('--db-url',
+ help="postgresql database connection string",
+ default="postgres:///sandcrawler")
+ parser.add_argument('--s3-url',
+ help="S3 (minio) backend URL",
+ default="localhost:9000")
+ parser.add_argument('--s3-access-key',
+ help="S3 (minio) credential",
+ default=os.environ.get('MINIO_ACCESS_KEY'))
+ parser.add_argument('--s3-secret-key',
+ help="S3 (minio) credential",
+ default=os.environ.get('MINIO_SECRET_KEY'))
+ parser.add_argument('--s3-bucket',
+ help="S3 (minio) bucket to persist into",
+ default="sandcrawler-dev")
subparsers = parser.add_subparsers()
sub_grobid_extract = subparsers.add_parser('grobid-extract',
help="daemon that consumes CDX JSON objects from Kafka, extracts, pushes to Kafka")
sub_grobid_extract.set_defaults(func=run_grobid_extract)
- sub_grobid_persist = subparsers.add_parser('grobid-persist',
+ sub_persist_grobid = subparsers.add_parser('persist-grobid',
help="daemon that consumes GROBID output from Kafka and pushes to minio and postgres")
- sub_grobid_persist.set_defaults(func=run_grobid_persist)
+ sub_persist_grobid.add_argument('--s3-only',
+ action='store_true',
+ help="only upload TEI-XML to S3 (don't write to database)")
+ sub_persist_grobid.set_defaults(func=run_persist_grobid)
sub_ingest_file = subparsers.add_parser('ingest-file',
help="daemon that consumes requests from Kafka, ingests, pushes results to Kafka")
sub_ingest_file.set_defaults(func=run_ingest_file)
- sub_ingest_file_persist = subparsers.add_parser('ingest-file-persist',
+ sub_persist_ingest_file = subparsers.add_parser('persist-ingest-file',
help="daemon that consumes ingest-file output from Kafka and pushes to postgres")
- sub_ingest_file_persist.set_defaults(func=run_ingest_file_persist)
+ sub_persist_ingest_file.set_defaults(func=run_persist_ingest_file)
args = parser.parse_args()
if not args.__dict__.get("func"):