aboutsummaryrefslogtreecommitdiffstats
path: root/python/fatcat_worker.py
diff options
context:
space:
mode:
authorBryan Newbold <bnewbold@robocracy.org>2021-11-02 18:14:09 -0700
committerBryan Newbold <bnewbold@robocracy.org>2021-11-02 18:14:09 -0700
commit6464631dbe5c4afeb76f2f3c9d63b89f917c9a3b (patch)
tree633303839cafc7d901cf8565e034542606a5bb27 /python/fatcat_worker.py
parentcdfd6b85b386b7bbf9d5a5179ef26970b6e5a4e7 (diff)
downloadfatcat-6464631dbe5c4afeb76f2f3c9d63b89f917c9a3b.tar.gz
fatcat-6464631dbe5c4afeb76f2f3c9d63b89f917c9a3b.zip
fmt (black): *.py
Diffstat (limited to 'python/fatcat_worker.py')
-rwxr-xr-xpython/fatcat_worker.py139
1 files changed, 92 insertions, 47 deletions
diff --git a/python/fatcat_worker.py b/python/fatcat_worker.py
index 397cf731..b776e0ce 100755
--- a/python/fatcat_worker.py
+++ b/python/fatcat_worker.py
@@ -20,10 +20,12 @@ sentry_client = raven.Client()
def run_changelog(args):
topic = "fatcat-{}.changelog".format(args.env)
- worker = ChangelogWorker(args.api, args.kafka_hosts, topic,
- poll_interval=args.poll_interval)
+ worker = ChangelogWorker(
+ args.api, args.kafka_hosts, topic, poll_interval=args.poll_interval
+ )
worker.run()
+
def run_entity_updates(args):
changelog_topic = "fatcat-{}.changelog".format(args.env)
release_topic = "fatcat-{}.release-updates-v03".format(args.env)
@@ -31,7 +33,9 @@ def run_entity_updates(args):
container_topic = "fatcat-{}.container-updates".format(args.env)
work_ident_topic = "fatcat-{}.work-ident-updates".format(args.env)
ingest_file_request_topic = "sandcrawler-{}.ingest-file-requests-daily".format(args.env)
- worker = EntityUpdatesWorker(args.api, args.kafka_hosts,
+ worker = EntityUpdatesWorker(
+ args.api,
+ args.kafka_hosts,
changelog_topic,
release_topic=release_topic,
file_topic=file_topic,
@@ -41,86 +45,126 @@ def run_entity_updates(args):
)
worker.run()
+
def run_elasticsearch_release(args):
consume_topic = "fatcat-{}.release-updates-v03".format(args.env)
- worker = ElasticsearchReleaseWorker(args.kafka_hosts, consume_topic,
+ worker = ElasticsearchReleaseWorker(
+ args.kafka_hosts,
+ consume_topic,
elasticsearch_backend=args.elasticsearch_backend,
- elasticsearch_index=args.elasticsearch_index)
+ elasticsearch_index=args.elasticsearch_index,
+ )
worker.run()
+
def run_elasticsearch_container(args):
consume_topic = "fatcat-{}.container-updates".format(args.env)
- worker = ElasticsearchContainerWorker(args.kafka_hosts, consume_topic,
+ worker = ElasticsearchContainerWorker(
+ args.kafka_hosts,
+ consume_topic,
query_stats=args.query_stats,
elasticsearch_release_index="fatcat_release",
elasticsearch_backend=args.elasticsearch_backend,
- elasticsearch_index=args.elasticsearch_index)
+ elasticsearch_index=args.elasticsearch_index,
+ )
worker.run()
+
def run_elasticsearch_changelog(args):
consume_topic = "fatcat-{}.changelog".format(args.env)
- worker = ElasticsearchChangelogWorker(args.kafka_hosts, consume_topic,
+ worker = ElasticsearchChangelogWorker(
+ args.kafka_hosts,
+ consume_topic,
elasticsearch_backend=args.elasticsearch_backend,
- elasticsearch_index=args.elasticsearch_index)
+ elasticsearch_index=args.elasticsearch_index,
+ )
worker.run()
+
def main():
- parser = argparse.ArgumentParser(
- formatter_class=argparse.ArgumentDefaultsHelpFormatter)
- parser.add_argument('--api-host-url',
- default="http://localhost:9411/v0",
- help="fatcat API host/port to use")
- parser.add_argument('--kafka-hosts',
+ parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)
+ parser.add_argument(
+ "--api-host-url", default="http://localhost:9411/v0", help="fatcat API host/port to use"
+ )
+ parser.add_argument(
+ "--kafka-hosts",
default="localhost:9092",
- help="list of Kafka brokers (host/port) to use")
- parser.add_argument('--env',
- default="dev",
- help="Kafka topic namespace to use (eg, prod, qa, dev)")
+ help="list of Kafka brokers (host/port) to use",
+ )
+ parser.add_argument(
+ "--env", default="dev", help="Kafka topic namespace to use (eg, prod, qa, dev)"
+ )
subparsers = parser.add_subparsers()
- sub_changelog = subparsers.add_parser('changelog',
- help="poll fatcat API for changelog entries, push to kafka")
+ sub_changelog = subparsers.add_parser(
+ "changelog", help="poll fatcat API for changelog entries, push to kafka"
+ )
sub_changelog.set_defaults(func=run_changelog)
- sub_changelog.add_argument('--poll-interval',
+ sub_changelog.add_argument(
+ "--poll-interval",
help="how long to wait between polling (seconds)",
- default=5.0, type=float)
+ default=5.0,
+ type=float,
+ )
- sub_entity_updates = subparsers.add_parser('entity-updates',
- help="poll kafka for changelog entries; push entity changes to various kafka topics")
+ sub_entity_updates = subparsers.add_parser(
+ "entity-updates",
+ help="poll kafka for changelog entries; push entity changes to various kafka topics",
+ )
sub_entity_updates.set_defaults(func=run_entity_updates)
- sub_elasticsearch_release = subparsers.add_parser('elasticsearch-release',
- help="consume kafka feed of new/updated releases, transform and push to search")
+ sub_elasticsearch_release = subparsers.add_parser(
+ "elasticsearch-release",
+ help="consume kafka feed of new/updated releases, transform and push to search",
+ )
sub_elasticsearch_release.set_defaults(func=run_elasticsearch_release)
- sub_elasticsearch_release.add_argument('--elasticsearch-backend',
+ sub_elasticsearch_release.add_argument(
+ "--elasticsearch-backend",
help="elasticsearch backend to connect to",
- default="http://localhost:9200")
- sub_elasticsearch_release.add_argument('--elasticsearch-index',
+ default="http://localhost:9200",
+ )
+ sub_elasticsearch_release.add_argument(
+ "--elasticsearch-index",
help="elasticsearch index to push into",
- default="fatcat_release_v03")
+ default="fatcat_release_v03",
+ )
- sub_elasticsearch_container = subparsers.add_parser('elasticsearch-container',
- help="consume kafka feed of new/updated containers, transform and push to search")
+ sub_elasticsearch_container = subparsers.add_parser(
+ "elasticsearch-container",
+ help="consume kafka feed of new/updated containers, transform and push to search",
+ )
sub_elasticsearch_container.set_defaults(func=run_elasticsearch_container)
- sub_elasticsearch_container.add_argument('--elasticsearch-backend',
+ sub_elasticsearch_container.add_argument(
+ "--elasticsearch-backend",
help="elasticsearch backend to connect to",
- default="http://localhost:9200")
- sub_elasticsearch_container.add_argument('--elasticsearch-index',
+ default="http://localhost:9200",
+ )
+ sub_elasticsearch_container.add_argument(
+ "--elasticsearch-index",
help="elasticsearch index to push into",
- default="fatcat_container")
- sub_elasticsearch_container.add_argument('--query-stats',
- action='store_true',
- help="whether to query release search index for container stats")
+ default="fatcat_container",
+ )
+ sub_elasticsearch_container.add_argument(
+ "--query-stats",
+ action="store_true",
+ help="whether to query release search index for container stats",
+ )
- sub_elasticsearch_changelog = subparsers.add_parser('elasticsearch-changelog',
- help="consume changelog kafka feed, transform and push to search")
+ sub_elasticsearch_changelog = subparsers.add_parser(
+ "elasticsearch-changelog",
+ help="consume changelog kafka feed, transform and push to search",
+ )
sub_elasticsearch_changelog.set_defaults(func=run_elasticsearch_changelog)
- sub_elasticsearch_changelog.add_argument('--elasticsearch-backend',
+ sub_elasticsearch_changelog.add_argument(
+ "--elasticsearch-backend",
help="elasticsearch backend to connect to",
- default="http://localhost:9200")
- sub_elasticsearch_changelog.add_argument('--elasticsearch-index',
+ default="http://localhost:9200",
+ )
+ sub_elasticsearch_changelog.add_argument(
+ "--elasticsearch-index",
help="elasticsearch index to push into",
- default="fatcat_changelog")
+ default="fatcat_changelog",
+ )
args = parser.parse_args()
if not args.__dict__.get("func"):
@@ -130,5 +174,6 @@ def main():
args.api = public_api(args.api_host_url)
args.func(args)
-if __name__ == '__main__':
+
+if __name__ == "__main__":
main()