aboutsummaryrefslogtreecommitdiffstats
path: root/python
diff options
context:
space:
mode:
authorBryan Newbold <bnewbold@robocracy.org>2021-11-23 19:29:49 -0800
committerBryan Newbold <bnewbold@robocracy.org>2021-11-23 19:32:21 -0800
commit112c41a1157862d2c8f758eac685b0b26c921797 (patch)
treeeb4511e6a8087ce30dcbeba74de6d65c2b636999 /python
parent8080eef139b5dcf6201e4f27076a879d0df20096 (diff)
downloadfatcat-112c41a1157862d2c8f758eac685b0b26c921797.tar.gz
fatcat-112c41a1157862d2c8f758eac685b0b26c921797.zip
file merges: fixes from testing in QA
Diffstat (limited to 'python')
-rw-r--r--python/fatcat_tools/mergers/files.py37
1 files changed, 23 insertions, 14 deletions
diff --git a/python/fatcat_tools/mergers/files.py b/python/fatcat_tools/mergers/files.py
index 32c7fcb6..30b31330 100644
--- a/python/fatcat_tools/mergers/files.py
+++ b/python/fatcat_tools/mergers/files.py
@@ -24,9 +24,12 @@ class FileMerger(EntityMerger):
def __init__(self, api: fatcat_openapi_client.ApiClient, **kwargs) -> None:
- eg_desc = kwargs.get("editgroup_description", "Automated merge of file entities")
- eg_extra = kwargs.get("editgroup_extra", dict())
+ eg_desc = (
+ kwargs.pop("editgroup_description", None) or "Automated merge of file entities"
+ )
+ eg_extra = kwargs.pop("editgroup_extra", dict())
eg_extra["agent"] = eg_extra.get("agent", "fatcat_tools.FileMerger")
+ self.dry_run_mode: bool = eg_extra.get("dry_run_mode", False)
super().__init__(api, editgroup_description=eg_desc, editgroup_extra=eg_extra, **kwargs)
self.entity_type_name = "file"
@@ -130,7 +133,7 @@ class FileMerger(EntityMerger):
if entities[ident].state != "active":
self.counts["skip-not-active-entity"] += 1
return 0
- if getattr(entities[ident].ext_ids, evidence["extid_type"]) != evidence["extid"]:
+ if getattr(entities[ident], evidence["extid_type"]) != evidence["extid"]:
self.counts["skip-extid-mismatch"] += 1
return 0
@@ -138,7 +141,6 @@ class FileMerger(EntityMerger):
primary_id = self.choose_primary_file(list(entities.values()))
dupe_ids = [d for d in dupe_ids if d != primary_id]
- # ensure primary is not in dupes
assert primary_id not in dupe_ids
primary = entities[primary_id]
@@ -146,25 +148,32 @@ class FileMerger(EntityMerger):
for other_id in dupe_ids:
other = entities[other_id]
primary_updated = self.merge_file_metadata_from(primary, other) or primary_updated
- self.api.update_file(
- eg_id,
- other.ident,
- FileEntity(
- redirect=primary.ident,
- edit_extra=evidence,
- ),
- )
+ if not self.dry_run_mode:
+ self.api.update_file(
+ eg_id,
+ other.ident,
+ FileEntity(
+ redirect=primary.ident,
+ edit_extra=evidence,
+ ),
+ )
updated_entities += 1
if primary_updated:
- self.api.update_file(eg_id, primary.ident, primary)
+ if not self.dry_run_mode:
+ self.api.update_file(eg_id, primary.ident, primary)
updated_entities += 1
return updated_entities
def run_merge_files(args: argparse.Namespace) -> None:
- em = FileMerger(args.api, edit_batch_size=args.batch_size, dry_run_mode=args.dry_run)
+ em = FileMerger(
+ args.api,
+ edit_batch_size=args.batch_size,
+ dry_run_mode=args.dry_run,
+ editgroup_description=args.editgroup_description_override,
+ )
JsonLinePusher(em, args.json_file).run()