diff options
author | Bryan Newbold <bnewbold@robocracy.org> | 2019-01-17 13:48:59 -0800 |
---|---|---|
committer | Bryan Newbold <bnewbold@robocracy.org> | 2019-01-17 13:48:59 -0800 |
commit | 8a36af1664f0399f29a57b639dc0c89033578878 (patch) | |
tree | cb35bd1fdf329431820db688dd94b62c842d9ac5 /python/fatcat_tools/importers | |
parent | 3c56a6c8da155f4675e48a6f5fa55b1cf9fb096a (diff) | |
download | fatcat-8a36af1664f0399f29a57b639dc0c89033578878.tar.gz fatcat-8a36af1664f0399f29a57b639dc0c89033578878.zip |
issn => journal_metadata in several places
Diffstat (limited to 'python/fatcat_tools/importers')
-rw-r--r-- | python/fatcat_tools/importers/__init__.py | 2 | ||||
-rw-r--r-- | python/fatcat_tools/importers/journal_metadata.py (renamed from python/fatcat_tools/importers/issn.py) | 10 |
2 files changed, 6 insertions, 6 deletions
diff --git a/python/fatcat_tools/importers/__init__.py b/python/fatcat_tools/importers/__init__.py index e6f081e5..47fc1fd3 100644 --- a/python/fatcat_tools/importers/__init__.py +++ b/python/fatcat_tools/importers/__init__.py @@ -2,6 +2,6 @@ from .common import FatcatImporter, make_kafka_consumer from .crossref import CrossrefImporter, CROSSREF_TYPE_MAP from .grobid_metadata import GrobidMetadataImporter -from .issn import IssnImporter +from .journal_metadata import JournalMetadataImporter from .matched import MatchedImporter from .orcid import OrcidImporter diff --git a/python/fatcat_tools/importers/issn.py b/python/fatcat_tools/importers/journal_metadata.py index f4d525a4..859662ae 100644 --- a/python/fatcat_tools/importers/issn.py +++ b/python/fatcat_tools/importers/journal_metadata.py @@ -25,7 +25,7 @@ def truthy(s): else: return None -class IssnImporter(FatcatImporter): +class JournalMetadataImporter(FatcatImporter): """ Imports journal metadata ("containers") by ISSN, currently from a custom (data munged) .csv file format @@ -40,12 +40,12 @@ class IssnImporter(FatcatImporter): eg_desc = kwargs.get('editgroup_description', "Automated import of container-level metadata, by ISSN. Metadata from Internet Archive munging.") eg_extra = kwargs.get('editgroup_extra', dict()) - eg_extra['agent'] = eg_extra.get('agent', 'fatcat_tools.IssnImporter') + eg_extra['agent'] = eg_extra.get('agent', 'fatcat_tools.JournalMetadataImporter') super().__init__(api, editgroup_description=eg_desc, editgroup_extra=eg_extra) - def parse_issn_row(self, row): + def parse_journal_metadata_row(self, row): """ row is a python dict (parsed from CSV). returns a ContainerEntity (or None if invalid or couldn't parse) @@ -75,14 +75,14 @@ class IssnImporter(FatcatImporter): return ce def create_row(self, row, editgroup_id=None): - ce = self.parse_issn_row(row) + ce = self.parse_journal_metadata_row(row) if ce is not None: self.api.create_container(ce, editgroup_id=editgroup_id) self.counts['insert'] += 1 def create_batch(self, batch): """Reads and processes in batches (not API-call-per-line)""" - objects = [self.parse_issn_row(l) + objects = [self.parse_journal_metadata_row(l) for l in batch if (l is not None)] objects = [o for o in objects if (o is not None)] self.api.create_container_batch(objects, autoaccept="true") |