diff options
author | Bryan Newbold <bnewbold@robocracy.org> | 2019-09-05 18:49:36 -0700 |
---|---|---|
committer | Bryan Newbold <bnewbold@robocracy.org> | 2019-09-05 18:49:36 -0700 |
commit | 28d2d94a9100c1a809955cde93fdbb7a36263057 (patch) | |
tree | b61675cbd21f28da15dabf4361cf86bad565f1c4 /python/fatcat_tools/importers | |
parent | 070069cb6eb71b92a9c4e46f3d4cfabb67f4eb3f (diff) | |
download | fatcat-28d2d94a9100c1a809955cde93fdbb7a36263057.tar.gz fatcat-28d2d94a9100c1a809955cde93fdbb7a36263057.zip |
refactor all python source for client lib name
Diffstat (limited to 'python/fatcat_tools/importers')
-rw-r--r-- | python/fatcat_tools/importers/arabesque.py | 16 | ||||
-rw-r--r-- | python/fatcat_tools/importers/arxiv.py | 22 | ||||
-rwxr-xr-x | python/fatcat_tools/importers/cdl_dash_dat.py | 6 | ||||
-rw-r--r-- | python/fatcat_tools/importers/chocula.py | 12 | ||||
-rw-r--r-- | python/fatcat_tools/importers/common.py | 6 | ||||
-rw-r--r-- | python/fatcat_tools/importers/crossref.py | 20 | ||||
-rw-r--r-- | python/fatcat_tools/importers/grobid_metadata.py | 24 | ||||
-rw-r--r-- | python/fatcat_tools/importers/jalc.py | 16 | ||||
-rw-r--r-- | python/fatcat_tools/importers/journal_metadata.py | 10 | ||||
-rw-r--r-- | python/fatcat_tools/importers/jstor.py | 18 | ||||
-rw-r--r-- | python/fatcat_tools/importers/matched.py | 18 | ||||
-rw-r--r-- | python/fatcat_tools/importers/orcid.py | 10 | ||||
-rw-r--r-- | python/fatcat_tools/importers/pubmed.py | 32 | ||||
-rwxr-xr-x | python/fatcat_tools/importers/wayback_static.py | 2 |
14 files changed, 106 insertions, 106 deletions
diff --git a/python/fatcat_tools/importers/arabesque.py b/python/fatcat_tools/importers/arabesque.py index e9376d96..65597d5c 100644 --- a/python/fatcat_tools/importers/arabesque.py +++ b/python/fatcat_tools/importers/arabesque.py @@ -4,7 +4,7 @@ import json import base64 import sqlite3 import itertools -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean, make_rel_url, SANE_MAX_RELEASES, SANE_MAX_URLS @@ -105,7 +105,7 @@ class ArabesqueMatchImporter(EntityImporter): # lookup extid try: re = self.api.lookup_release(**{self.extid_type: extid}) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status == 404: # bail on 404 (release not in DB) self.counts['skip-extid-not-found'] += 1 @@ -128,13 +128,13 @@ class ArabesqueMatchImporter(EntityImporter): row['final_url']) urls = [url, ("webarchive", wayback)] - urls = [fatcat_client.FileUrl(rel=rel, url=url) for (rel, url) in urls] + urls = [fatcat_openapi_client.FileUrl(rel=rel, url=url) for (rel, url) in urls] if len(urls) > SANE_MAX_URLS: self.counts['skip-too-many-url'] += 1 return None - fe = fatcat_client.FileEntity( + fe = fatcat_openapi_client.FileEntity( sha1=b32_hex(row['final_sha1']), mimetype=row['final_mimetype'] or self.default_mimetype, release_ids=[re.ident], @@ -147,7 +147,7 @@ class ArabesqueMatchImporter(EntityImporter): existing = None try: existing = self.api.lookup_file(sha1=fe.sha1) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -177,7 +177,7 @@ class ArabesqueMatchImporter(EntityImporter): # merge the existing into this one and update existing.urls = list(set([(u.rel, u.url) for u in fe.urls + existing.urls])) - existing.urls = [fatcat_client.FileUrl(rel=rel, url=url) for (rel, url) in existing.urls] + existing.urls = [fatcat_openapi_client.FileUrl(rel=rel, url=url) for (rel, url) in existing.urls] if len(existing.urls) > SANE_MAX_URLS: self.counts['skip-update-too-many-url'] += 1 return None @@ -192,8 +192,8 @@ class ArabesqueMatchImporter(EntityImporter): return False def insert_batch(self, batch): - self.api.create_file_auto_batch(fatcat_client.FileAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_file_auto_batch(fatcat_openapi_client.FileAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/arxiv.py b/python/fatcat_tools/importers/arxiv.py index 71b2d134..c69ee16a 100644 --- a/python/fatcat_tools/importers/arxiv.py +++ b/python/fatcat_tools/importers/arxiv.py @@ -6,7 +6,7 @@ import datetime from bs4 import BeautifulSoup from pylatexenc.latex2text import LatexNodes2Text -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean from .crossref import lookup_license_slug @@ -120,7 +120,7 @@ class ArxivRawImporter(EntityImporter): doi = None title = latex_to_text(metadata.title.string) authors = parse_arxiv_authors(metadata.authors.string) - contribs = [fatcat_client.ReleaseContrib(index=i, raw_name=a, role='author') for i, a in enumerate(authors)] + contribs = [fatcat_openapi_client.ReleaseContrib(index=i, raw_name=a, role='author') for i, a in enumerate(authors)] lang = "en" # the vast majority in english if metadata.comments and metadata.comments.string: @@ -178,12 +178,12 @@ class ArxivRawImporter(EntityImporter): if '$' in abst or '{' in abst: mime = "application/x-latex" abst_plain = latex_to_text(abst) - abstracts.append(fatcat_client.ReleaseAbstract(content=abst_plain, mimetype="text/plain", lang="en")) + abstracts.append(fatcat_openapi_client.ReleaseAbstract(content=abst_plain, mimetype="text/plain", lang="en")) else: mime = "text/plain" - abstracts.append(fatcat_client.ReleaseAbstract(content=abst, mimetype=mime, lang="en")) + abstracts.append(fatcat_openapi_client.ReleaseAbstract(content=abst, mimetype=mime, lang="en")) if orig: - abstracts.append(fatcat_client.ReleaseAbstract(content=orig, mimetype=mime)) + abstracts.append(fatcat_openapi_client.ReleaseAbstract(content=orig, mimetype=mime)) # indicates that fulltext probably isn't english either if lang == 'en': lang = None @@ -207,7 +207,7 @@ class ArxivRawImporter(EntityImporter): release_date = version.date.string.strip() release_date = datetime.datetime.strptime(release_date, "%a, %d %b %Y %H:%M:%S %Z").date() # TODO: source_type? - versions.append(fatcat_client.ReleaseEntity( + versions.append(fatcat_openapi_client.ReleaseEntity( work_id=None, title=title, #original_title @@ -216,7 +216,7 @@ class ArxivRawImporter(EntityImporter): release_stage='submitted', release_date=release_date.isoformat(), release_year=release_date.year, - ext_ids=fatcat_client.ReleaseExtIds( + ext_ids=fatcat_openapi_client.ReleaseExtIds( arxiv=arxiv_id, ), number=number, @@ -270,7 +270,7 @@ class ArxivRawImporter(EntityImporter): existing_doi = None try: existing = self.api.lookup_release(arxiv=v.ext_ids.arxiv) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -281,7 +281,7 @@ class ArxivRawImporter(EntityImporter): if v.ext_ids.doi: try: existing_doi = self.api.lookup_release(doi=v.ext_ids.doi) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err if existing_doi: @@ -325,8 +325,8 @@ class ArxivRawImporter(EntityImporter): # there is no batch/bezerk mode for arxiv importer, except for testing if self._test_override: for batch in batch_batch: - self.api.create_release_auto_batch(fatcat_client.ReleaseAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_release_auto_batch(fatcat_openapi_client.ReleaseAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/cdl_dash_dat.py b/python/fatcat_tools/importers/cdl_dash_dat.py index aa7fab2b..536c013b 100755 --- a/python/fatcat_tools/importers/cdl_dash_dat.py +++ b/python/fatcat_tools/importers/cdl_dash_dat.py @@ -9,8 +9,8 @@ import hashlib import mimetypes import subprocess -import fatcat_client -from fatcat_client import * +import fatcat_openapi_client +from fatcat_openapi_client import * from .common import clean from .crossref import lookup_license_slug @@ -176,7 +176,7 @@ def auto_cdl_dash_dat(api, dat_path, release_id=None, editgroup_id=None): try: r = api.lookup_release(doi=release.ext_ids.doi) release_id = r.ident - except fatcat_client.rest.ApiException: + except fatcat_openapi_client.rest.ApiException: pass if not release_id: edit = api.create_release(eg.editgroup_id, release) diff --git a/python/fatcat_tools/importers/chocula.py b/python/fatcat_tools/importers/chocula.py index d7044ff4..6915ba98 100644 --- a/python/fatcat_tools/importers/chocula.py +++ b/python/fatcat_tools/importers/chocula.py @@ -2,7 +2,7 @@ import sys import json import itertools -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean @@ -66,7 +66,7 @@ class ChoculaImporter(EntityImporter): elif 'journal ' in name.lower(): container_type = 'journal' - ce = fatcat_client.ContainerEntity( + ce = fatcat_openapi_client.ContainerEntity( issnl=row['issnl'], ident=row['ident'], name=name, @@ -82,7 +82,7 @@ class ChoculaImporter(EntityImporter): if ce.ident: try: existing = self.api.get_container(ce.ident) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err self.counts['exists'] += 1 @@ -97,7 +97,7 @@ class ChoculaImporter(EntityImporter): # check if existing by ISSN-L try: existing = self.api.lookup_container(issnl=ce.issnl) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err if existing: @@ -157,8 +157,8 @@ class ChoculaImporter(EntityImporter): raise NotImplementedError def insert_batch(self, batch): - self.api.create_container_auto_batch(fatcat_client.ContainerAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_container_auto_batch(fatcat_openapi_client.ContainerAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/common.py b/python/fatcat_tools/importers/common.py index 2f677be4..093569e1 100644 --- a/python/fatcat_tools/importers/common.py +++ b/python/fatcat_tools/importers/common.py @@ -13,8 +13,8 @@ import xml.etree.ElementTree as ET import pykafka from bs4 import BeautifulSoup -import fatcat_client -from fatcat_client.rest import ApiException +import fatcat_openapi_client +from fatcat_openapi_client.rest import ApiException DATE_FMT = "%Y-%m-%d" @@ -337,7 +337,7 @@ class EntityImporter: if not self._editgroup_id: eg = self.api.create_editgroup( - fatcat_client.Editgroup( + fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra)) self._editgroup_id = eg.editgroup_id diff --git a/python/fatcat_tools/importers/crossref.py b/python/fatcat_tools/importers/crossref.py index 4074dfb8..faee6aac 100644 --- a/python/fatcat_tools/importers/crossref.py +++ b/python/fatcat_tools/importers/crossref.py @@ -5,7 +5,7 @@ import sqlite3 import datetime import itertools import subprocess -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean @@ -215,7 +215,7 @@ class CrossrefImporter(EntityImporter): extra = None assert ctype in ("author", "editor", "translator") raw_name = clean(raw_name) - contribs.append(fatcat_client.ReleaseContrib( + contribs.append(fatcat_openapi_client.ReleaseContrib( creator_id=creator_id, index=index, raw_name=raw_name, @@ -244,7 +244,7 @@ class CrossrefImporter(EntityImporter): container_name = None if (container_id is None and self.create_containers and (issnl is not None) and container_name): - ce = fatcat_client.ContainerEntity( + ce = fatcat_openapi_client.ContainerEntity( issnl=issnl, publisher=publisher, container_type=self.map_container_type(release_type), @@ -300,7 +300,7 @@ class CrossrefImporter(EntityImporter): ref_extra[k] = clean(rm[k]) if not ref_extra: ref_extra = None - refs.append(fatcat_client.ReleaseRef( + refs.append(fatcat_openapi_client.ReleaseRef( index=i, # doing lookups would be a second import pass target_release_id=None, @@ -316,7 +316,7 @@ class CrossrefImporter(EntityImporter): abstracts = [] abstract = clean(obj.get('abstract')) if abstract and len(abstract) > 10: - abstracts.append(fatcat_client.ReleaseAbstract( + abstracts.append(fatcat_openapi_client.ReleaseAbstract( mimetype="application/xml+jats", content=abstract)) @@ -420,7 +420,7 @@ class CrossrefImporter(EntityImporter): if not extra: extra = None - re = fatcat_client.ReleaseEntity( + re = fatcat_openapi_client.ReleaseEntity( work_id=None, container_id=container_id, title=title, @@ -431,7 +431,7 @@ class CrossrefImporter(EntityImporter): release_date=release_date, release_year=release_year, publisher=publisher, - ext_ids=fatcat_client.ReleaseExtIds( + ext_ids=fatcat_openapi_client.ReleaseExtIds( doi=obj['DOI'].lower(), pmid=extids['pmid'], pmcid=extids['pmcid'], @@ -459,7 +459,7 @@ class CrossrefImporter(EntityImporter): existing = None try: existing = self.api.lookup_release(doi=re.ext_ids.doi) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err # doesn't exist, need to update @@ -474,8 +474,8 @@ class CrossrefImporter(EntityImporter): return True def insert_batch(self, batch): - self.api.create_release_auto_batch(fatcat_client.ReleaseAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_release_auto_batch(fatcat_openapi_client.ReleaseAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/grobid_metadata.py b/python/fatcat_tools/importers/grobid_metadata.py index ea627dd6..2077eae4 100644 --- a/python/fatcat_tools/importers/grobid_metadata.py +++ b/python/fatcat_tools/importers/grobid_metadata.py @@ -4,7 +4,7 @@ import sys import json import base64 import datetime -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean, make_rel_url MAX_ABSTRACT_BYTES=4096 @@ -56,7 +56,7 @@ class GrobidMetadataImporter(EntityImporter): existing = None try: existing = self.api.lookup_file(sha1=fe.sha1) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -82,7 +82,7 @@ class GrobidMetadataImporter(EntityImporter): abstract = obj.get('abstract') if abstract and len(abstract) < MAX_ABSTRACT_BYTES and len(abstract) > 10: - abobj = fatcat_client.ReleaseAbstract( + abobj = fatcat_openapi_client.ReleaseAbstract( mimetype="text/plain", content=clean(obj.get('abstract'))) abstracts = [abobj] @@ -91,7 +91,7 @@ class GrobidMetadataImporter(EntityImporter): contribs = [] for i, a in enumerate(obj.get('authors', [])): - contribs.append(fatcat_client.ReleaseContrib( + contribs.append(fatcat_openapi_client.ReleaseContrib( index=i, raw_name=clean(a['name']), given_name=clean(a.get('given_name')), @@ -116,7 +116,7 @@ class GrobidMetadataImporter(EntityImporter): if not cite_extra: cite_extra = None - refs.append(fatcat_client.ReleaseRef( + refs.append(fatcat_openapi_client.ReleaseRef( key=clean(raw.get('id')), year=year, title=clean(raw['title']), @@ -147,7 +147,7 @@ class GrobidMetadataImporter(EntityImporter): if not title or len(title) < 2: return None - re = fatcat_client.ReleaseEntity( + re = fatcat_openapi_client.ReleaseEntity( title=title, release_type="article-journal", release_date=release_date, @@ -158,7 +158,7 @@ class GrobidMetadataImporter(EntityImporter): volume=clean(obj['journal'].get('volume')), issue=clean(obj['journal'].get('issue')), abstracts=abstracts, - ext_ids=fatcat_client.ReleaseExtIds(), + ext_ids=fatcat_openapi_client.ReleaseExtIds(), extra=extra) return re @@ -166,7 +166,7 @@ class GrobidMetadataImporter(EntityImporter): sha1 = base64.b16encode(base64.b32decode(sha1_key.replace('sha1:', ''))).decode('ascii').lower() - fe = fatcat_client.FileEntity( + fe = fatcat_openapi_client.FileEntity( sha1=sha1, size=int(file_size), mimetype=mimetype, @@ -181,10 +181,10 @@ class GrobidMetadataImporter(EntityImporter): cdx['dt'], original) fe.urls.append( - fatcat_client.FileUrl(url=wayback, rel="webarchive")) + fatcat_openapi_client.FileUrl(url=wayback, rel="webarchive")) original_url = make_rel_url(original, default_link_rel=self.default_link_rel) if original_url is not None: - fe.urls.append(fatcat_client.FileUrl(rel=original_url[0], url=original_url[1])) + fe.urls.append(fatcat_openapi_client.FileUrl(rel=original_url[0], url=original_url[1])) return fe @@ -193,8 +193,8 @@ class GrobidMetadataImporter(EntityImporter): return True def insert_batch(self, batch): - self.api.create_file_auto_batch(fatcat_client.FileAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_file_auto_batch(fatcat_openapi_client.FileAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/jalc.py b/python/fatcat_tools/importers/jalc.py index c1c13bb7..a0e0086b 100644 --- a/python/fatcat_tools/importers/jalc.py +++ b/python/fatcat_tools/importers/jalc.py @@ -7,7 +7,7 @@ import itertools import subprocess from bs4 import BeautifulSoup -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean, is_cjk, DATE_FMT @@ -48,7 +48,7 @@ def parse_jalc_persons(raw_persons): if lang == 'en' and surname and given_name: # english names order is flipped name = "{} {}".format(given_name, surname) - rc = fatcat_client.ReleaseContrib( + rc = fatcat_openapi_client.ReleaseContrib( raw_name=name, surname=surname, given_name=given_name, @@ -269,7 +269,7 @@ class JalcImporter(EntityImporter): # extra: issnp, issne, original_name, languages, country container_extra['country'] = 'jp' container_extra['languages'] = ['ja'] - ce = fatcat_client.ContainerEntity( + ce = fatcat_openapi_client.ContainerEntity( name=container_name, container_type='journal', publisher=publisher, @@ -303,7 +303,7 @@ class JalcImporter(EntityImporter): if not title: return None - re = fatcat_client.ReleaseEntity( + re = fatcat_openapi_client.ReleaseEntity( work_id=None, title=title, original_title=clean(original_title), @@ -311,7 +311,7 @@ class JalcImporter(EntityImporter): release_stage='published', release_date=release_date, release_year=release_year, - ext_ids=fatcat_client.ReleaseExtIds( + ext_ids=fatcat_openapi_client.ReleaseExtIds( doi=doi, pmid=extids['pmid'], pmcid=extids['pmcid'], @@ -338,7 +338,7 @@ class JalcImporter(EntityImporter): existing = None try: existing = self.api.lookup_release(doi=re.ext_ids.doi) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err # doesn't exist, need to insert @@ -353,8 +353,8 @@ class JalcImporter(EntityImporter): return True def insert_batch(self, batch): - self.api.create_release_auto_batch(fatcat_client.ReleaseAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_release_auto_batch(fatcat_openapi_client.ReleaseAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/journal_metadata.py b/python/fatcat_tools/importers/journal_metadata.py index d24b4103..d439c80a 100644 --- a/python/fatcat_tools/importers/journal_metadata.py +++ b/python/fatcat_tools/importers/journal_metadata.py @@ -2,7 +2,7 @@ import sys import json import itertools -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean @@ -93,7 +93,7 @@ class JournalMetadataImporter(EntityImporter): if not name: return None - ce = fatcat_client.ContainerEntity( + ce = fatcat_openapi_client.ContainerEntity( issnl=row['issnl'], container_type=None, # TODO name=name, @@ -107,7 +107,7 @@ class JournalMetadataImporter(EntityImporter): existing = None try: existing = self.api.lookup_container(issnl=ce.issnl) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -133,8 +133,8 @@ class JournalMetadataImporter(EntityImporter): raise NotImplementedError def insert_batch(self, batch): - self.api.create_container_auto_batch(fatcat_client.ContainerAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_container_auto_batch(fatcat_openapi_client.ContainerAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/jstor.py b/python/fatcat_tools/importers/jstor.py index 5016e813..5ff1ecd9 100644 --- a/python/fatcat_tools/importers/jstor.py +++ b/python/fatcat_tools/importers/jstor.py @@ -5,7 +5,7 @@ import datetime import warnings from bs4 import BeautifulSoup -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean, LANG_MAP_MARC from .crossref import CONTAINER_TYPE_MAP @@ -114,7 +114,7 @@ class JstorImporter(EntityImporter): # create container if it doesn't exist if (container_id is None and self.create_containers and (issnl is not None) and journal_title): - ce = fatcat_client.ContainerEntity( + ce = fatcat_openapi_client.ContainerEntity( issnl=issnl, publisher=publisher, container_type=self.map_container_type(release_type), @@ -158,7 +158,7 @@ class JstorImporter(EntityImporter): role = JSTOR_CONTRIB_MAP.get(c.get('contrib-type', 'author')) if not role and c.get('contrib-type'): sys.stderr.write("NOT IN JSTOR_CONTRIB_MAP: {}\n".format(c['contrib-type'])) - contribs.append(fatcat_client.ReleaseContrib( + contribs.append(fatcat_openapi_client.ReleaseContrib( role=role, raw_name=raw_name, given_name=given, @@ -228,7 +228,7 @@ class JstorImporter(EntityImporter): if not extra: extra = None - re = fatcat_client.ReleaseEntity( + re = fatcat_openapi_client.ReleaseEntity( #work_id title=title, #original_title @@ -236,7 +236,7 @@ class JstorImporter(EntityImporter): release_stage=release_stage, release_date=release_date, release_year=release_year, - ext_ids=fatcat_client.ReleaseExtIds( + ext_ids=fatcat_openapi_client.ReleaseExtIds( doi=doi, jstor=jstor_id, ), @@ -270,7 +270,7 @@ class JstorImporter(EntityImporter): existing = None try: existing = self.api.lookup_release(jstor=re.ext_ids.jstor) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -282,7 +282,7 @@ class JstorImporter(EntityImporter): doi = "10.2307/{}".format(re.ext_ids.jstor) try: existing = self.api.lookup_release(doi=doi) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -309,8 +309,8 @@ class JstorImporter(EntityImporter): return True def insert_batch(self, batch): - self.api.create_release_auto_batch(fatcat_client.ReleaseAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_release_auto_batch(fatcat_openapi_client.ReleaseAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/matched.py b/python/fatcat_tools/importers/matched.py index a94584cd..dbb78ec9 100644 --- a/python/fatcat_tools/importers/matched.py +++ b/python/fatcat_tools/importers/matched.py @@ -3,7 +3,7 @@ import sys import json import sqlite3 import itertools -import fatcat_client +import fatcat_openapi_client from fatcat_tools.normal import * from .common import EntityImporter, clean, make_rel_url, SANE_MAX_RELEASES, SANE_MAX_URLS @@ -61,7 +61,7 @@ class MatchedImporter(EntityImporter): return None try: re = self.api.lookup_release(doi=doi) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err re = None @@ -77,7 +77,7 @@ class MatchedImporter(EntityImporter): if extid: try: re = self.api.lookup_release(**{extid_type: extid}) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err re = None @@ -110,7 +110,7 @@ class MatchedImporter(EntityImporter): url = make_rel_url(original, default_link_rel=self.default_link_rel) if url != None: urls.add(url) - urls = [fatcat_client.FileUrl(rel=rel, url=url) for (rel, url) in urls] + urls = [fatcat_openapi_client.FileUrl(rel=rel, url=url) for (rel, url) in urls] if len(urls) == 0: self.counts['skip-no-urls'] += 1 return None @@ -127,7 +127,7 @@ class MatchedImporter(EntityImporter): if urls[0].url.endswith('.pdf'): mimetype = 'application/pdf' - fe = fatcat_client.FileEntity( + fe = fatcat_openapi_client.FileEntity( md5=obj.get('md5'), sha1=obj['sha1'], sha256=obj.get('sha256'), @@ -143,7 +143,7 @@ class MatchedImporter(EntityImporter): existing = None try: existing = self.api.lookup_file(sha1=fe.sha1) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -183,7 +183,7 @@ class MatchedImporter(EntityImporter): # merge the existing into this one and update existing.urls = list(set([(u.rel, u.url) for u in fe.urls + existing.urls])) - existing.urls = [fatcat_client.FileUrl(rel=rel, url=url) for (rel, url) in existing.urls] + existing.urls = [fatcat_openapi_client.FileUrl(rel=rel, url=url) for (rel, url) in existing.urls] if len(existing.urls) > SANE_MAX_URLS: self.counts['skip-update-too-many-url'] += 1 @@ -203,8 +203,8 @@ class MatchedImporter(EntityImporter): return False def insert_batch(self, batch): - self.api.create_file_auto_batch(fatcat_client.FileAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_file_auto_batch(fatcat_openapi_client.FileAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/orcid.py b/python/fatcat_tools/importers/orcid.py index 0bdd5d19..0a2c8610 100644 --- a/python/fatcat_tools/importers/orcid.py +++ b/python/fatcat_tools/importers/orcid.py @@ -2,7 +2,7 @@ import sys import json import itertools -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean def value_or_none(e): @@ -64,7 +64,7 @@ class OrcidImporter(EntityImporter): if not display: # must have *some* name return None - ce = fatcat_client.CreatorEntity( + ce = fatcat_openapi_client.CreatorEntity( orcid=orcid, given_name=clean(given), surname=clean(sur), @@ -76,7 +76,7 @@ class OrcidImporter(EntityImporter): existing = None try: existing = self.api.lookup_creator(orcid=raw_record.orcid) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -89,8 +89,8 @@ class OrcidImporter(EntityImporter): return True def insert_batch(self, batch): - self.api.create_creator_auto_batch(fatcat_client.CreatorAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_creator_auto_batch(fatcat_openapi_client.CreatorAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/pubmed.py b/python/fatcat_tools/importers/pubmed.py index 4bfbbc79..80cf986c 100644 --- a/python/fatcat_tools/importers/pubmed.py +++ b/python/fatcat_tools/importers/pubmed.py @@ -7,7 +7,7 @@ import warnings from bs4 import BeautifulSoup from bs4.element import NavigableString -import fatcat_client +import fatcat_openapi_client from .common import EntityImporter, clean, LANG_MAP_MARC # from: https://www.ncbi.nlm.nih.gov/books/NBK3827/table/pubmedhelp.T.publication_types/?report=objectonly @@ -508,7 +508,7 @@ class PubmedImporter(EntityImporter): and container_name): # name, type, publisher, issnl # extra: issnp, issne, original_name, languages, country - ce = fatcat_client.ContainerEntity( + ce = fatcat_openapi_client.ContainerEntity( name=container_name, container_type='journal', #NOTE: publisher not included @@ -532,7 +532,7 @@ class PubmedImporter(EntityImporter): primary_abstract = medline.find("Abstract") if primary_abstract and primary_abstract.AbstractText.get('NlmCategory'): joined = "\n".join([m.get_text() for m in primary_abstract.find_all("AbstractText")]) - abst = fatcat_client.ReleaseAbstract( + abst = fatcat_openapi_client.ReleaseAbstract( content=joined, mimetype="text/plain", lang="en", @@ -541,7 +541,7 @@ class PubmedImporter(EntityImporter): abstracts.append(abst) elif primary_abstract: for abstract in primary_abstract.find_all("AbstractText"): - abst = fatcat_client.ReleaseAbstract( + abst = fatcat_openapi_client.ReleaseAbstract( content=abstract.get_text().strip(), mimetype="text/plain", lang="en", @@ -549,7 +549,7 @@ class PubmedImporter(EntityImporter): if abst.content: abstracts.append(abst) if abstract.find('math'): - abst = fatcat_client.ReleaseAbstract( + abst = fatcat_openapi_client.ReleaseAbstract( # strip the <AbstractText> tags content=str(abstract)[14:-15], mimetype="application/mathml+xml", @@ -562,7 +562,7 @@ class PubmedImporter(EntityImporter): lang = "en" if other.get('Language'): lang = LANG_MAP_MARC.get(other['Language']) - abst = fatcat_client.ReleaseAbstract( + abst = fatcat_openapi_client.ReleaseAbstract( content=other.AbstractText.get_text().strip(), mimetype="text/plain", lang=lang, @@ -617,7 +617,7 @@ class PubmedImporter(EntityImporter): if author.find("EqualContrib"): # TODO: schema for this? contrib_extra['equal'] = True - contribs.append(fatcat_client.ReleaseContrib( + contribs.append(fatcat_openapi_client.ReleaseContrib( raw_name=raw_name, given_name=given_name, surname=surname, @@ -628,7 +628,7 @@ class PubmedImporter(EntityImporter): )) if medline.AuthorList['CompleteYN'] == 'N': - contribs.append(fatcat_client.ReleaseContrib(raw_name="et al.")) + contribs.append(fatcat_openapi_client.ReleaseContrib(raw_name="et al.")) for i, contrib in enumerate(contribs): if contrib.raw_name != "et al.": @@ -659,7 +659,7 @@ class PubmedImporter(EntityImporter): ref_extra['unstructured'] = ref_raw.string if not ref_extra: ref_extra = None - refs.append(fatcat_client.ReleaseRef( + refs.append(fatcat_openapi_client.ReleaseRef( target_release_id=ref_release_id, extra=ref_extra, )) @@ -682,7 +682,7 @@ class PubmedImporter(EntityImporter): if not title: return None - re = fatcat_client.ReleaseEntity( + re = fatcat_openapi_client.ReleaseEntity( work_id=None, title=title, original_title=clean(original_title), @@ -691,7 +691,7 @@ class PubmedImporter(EntityImporter): release_date=release_date, release_year=release_year, withdrawn_status=withdrawn_status, - ext_ids=fatcat_client.ReleaseExtIds( + ext_ids=fatcat_openapi_client.ReleaseExtIds( doi=doi, pmid=pmid, pmcid=pmcid, @@ -717,7 +717,7 @@ class PubmedImporter(EntityImporter): existing = None try: existing = self.api.lookup_release(pmid=re.ext_ids.pmid) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err @@ -725,7 +725,7 @@ class PubmedImporter(EntityImporter): if not existing and re.ext_ids.doi: try: existing = self.api.lookup_release(doi=re.ext_ids.doi) - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: if err.status != 404: raise err if existing and existing.ext_ids.pmid and existing.ext_ids.pmid != re.ext_ids.pmid: @@ -751,7 +751,7 @@ class PubmedImporter(EntityImporter): try: self.api.update_release(self.get_editgroup_id(), existing.ident, existing) self.counts['update'] += 1 - except fatcat_client.rest.ApiException as err: + except fatcat_openapi_client.rest.ApiException as err: # there is a code path where we try to update the same release # twice in a row; if that happens, just skip # NOTE: API behavior might change in the future? @@ -765,8 +765,8 @@ class PubmedImporter(EntityImporter): return True def insert_batch(self, batch): - self.api.create_release_auto_batch(fatcat_client.ReleaseAutoBatch( - editgroup=fatcat_client.Editgroup( + self.api.create_release_auto_batch(fatcat_openapi_client.ReleaseAutoBatch( + editgroup=fatcat_openapi_client.Editgroup( description=self.editgroup_description, extra=self.editgroup_extra), entity_list=batch)) diff --git a/python/fatcat_tools/importers/wayback_static.py b/python/fatcat_tools/importers/wayback_static.py index 2125017a..fa0bd4ef 100755 --- a/python/fatcat_tools/importers/wayback_static.py +++ b/python/fatcat_tools/importers/wayback_static.py @@ -16,7 +16,7 @@ import argparse import subprocess from bs4 import BeautifulSoup -from fatcat_client import * +from fatcat_openapi_client import * CDX_API_BASE = "https://web.archive.org/cdx/search/cdx" GWB_URL_BASE = "https://web.archive.org/web" |