import collections from fatcat_client import ReleaseEntity, ApiClient def entity_to_dict(entity): """ Hack to take advantage of the code-generated serialization code """ ac = ApiClient() return ac.sanitize_for_serialization(entity) def entity_from_json(json_str, entity_type): """ Hack to take advantage of the code-generated deserialization code """ ac = ApiClient() thing = collections.namedtuple('Thing', ['data']) thing.data = json_str return ac.deserialize(thing, entity_type) def release_to_elasticsearch(release): """ Converts from an entity model/schema to elasticsearch oriented schema. Returns: dict Raises exception on error (never returns None) """ if release.state in ('redirect', 'deleted'): return dict( ident = release.ident, state = release.state, ) elif release.state != 'active': raise ValueError("Unhandled release state: {}".format(release.state)) # First, the easy ones (direct copy) t = dict( ident = release.ident, state = release.state, revision = release.revision, title = release.title, release_type = release.release_type, release_status = release.release_status, language = release.language, doi = release.doi, pmid = release.pmid, pmcid = release.pmcid, isbn13 = release.isbn13, core_id = release.core_id, wikidata_qid = release.wikidata_qid ) if release.release_date: # .isoformat() results in, eg, '2010-10-22' (YYYY-MM-DD) t['release_date'] = release.release_date.isoformat() if release.release_year is None: t['release_year'] = release.release_date.year if release.release_year is not None: t['release_year'] = release.release_year container = release.container container_is_kept = False if container: t['publisher'] = container.publisher t['container_name'] = container.name t['container_issnl'] = container.issnl container_extra = container.extra if container_extra: t['container_is_oa'] = container_extra.get('is_oa') container_is_kept = container_extra.get('is_kept', False) t['container_is_longtail_oa'] = container_extra.get('is_longtail_oa') else: t['publisher'] = release.publisher files = release.files or [] t['file_count'] = len(files) in_wa = False in_ia = False t['file_pdf_url'] = None for f in files: is_pdf = 'pdf' in (f.mimetype or '') for url in (f.urls or []): if url.rel == 'webarchive': in_wa = True if '//web.archive.org/' in (url.url or '') or '//archive.org/' in (url.url or ''): in_ia = True if is_pdf: t['file_pdf_url'] = url.url if not t['file_pdf_url'] and is_pdf: t['file_pdf_url'] = url.url t['file_in_webarchive'] = in_wa t['file_in_ia'] = in_ia extra = release.extra or dict() if extra: t['in_shadow'] = extra.get('in_shadow') if extra.get('grobid') and extra['grobid'].get('is_longtail_oa'): t['container_is_longtail_oa'] = True t['any_abstract'] = len(release.abstracts) > 0 t['is_kept'] = container_is_kept or extra.get('is_kept', False) t['ref_count'] = len(release.refs or []) t['contrib_count'] = len(release.contribs or []) contrib_names = [] for c in (release.contribs or []): if c.raw_name: contrib_names.append(c.raw_name) t['contrib_names'] = contrib_names return t