aboutsummaryrefslogtreecommitdiffstats
path: root/python/fatcat_tools/importers/dblp_container.py
blob: 603a627105ce6079f4e6038ad0915813b2dcf238 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
"""
Importer for DBLP container-level (journal/conference/series) metadata,
pre-scraped in to JSON from HTML pages.
"""

import sys  # noqa: F401

import fatcat_openapi_client

from fatcat_tools.importers.common import EntityImporter
from fatcat_tools.normal import clean_str


class DblpContainerImporter(EntityImporter):
    def __init__(
        self, api, issn_map_file, dblp_container_map_file, dblp_container_map_output, **kwargs
    ):

        eg_desc = kwargs.get(
            "editgroup_description",
            "Automated import of container-level metadata scraped from dblp HTML",
        )
        eg_extra = kwargs.get("editgroup_extra", dict())
        eg_extra["agent"] = eg_extra.get("agent", "fatcat_tools.DblpContainerImporter")
        super().__init__(api, editgroup_description=eg_desc, editgroup_extra=eg_extra, **kwargs)

        self.dblp_container_map_output = dblp_container_map_output
        self.read_dblp_container_map_file(dblp_container_map_file)
        self.read_issn_map_file(issn_map_file)
        print("\t".join(["dblp_prefix", "container_id"]), file=self.dblp_container_map_output)

    def read_dblp_container_map_file(self, dblp_container_map_file) -> None:
        self._dblp_container_map = dict()
        print("Loading existing dblp prefix container map file...", file=sys.stderr)
        for line in dblp_container_map_file:
            if line.startswith("dblp_prefix") or len(line) == 0:
                continue
            (prefix, container_id) = line.split()[0:2]
            assert len(container_id) == 26
            self._dblp_container_map[prefix] = container_id
            print("\t".join([prefix, container_id]), file=self.dblp_container_map_output)
        print(
            "Got {} existing dblp container mappings.".format(len(self._dblp_container_map)),
            file=sys.stderr,
        )

    def lookup_dblp_prefix(self, prefix):
        if not prefix:
            return None
        return self._dblp_container_map.get(prefix)

    def want(self, raw_record):
        return True

    def parse_record(self, row):
        """
        row is a python dict (parsed from JSON).

        returns a ContainerEntity (or None if invalid or couldn't parse)
        """

        dblp_prefix = row.get("key") or row.get("dblp_prefix")
        assert dblp_prefix
        assert row["title"]

        container_type = None
        if dblp_prefix.startswith("conf/"):
            container_type = "conference-series"
        elif dblp_prefix.startswith("journals/"):
            container_type = "journal"
        elif dblp_prefix.startswith("series/"):
            container_type = "book-series"

        issnl = None
        for issn in row.get("issns", []):
            issnl = self.issn2issnl(issn)
            if issnl:
                break

        extra = {
            "dblp": {
                "prefix": dblp_prefix,
            },
        }

        if row.get("homepage_url"):
            extra["urls"] = [row["homepage_url"]]

        if row.get("acronym"):
            extra["acronym"] = row["acronym"]

        ce = fatcat_openapi_client.ContainerEntity(
            name=clean_str(row["title"]),
            container_type=container_type,
            issnl=issnl,
            wikidata_qid=row.get("wikidata_qid"),
            extra=extra,
        )
        return ce

    def try_update(self, ce):

        dblp_prefix = ce.extra["dblp"]["prefix"]
        existing = None
        existing_container_id = self.lookup_dblp_prefix(dblp_prefix)
        if existing_container_id:
            existing = self.api.get_container(existing_container_id)
        if not existing and ce.issnl:
            # check if existing by ISSN-L
            try:
                existing = self.api.lookup_container(issnl=ce.issnl)
            except fatcat_openapi_client.rest.ApiException as err:
                if err.status != 404:
                    raise err
        if not existing and ce.wikidata_qid:
            try:
                existing = self.api.lookup_container(wikidata_qid=ce.wikidata_qid)
            except fatcat_openapi_client.rest.ApiException as err:
                if err.status != 404:
                    raise err

        # TODO: plan to add a fuzzy match check here

        if not existing:
            return True

        if existing:
            self.counts["exists"] += 1
            print(
                "\t".join([ce.extra["dblp"]["prefix"], existing.ident]),
                file=self.dblp_container_map_output,
            )
            return False

        # shouldn't get here
        raise NotImplementedError()

    def insert_batch(self, batch):
        """
        Because we want to print a prefix/container_id match for each row, we
        require a special batch insert method
        """
        eg = self.api.create_container_auto_batch(
            fatcat_openapi_client.ContainerAutoBatch(
                editgroup=fatcat_openapi_client.Editgroup(
                    description=self.editgroup_description, extra=self.editgroup_extra
                ),
                entity_list=batch,
            )
        )
        for c_edit in eg.edits.containers:
            c = self.api.get_container(c_edit.ident)
            print(
                "\t".join([c.extra["dblp"]["prefix"], c.ident]),
                file=self.dblp_container_map_output,
            )