1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
|
import csv
from typing import Iterable, Optional
from chocula.util import clean_str, clean_issn, parse_lang, parse_country
from chocula.common import DirectoryLoader
from chocula.database import DirectoryInfo, HomepageUrl
class VanishedDisapearedLoader(DirectoryLoader):
"""
Journal-level metadata from the "Vanished Journals" project. This is the
"disapeared" dataset, with many homepage URLs in wayback (web.archive.org).
CSV headers:
- Source
- If Identified by second source
- Journal Name
- ISSN
- E-ISSN
- URL
- Publisher
- <blank>
- Language(s)
- Country
- society_affiliation
- other_sci_affiliation
- Discipline Group
- Start Year
- End Year
- Last Year Online
- Actively Publishing
- Internet Archive Link
"""
source_slug = "vanished_disapeared"
def open_file(self) -> Iterable:
return csv.DictReader(open(self.config.vanished_disapeared.filepath), delimiter=";")
def parse_record(self, record) -> Optional[DirectoryInfo]:
if not record["Journal Name"]:
return None
info = DirectoryInfo(
directory_slug=self.source_slug,
raw_issn=clean_issn(record["ISSN"]),
issne=clean_issn(record["E-ISSN"]),
name=clean_str(record["Journal Name"]),
publisher=clean_str(record["Publisher"]),
langs=[lang for lang in [parse_lang(record["Language(s)"])] if lang],
country=parse_country(record["Country"]),
)
homepage = HomepageUrl.from_url(record["Internet Archive Link"])
if homepage:
info.homepage_urls.append(homepage)
return info
|