Title: Bibliographic Metadata Dumps Author: bnewbold Date: 2017-06-07 Tags: tech, archive, scholar Status: draft # TODO: # - does BASE link to fulltext PDFs? is that helpful? # - can we actually get academia.edu and researchgate.net papers? maybe? I've recently been lucky enough to start working on a new big project at the [Internet Archive][]: collecting, indexing, and expanding access to research publications and datasets in the open world. This is perhaps *the* original goal of networked information technology, and thanks to a decade of hard work by the Open Access movement it feels like inertia [is building][nature-elsevier] towards this one small piece of "universal access to all knowledge". [Internet Archive]: https://archive.org [nature-elsevier]: http://www.nature.com/news/scientists-in-germany-peru-and-taiwan-to-lose-access-to-elsevier-journals-1.21223 This is a snapshot-in-time look at "what's already out there" regarding indexes of scholarly papers and books (aka, "things that get cited"). There are a ton of resources out there, and many of them are just re-aggregating or building on top of each other. Here's a table of index-only resources for papers. These are databases or corpuses of metadata that might include links/URLs to full text, but don't seem to host fulltext copies themselves:
What Record Count (millions) Notes
Total digital English language papers 114 estimated[0], 2014
Total open access 27 estimated[0], 2014. Meaning "available somewhere"? MS academic had 35 million.
Number of DOIs 143 Global; includes non-journals.
CrossRef DOIs 88 Primary registrar for journals/paper in western world
BASE Search 109 Data from OAI-PMH
Google Scholar 100 "records", not URLs
Web of Science 90 proprietary; 1 billion citation graph
Scopus 55 proprietary/Elsevier
PubMed 26 Only half (13mil) have abstract or link to fulltext
CORE 24
Semantic Scholar 10 to 20 Sometimes mirror fulltext?
OpenCitations 5 Paper entries; Spring 2017
dblp 3.7 computer science bibliography; Spring 2017
A big open question to me is how many pre-digital scholarly articles there are which have not been digitized or assigned DOI numbers. Eg, how good is JSTOR coverage? I'm unsure how to even compute this number. And here are full-text collections of papers (which also include metadata):
What Fulltext Count (millions) Notes
Sci-Hub/scimag 62 one-file-per-DOI, 2017
CiteSeerX 6 (2010; presumably many more now?). Crawled from the web
CORE 4 Extracted fulltext, not PDF? Complete "gold" OA?
PubMed Central 4 Open Access. 2017
OSF Preprints (COS) 2 2017
Internet Archive 1.5 "Clean" mirrored items in Journal collections; we probably have far more
arxiv.org 1.2 physics+math. articles, not files, 2017
JSTOR Total 10 mostly locked down. includes books, grey lit
JSTOR Early Articles 0.5 open access subset
biorxiv.org 0.01 2017
Numbers aside, here are the useful resources to build on top of: **CrossRef** is the primary **DOI** registrar in the western (english speaking world). They are a non-profit, one of only a dozen or so DOI registrars; almost all scholarly publishers go through them. They provide some basic metadata (title, authors, publication), and have excellent data access: bulk datasets, a query API, and a streaming update API. This is a good, authoritative foundation for building indexes. China, Korea, and Japan have their own DOI registries, and published datasets end up in DataCite instead of CrossRef. Other holes in DOI coverage are "grey literature" (unpublished or informally published documents, like government reports or technical memos), documents pre-2000 with absentee publishers, and books (only a small fraction of books/chapters have DOIs). Publishers and repositories seem to be pretty good about providing **OAI-PMH** API access to their metadata and records (and sometimes fulltext). Directories make it possible to look up thousands of API endpoints. **BASE** seems to be the best aggregation of all this metadata, and some projects build on top of BASE (eg, oaDOI). **CORE** finds all of it's fulltext this way. It's not clear if BASE is a good place to pull bulk metadata from; they seem to re-index from scratch occasionally. **oaDOI** and **dissem.in** are services that provide an API and search interface over metadata and point to Open Access copies of the results. **PubMed** (index) and **PubMed Central** (fulltext) are large and well maintained. There are Pubmed records and identifiers ("PMID") going far back in history, though only for medical texts (there is increasing contemporary coversage out of medicine/biology, but only very recently). Annual and daily database dumps are available, so a good resource to pull from. **CiteSeerX** has been crawling the web for PDFs for a long time. Other than **Google Scholar** and maybe the **Internet Archive** I think they do the most serious paper crawling, though many folks do smaller or one-off crawls. They are academic/non-profit and are willing to share metadata and their collected papers; their systems are documented and open-source. Metadata and citations are extracted from PDFs themselves. They have collaborated with the Microsoft Research and the Allen Institute; I suspect they provided most or all content for **Semantic Scholar** and **Microsoft Academic Knowledge** (the later now defunct). NB: there are some interesting per-domain crawl statistics [available](http://csxcrawlweb01.ist.psu.edu//), though half-broken. It's worth noting that there is probably a lot of redundancy between **pre-prints** and the final published papers, even though semantically most people would consider them versions or editions of the same paper, not totally distinct works. This might inflate both the record counts and the DOI counts. A large number of other resources are not listed because they are very subject-specific or relatively small. They may or may not be worth pursuing, depending on how redundant they are with the larger resources. Eg, CogPrints (cognative science, ~thousands of fulltext), MathSciNet (proprietary math bibliogrpahy, ERIC (educational resources and grey lit), paperity.org (similar to CORE), etc. *Note: We don't do a very good job promoting it, but as of June 2017 The Internet Archive is hiring! In particular we're looking for an all-around web designer and a project manager for an existing 5 person python-web-app team. Check out those and more on our [jobs page](https://archive.org/about/jobs.php)* [0]: "The Number of Scholarly Documents on the Public Web", PLoS One, 1994, Khabsa and Giles. http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0093949