summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorBryan Newbold <bnewbold@archive.org>2020-05-19 11:36:58 -0700
committerBryan Newbold <bnewbold@archive.org>2020-05-19 11:36:58 -0700
commit3ee18580dd108c69c01cdf838a7f1a7d3d181629 (patch)
tree3926506bea54c596e3df61c3c047f2c14ac2c5c1
parentc5faf30df8772af607689a3f65241f34b365325b (diff)
downloadfatcat-scholar-3ee18580dd108c69c01cdf838a7f1a7d3d181629.tar.gz
fatcat-scholar-3ee18580dd108c69c01cdf838a7f1a7d3d181629.zip
WIP on SIM pipeline
-rw-r--r--fatcat_scholar/sim_pipeline.py173
-rw-r--r--fatcat_scholar/work_pipeline.py4
2 files changed, 175 insertions, 2 deletions
diff --git a/fatcat_scholar/sim_pipeline.py b/fatcat_scholar/sim_pipeline.py
new file mode 100644
index 0000000..1dd6476
--- /dev/null
+++ b/fatcat_scholar/sim_pipeline.py
@@ -0,0 +1,173 @@
+
+import os
+import io
+import sys
+import sqlite3
+import argparse
+from pydantic import BaseModel, validator
+from typing import List, Dict, Tuple, Optional, Any, Sequence
+from fatcat_openapi_client import ReleaseEntity, FileEntity
+import internetarchive
+
+from fatcat_scholar.api_entities import *
+from fatcat_scholar.djvu import djvu_extract_leaf_texts
+from fatcat_scholar.sandcrawler import SandcrawlerPostgrestClient, SandcrawlerMinioClient
+from fatcat_scholar.issue_db import IssueDB, SimIssueRow
+from fatcat_scholar.es_transform import es_biblio_from_release, es_release_from_release, DocType
+from fatcat_scholar.work_pipeline import IntermediateBundle
+
+
+def truncate_pub_meta(full: Dict[str, Any]) -> Dict[str, Any]:
+ """
+ Takes a complete archive.org metadata dictionary for a publication
+ collection, and simplifies it by removing fields. Motivation is to make
+ intermediate bundle files smaller.
+ """
+ full.pop('files')
+ if 'ulrichs' in full and full['ulrichs']:
+ full['ulrichs'][0].pop('reviews_mfl')
+ full['ulrichs'][0].pop('editorial_description')
+
+ # these are interesting, but just too long
+ full['ulrichs'][0].pop('online_availability_full_text')
+ full['ulrichs'][0].pop('abstracting_indexing')
+ full['ulrichs'][0].pop('publisher_and_ordering_details')
+ return full
+
+def truncate_issue_meta(full: Dict[str, Any]) -> Dict[str, Any]:
+ """
+ Same as truncate_pub_meta() but for issue item metadata
+ """
+ full.pop('files')
+ return full
+
+class SimPipeline():
+
+ def __init__(self, issue_db: IssueDB):
+ self.issue_db: IssueDB = issue_db
+ self.ia_client = internetarchive.get_session()
+
+ def fetch_sim_issue(self, issue_db_row: Any) -> Optional[Any]:
+ """
+ issue_item
+ pages: str
+ page_texts: list
+ page_number
+ raw_text
+ release_ident: Optional[str]
+ pub_item_metadata
+ issue_item_metadata
+ """
+ # fetch full metadata from API
+ issue_meta = self.ia_client.get_metadata(issue_db_row['issue_item'])
+ pub_meta = self.ia_client.get_metadata(issue_db_row['pub_collection'])
+
+ leaf_index = dict()
+ leaf_list = []
+ assert 'page_numbers' in issue_meta
+ for entry in issue_meta['page_numbers'].get('pages', []):
+ page_num = entry['pageNumber']
+ leaf_index[entry['leafNum']] = page_num
+ if not (page_num and page_num.isdigit()):
+ continue
+ page_num = int(page_num)
+ leaf_list.append(entry['leafNum'])
+
+ if not leaf_list:
+ return None
+
+ page_texts: List[Dict[str, Any]] = []
+ issue_item = self.ia_client.get_item(issue_db_row['issue_item'])
+ issue_item_djvu = issue_item.get_file(issue_db_row['issue_item'] + "_djvu.xml")
+
+ # override 'close()' method so we can still read out contents
+ djvu_bytes = io.BytesIO()
+ djvu_bytes.close = lambda: None # type: ignore
+ assert issue_item_djvu.download(fileobj=djvu_bytes) == True
+ djvu_bytes.seek(0)
+ djvu_xml = io.StringIO(djvu_bytes.read().decode("UTF-8"))
+ del(djvu_bytes)
+
+ leaf_dict = djvu_extract_leaf_texts(djvu_xml)
+
+ for leaf_num, raw_text in leaf_dict.items():
+ page_texts.append(dict(page_num=leaf_index.get(leaf_num), leaf_num=leaf_num, raw_text=raw_text))
+
+ return dict(
+ issue_item=issue_db_row['issue_item'],
+ pages=None,
+ page_texts=page_texts,
+ release_ident=None,
+ pub_item_metadata=truncate_pub_meta(pub_meta),
+ issue_item_metadata=truncate_issue_meta(issue_meta),
+ )
+
+ def run_issue_db(self, limit: int = None):
+ count = 0
+ self.issue_db.db.row_factory = sqlite3.Row
+ cur = self.issue_db.db.cursor()
+ for row in cur.execute('SELECT * FROM sim_issue LEFT JOIN sim_pub ON sim_issue.sim_pubid = sim_pub.sim_pubid WHERE sim_issue.release_count < 3'):
+ full_issue = self.fetch_sim_issue(row)
+ if not full_issue:
+ continue
+ for leaf in full_issue['page_texts']:
+ bundle = IntermediateBundle(
+ doc_type=DocType.sim_page,
+ releases=[],
+ biblio_release_ident=None,
+ grobid_fulltext=None,
+ pdftotext_fulltext=None,
+ sim_fulltext=dict(
+ issue_item=full_issue['issue_item'],
+ pages=str(leaf['page_num']),
+ page_texts=[leaf['raw_text']],
+ release_ident=None,
+ pub_item_metadata=full_issue['pub_item_metadata'],
+ issue_item_metadata=full_issue['issue_item_metadata'],
+ )
+ )
+ print(bundle.json())
+ count += 1
+ if limit is not None and count >= limit:
+ break
+ if limit is not None and count >= limit:
+ break
+
+def main():
+ """
+ Run this command like:
+
+ python -m fatcat_scholar.sim_pipeline
+ """
+
+ parser = argparse.ArgumentParser(
+ formatter_class=argparse.ArgumentDefaultsHelpFormatter)
+ subparsers = parser.add_subparsers()
+
+ parser.add_argument("--issue-db-file",
+ help="sqlite3 database file to open",
+ default='data/issue_db.sqlite',
+ type=str)
+
+ sub = subparsers.add_parser('run_issue_db',
+ help="iterates through entire IssueDB")
+ sub.set_defaults(func='run_issue_db')
+ sub.add_argument("--limit",
+ help="maximum number of pages to index",
+ type=int)
+
+ args = parser.parse_args()
+ if not args.__dict__.get("func"):
+ print("tell me what to do! (try --help)")
+ sys.exit(-1)
+
+ sp = SimPipeline(issue_db=IssueDB(args.issue_db_file))
+
+ if args.func == 'run_issue_db':
+ sp.run_issue_db(limit=args.limit)
+ else:
+ func = getattr(sp, args.func)
+ func()
+
+if __name__=="__main__":
+ main()
diff --git a/fatcat_scholar/work_pipeline.py b/fatcat_scholar/work_pipeline.py
index 161327a..b9dcbe8 100644
--- a/fatcat_scholar/work_pipeline.py
+++ b/fatcat_scholar/work_pipeline.py
@@ -98,7 +98,7 @@ class WorkPipeline():
if not fe.urls:
return None
grobid_meta = self.sandcrawler_db_client.get_grobid(fe.sha1)
- if grobid_meta['status'] != 'success':
+ if not grobid_meta or grobid_meta['status'] != 'success':
return None
#print(grobid_meta)
grobid_xml = self.sandcrawler_s3_client.get_blob(
@@ -290,7 +290,7 @@ def main():
"""
Run this command like:
- python -m fatcat_scholar.release_pipeline
+ python -m fatcat_scholar.work_pipeline
"""
parser = argparse.ArgumentParser(