blob: a5c16b4aa5a0e49cb7019ac9a1e210e99dda908b (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
|
SHELL = /bin/bash
.SHELLFLAGS = -o pipefail -c
TODAY ?= $(shell date --iso --utc)
.PHONY: help
help: ## Print info about all commands
@echo "Commands:"
@echo
@grep -E '^[a-zA-Z0-9_-]+:.*?## .*$$' $(MAKEFILE_LIST) | awk 'BEGIN {FS = ":.*?## "}; {printf " \033[01;32m%-20s\033[0m %s\n", $$1, $$2}'
.PHONY: dep
dep: ## Install dependencies using pipenv
pipenv install --dev
.PHONY: lint
lint: ## Run lints (eg, flake8, mypy)
pipenv run flake8 fatcat_scholar/ tests/ --exit-zero
pipenv run mypy fatcat_scholar/ tests/ --ignore-missing-imports
#pipenv run pytype fatcat_scholar/
.PHONY: fmt
fmt: ## Run code formating on all source code
pipenv run black fatcat_scholar/ tests/
.PHONY: test
test: lint ## Run all tests and lints
PIPENV_DONT_LOAD_ENV=1 ENV_FOR_DYNACONF=test pipenv run pytest
.PHONY: coverage
coverage: lint ## Run all tests with coverage
PIPENV_DONT_LOAD_ENV=1 ENV_FOR_DYNACONF=test pipenv run pytest --cov
.PHONY: dev
dev: ## Run web service locally, with reloading
ENV_FOR_DYNACONF=development pipenv run uvicorn fatcat_scholar.web:app --reload --port 9819
.PHONY: dev-qa
dev-qa: ## Run web service locally, with reloading, but point search queries to QA search index
ENV_FOR_DYNACONF=qa pipenv run uvicorn fatcat_scholar.web:app --reload --port 9819
.PHONY: run
run: ## Run web service under gunicorn
pipenv run gunicorn fatcat_scholar.web:app -w 4 -k uvicorn.workers.UvicornWorker
.PHONY: fetch-works
fetch-works: ## Fetches some works from any release .json in the data dir
cat data/release_*.json | jq . -c | pipenv run python -m fatcat_scholar.work_pipeline run_releases | pv -l > data/work_intermediate.json
.PHONY: fetch-sim
fetch-sim: ## Fetches some SIM pages
pipenv run python -m fatcat_scholar.sim_pipeline run_issue_db --limit 500 | pv -l > data/sim_intermediate.json
.PHONY: dev-index
dev-index: ## Delete/Create DEV elasticsearch fulltext index locally
http delete ":9200/dev_scholar_fulltext_v01" && true
http put ":9200/dev_scholar_fulltext_v01?include_type_name=true" < schema/scholar_fulltext.v01.json
http put ":9200/dev_scholar_fulltext_v01/_alias/dev_scholar_fulltext"
cat data/sim_intermediate.json data/work_intermediate.json | pipenv run python -m fatcat_scholar.transform run_transform | esbulk -verbose -size 200 -id key -w 4 -index dev_scholar_fulltext_v01 -type _doc
.PHONY: update-i18n
update-i18n: ## Re-extract and compile translation files
pipenv run pybabel extract -F extra/i18n/babel.cfg -o extra/i18n/web_interface.pot fatcat_scholar/
pipenv run pybabel update -i extra/i18n/web_interface.pot -d fatcat_scholar/translations
pipenv run pybabel compile -d fatcat_scholar/translations
data/$(TODAY)/sim_collections.tsv:
mkdir -p data/$(TODAY)
pipenv run ia search "collection:periodicals collection:sim_microfilm mediatype:collection" --itemlist | rg "^pub_" > $@.wip
mv $@.wip $@
data/$(TODAY)/sim_items.tsv:
mkdir -p data/$(TODAY)
pipenv run ia search "collection:periodicals collection:sim_microfilm mediatype:texts" --itemlist | rg "^sim_" > $@.wip
mv $@.wip $@
data/$(TODAY)/sim_collections.json: data/$(TODAY)/sim_collections.tsv
cat data/$(TODAY)/sim_collections.tsv | pipenv run parallel -j4 ia metadata {} | jq . -c | pv -l > $@.wip
mv $@.wip $@
data/$(TODAY)/sim_items.json: data/$(TODAY)/sim_items.tsv
cat data/$(TODAY)/sim_items.tsv | pipenv run parallel -j8 ia metadata {} | jq -c 'del(.histograms, .rotations)' | pv -l > $@.wip
mv $@.wip $@
data/$(TODAY)/issue_db.sqlite: data/$(TODAY)/sim_collections.json data/$(TODAY)/sim_items.json
pipenv run python -m fatcat_scholar.issue_db --db-file $@.wip init_db
cat data/$(TODAY)/sim_collections.json | pv -l | pipenv run python -m fatcat_scholar.issue_db --db-file $@.wip load_pubs -
cat data/$(TODAY)/sim_items.json | pv -l | pipenv run python -m fatcat_scholar.issue_db --db-file $@.wip load_issues -
pipenv run python -m fatcat_scholar.issue_db --db-file $@.wip load_counts
mv $@.wip $@
data/issue_db.sqlite: data/$(TODAY)/issue_db.sqlite
cp data/$(TODAY)/issue_db.sqlite data/issue_db.sqlite
.PHONY: issue-db
issue-db: data/issue_db.sqlite ## Build SIM issue database with today's metadata, then move to default location
|