| Commit message (Collapse) | Author | Age | Files | Lines |
... | |
| |
| |
| |
| | |
Also fix a spurious typo.
|
| | |
|
| | |
|
| | |
|
| |
| |
| |
| | |
As a form of documentation
|
| |
| |
| |
| | |
Based on ingest-file-results importer
|
| | |
|
|/
|
|
|
| |
For use with bots that don't have admin privileges, or where human
follow-up review is desired.
|
|\
| |
| |
| |
| | |
container-ingest tool
See merge request webgroup/fatcat!8
|
| |
| |
| |
| | |
Caught by Martin in review; Thanks!
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
--fatcat-api-url is clearer than --host-url
remove unimplemented --debug (copy/paste from webface argparse)
use formater which will display 'default' parameters with --help
Thanks to Martin for pointing out the later, which i've always wanted!
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
This gets rid of some mess error handling code by properly configuring
the elasticsearch client to just not clean up scroll iterators when
accessing the public (prod or qa) search interfaces.
Leaving the scroll state around isn't ideal, so we still delete them if
possible (eg, connecting directly to elasticsearch).
Thanks to Martin for pointing out this solution in review.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
The intent of this tool is to make it easy to enque ingest requests into
kafka, to be processed by a worker pool and eventually end up inserted
into fatcat (for ingest hits that pass various checks).
As a specific example use-case, we have pretty good coverage of eLife (a
prominent OA publisher), but have missed some publications in the past,
and have a large gap for the year 2019:
https://fatcat.wiki/container/en4qj5ijrbf5djxx7p5zzpjyoq/coverage
This tool would make it trivial to enqueue all the missing releases to
be crawled.
Future variants on this tool could query for, eg, long-tail OA works.
|
| | |
|
| | |
|
| |
| |
| |
| |
| | |
These are low-level and high-level (respectively)
client wrappers for elasticsearch
|
|/
|
|
|
|
|
|
|
|
|
|
|
|
| |
Use --fatcat-api-url instead of (ambiguous) --host-url for commands that
aren't deployed/running via systemd.
TODO: update the other --host-url usage, and either roll-out change
consistently or support the old arg as an alias during cut-over
Use argparse.ArgumentDefaultsHelpFormatter (thanks Martin!)
Add help messages for all sub-commands, both as documentation and as a
way to get argparse to print available commands in a more readable
format.
|
|
|
|
|
|
| |
This was causing 5xx errors in production and qa. Eg, at:
https://qa.fatcat.wiki/release/aaaaaaaaaaaaarceaaaaaaaaai/history
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
| |
producer creation/configuration should be happening in __init__() time,
not 'daily' call.
This specific refactor motivated by mocking out the producer in unit
tests.
|
| |
|
| |
|
|
|
|
|
|
| |
The more complete fix is to actually render the JATS to HTML and display
that. This is just to fix a nit with the most common case of XML tags in
abstracts.
|
|
|
|
|
| |
- allow overriding source filter whitelist (common case for CLI use)
- fix editgroup description env variable pass-through
|
|
|
|
|
|
|
|
| |
I thought this would filter for metadata updates to an existing DOI, but
actually "updates" are a type of DOI (eg, a retraction).
TODO: handle 'updates' field. Should both do a lookup and set work_ident
appropriately, and store in crossref-specific metadata.
|
|
|
|
|
|
|
| |
Test in previous commit.
This fixes a user-reported 500 error when creating a file with
SHA1/SHA256/MD5 hashes in upper-case.
|
| |
|
| |
|
| |
|
|
|
|
| |
As opposed to sandcrawler-bot
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Initially was going to create a new worker to consume from the release
update channel, but couldn't get the edit context ("is this a new
release, or update to an existing") from that context.
Currently there is a flag in source code to control whether we only do
OA releases or all releases. Starting with OA only to start slow, but
should probably default to all, and make this a config flag. Should
probably also have a config flag to control this entire feature.
Tested locally in dev.
|
| |
|
|\
| |
| |
| |
| | |
Add basic pagination to search results
See merge request webgroup/fatcat!4
|
| | |
|
| |
| |
| |
| | |
> "Showing top " -> "Showing first "
|
| |
| |
| |
| |
| |
| |
| |
| | |
As per [this
issue](https://github.com/Semantic-Org/Semantic-UI/issues/1885#issuecomment-77619519),
text colors are not supported in semantic ui.
To not move text too much, gray out inactive links.
|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Two new macros:
* top_results(found)
* bottom_results(found)
wip: move pagination into macro
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
The "deep paging problem" imposes some limit, which currently is a
hardcoded default value, `deep_page_limit=2000` in `do_search`.
Elasticsearch can be configured, too:
> Note that from + size can not be more than the index.max_result_window
index setting, which defaults to 10,000.
-- https://www.elastic.co/guide/en/elasticsearch/reference/current/search-request-body.html#request-body-search-from-size
|
| |
| |
| |
| |
| | |
Caught one of these in sentry. Probably due to a crawler? Or typing
gibberish in the token form.
|
|\ \
| | |
| | |
| | |
| | | |
mention elasticsearch empty index setup
See merge request webgroup/fatcat!3
|
| |/
| |
| |
| |
| |
| | |
When setting up with the defaults, all works fine, except that the web
search will try to access a local elasticsearch. Mention in README, how
to create empty indices.
|