aboutsummaryrefslogtreecommitdiffstats
path: root/python
Commit message (Collapse)AuthorAgeFilesLines
* fixes for large GROBID result skipBryan Newbold2019-12-021-2/+2
|
* count empty blobs as 'failed' instead of crashingBryan Newbold2019-12-011-1/+2
| | | | Might be better to record an artificial kafka response instead?
* cleanup unused importBryan Newbold2019-12-011-1/+0
|
* filter out very large GROBID XML bodiesBryan Newbold2019-12-011-0/+6
| | | | | | | | | | This is to prevent Kafka MSG_SIZE_TOO_LARGE publish errors. We should probably bump this in the future. Open problems: hand-coding this size number isn't good, need to update in two places. Shouldn't filter out for non-Kafka sinks. Might still exist a corner-case where JSON encoded XML is larger than XML character string, due to encoding (eg, for unicode characters).
* CI: make some jobs manualBryan Newbold2019-11-151-0/+2
| | | | | Scalding test is broken :( But we aren't even using that code much these days.
* handle wayback fetch redirect loop in ingest codeBryan Newbold2019-11-141-2/+5
|
* bump kafka max poll interval for consumersBryan Newbold2019-11-141-2/+2
| | | | | The ingest worker keeps timing out at just over 5 minutes, so bump it just a bit.
* handle WaybackError during ingestBryan Newbold2019-11-141-0/+4
|
* handle SPNv1 redirect loopBryan Newbold2019-11-141-0/+2
|
* handle SPNv2 polling timeoutBryan Newbold2019-11-141-6/+10
|
* update ingest-file batch size to 1Bryan Newbold2019-11-142-4/+4
| | | | | | | | Was defaulting to 100, which I think was resulting in lots of consumer group timeouts, resulting in UNKNOWN_MEMBER_ID errors. Will probably switch back to batches of 10 or so, but multi-processing or some other concurrent dispatch/processing.
* start of hrmars.com ingest supportBryan Newbold2019-11-142-2/+7
|
* treat failure to get terminal capture as a SavePageNowErrorBryan Newbold2019-11-131-1/+1
|
* citation_pdf_url with host-relative URLsBryan Newbold2019-11-131-1/+3
|
* status_forcelist is on session, not requestBryan Newbold2019-11-131-2/+2
|
* handle SPNv1 remote server HTTP status codes betterBryan Newbold2019-11-131-8/+15
|
* grobid2json: make lang detection flexibleBryan Newbold2019-11-131-1/+2
|
* handle requests (http) redirect loop from waybackBryan Newbold2019-11-131-1/+4
|
* handle wayback client return status correctlyBryan Newbold2019-11-131-2/+2
|
* allow way more errors in SPN pathBryan Newbold2019-11-131-2/+11
|
* clean up redirect-following CDX API pathBryan Newbold2019-11-131-8/+15
|
* fix lint errorsBryan Newbold2019-11-132-6/+11
|
* improve ingest worker remote failure behaviorBryan Newbold2019-11-131-5/+12
|
* have SPN client differentiate between SPN and remote errorsBryan Newbold2019-11-132-3/+11
| | | | | | | | This is only a partial implementation. The requests client will still make way too many SPN requests trying to figure out if this is a real error or not (eg, if remote was a 502, we'll retry many times). We may just want to switch to SPNv2 for everything.
* correct ingest-file consumer groupBryan Newbold2019-11-131-1/+1
|
* add basic sandcrawler worker (kafka)Bryan Newbold2019-11-131-0/+74
|
* note that kafka_grobid.py is deprecatedBryan Newbold2019-11-131-0/+3
|
* rename FileIngestWorkerBryan Newbold2019-11-133-10/+16
|
* refactor consume_topic name out of make_kafka_consumer()Bryan Newbold2019-11-131-5/+5
| | | | Best to do this in wrapping code for full flexibility.
* more progress on file ingestBryan Newbold2019-11-134-17/+75
|
* much progress on file ingest pathBryan Newbold2019-10-226-335/+338
|
* remove spurious debug print from grobid2jsonBryan Newbold2019-10-221-1/+1
|
* we do actually want consolidateHeader=2, not 1Bryan Newbold2019-10-042-4/+4
|
* remove any trailing newlineBryan Newbold2019-10-041-2/+2
|
* grobid: consolidateHeaders typoBryan Newbold2019-10-041-1/+1
|
* grobid_tool: don't wrap multiprocess if we don't need toBryan Newbold2019-10-041-2/+4
|
* disable citation consolidation by defaultBryan Newbold2019-10-041-1/+1
| | | | | | | with this consolidation enabled, the glutton_fatcat elasticsearch server was totally pegged over 90% CPU with only 10 PDF worker threads; the glutton load seemed to be the bottleneck even for this low degree of parallelism. Disabled for now, will debug with GROBID/glutton folks.
* grobid-output-pg, not grobid-output-jsonBryan Newbold2019-10-041-4/+2
|
* grobid_tool: don't always insert multi wrapperBryan Newbold2019-10-041-6/+13
|
* grobid2json: language_codeBryan Newbold2019-10-042-1/+7
|
* fix GROBID POST flagsBryan Newbold2019-10-041-1/+3
|
* workers: better generic batch-size arg handlingBryan Newbold2019-10-031-0/+6
|
* handle GROBID fetch empty blob conditionBryan Newbold2019-10-031-1/+2
|
* grobid_affiliations fix from prod, and usage exampleBryan Newbold2019-10-021-0/+5
|
* deliver_dumpgrobid_to_s3: typo fix from old prodBryan Newbold2019-10-021-3/+4
|
* grobid affiliation extractor (script)Bryan Newbold2019-10-021-0/+47
|
* python tests for pusher classesBryan Newbold2019-10-022-0/+28
|
* have grobidworker error status indicate issues instead of bailingBryan Newbold2019-10-021-4/+13
|
* grobid_tool.py example usage in docstringBryan Newbold2019-10-021-0/+6
|
* add tests for affiliation extractionBryan Newbold2019-10-022-1/+25
|