aboutsummaryrefslogtreecommitdiffstats
path: root/python/sandcrawler
Commit message (Collapse)AuthorAgeFilesLines
...
* many publisher-specific ingest improvementsBryan Newbold2020-01-101-4/+96
|
* improve ingest robustness (for legacy requests)Bryan Newbold2020-01-101-6/+12
|
* support forwarding url types other than pdf_urlBryan Newbold2020-01-091-4/+5
|
* wayback: datetime mismatch as an errorBryan Newbold2020-01-091-1/+2
|
* fill in more html extraction techniquesBryan Newbold2020-01-091-7/+6
|
* refactor ingest to a loop, allowing multiple hopsBryan Newbold2020-01-091-25/+48
|
* lots of progress on wayback refactoringBryan Newbold2020-01-092-50/+138
| | | | | | - too much to list - canonical flags to control crawling - cdx_to_dict helper
* location comes as a string, not listBryan Newbold2020-01-091-1/+1
|
* fix http/https issue with GlobalWayback libraryBryan Newbold2020-01-091-1/+2
|
* wayback fetch via replay; confirm hashes in crawl_resource()Bryan Newbold2020-01-091-5/+40
|
* wrap up basic (locally testable) ingest refactorBryan Newbold2020-01-092-178/+219
|
* fix grobid tests for new wayback refactorsBryan Newbold2020-01-091-3/+3
|
* more wayback and SPN tests and fixesBryan Newbold2020-01-092-39/+153
|
* refactor CdxApiClient, add testsBryan Newbold2020-01-081-40/+130
| | | | | | - always use auth token and get full CDX rows - simplify to "fetch" (exact url/dt match) and "lookup_best" methods - all redirect stuff will be moved to a higher level
* refactor SavePaperNowClient and add testBryan Newbold2020-01-071-28/+154
| | | | | | - response as a namedtuple - "remote" errors (aka, SPN API was HTTP 200 but returned error) aren't an exception
* remove SPNv1 code pathsBryan Newbold2020-01-072-65/+25
|
* handle grobid2json errors in calling code insteadBryan Newbold2020-01-021-1/+7
|
* db: move duplicate row filtering into DB insert helpersBryan Newbold2020-01-022-15/+26
|
* remove unused filter in grobid workerBryan Newbold2020-01-021-1/+0
|
* fix dict typoBryan Newbold2020-01-021-1/+1
|
* improvements to grobid persist workerBryan Newbold2020-01-021-13/+16
|
* set mimetype when PUT to minioBryan Newbold2020-01-021-0/+4
|
* fix DB import countingBryan Newbold2020-01-021-4/+5
|
* fix small errors found by pylintBryan Newbold2020-01-022-1/+2
|
* fix sandcrawler persist workersBryan Newbold2020-01-021-0/+1
|
* filter ingest results to not have key conflicts within batchBryan Newbold2020-01-021-1/+16
| | | | | This handles a corner case with ON CONFLICT ... DO UPDATE where you can't do multiple such updates in the same batch transaction.
* db: fancy insert/update separation using postgres xmaxBryan Newbold2020-01-022-24/+45
|
* add PersistGrobidDiskWorkerBryan Newbold2020-01-021-0/+33
| | | | To help with making dumps directly from Kafka (eg, for partner delivery)
* flush out minio helper, add to grobid persistBryan Newbold2020-01-022-22/+71
|
* implement counts properly for persist workersBryan Newbold2020-01-021-15/+19
|
* improve DB helpersBryan Newbold2020-01-021-26/+81
| | | | | - return insert/update row counts - implement ON CONFLICT ... DO UPDATE on some tables
* be more parsimonious with GROBID metadataBryan Newbold2020-01-021-2/+4
| | | | | Because these are getting persisted in database (as well as kafka), don't write out empty keys.
* start work on DB connector and minio clientBryan Newbold2020-01-022-0/+200
|
* have JsonLinePusher continue on JSON decode errors (but count)Bryan Newbold2020-01-021-1/+5
|
* start work on persist workers and toolBryan Newbold2020-01-021-0/+223
|
* refactor: use print(..., file=sys.stderr)Bryan Newbold2019-12-183-25/+27
| | | | Should use logging soon, but this seems more idiomatic in the meanwhile.
* fixes for large GROBID result skipBryan Newbold2019-12-021-2/+2
|
* count empty blobs as 'failed' instead of crashingBryan Newbold2019-12-011-1/+2
| | | | Might be better to record an artificial kafka response instead?
* cleanup unused importBryan Newbold2019-12-011-1/+0
|
* filter out very large GROBID XML bodiesBryan Newbold2019-12-011-0/+6
| | | | | | | | | | This is to prevent Kafka MSG_SIZE_TOO_LARGE publish errors. We should probably bump this in the future. Open problems: hand-coding this size number isn't good, need to update in two places. Shouldn't filter out for non-Kafka sinks. Might still exist a corner-case where JSON encoded XML is larger than XML character string, due to encoding (eg, for unicode characters).
* CI: make some jobs manualBryan Newbold2019-11-151-0/+2
| | | | | Scalding test is broken :( But we aren't even using that code much these days.
* handle wayback fetch redirect loop in ingest codeBryan Newbold2019-11-141-2/+5
|
* bump kafka max poll interval for consumersBryan Newbold2019-11-141-2/+2
| | | | | The ingest worker keeps timing out at just over 5 minutes, so bump it just a bit.
* handle WaybackError during ingestBryan Newbold2019-11-141-0/+4
|
* handle SPNv1 redirect loopBryan Newbold2019-11-141-0/+2
|
* handle SPNv2 polling timeoutBryan Newbold2019-11-141-6/+10
|
* update ingest-file batch size to 1Bryan Newbold2019-11-141-3/+3
| | | | | | | | Was defaulting to 100, which I think was resulting in lots of consumer group timeouts, resulting in UNKNOWN_MEMBER_ID errors. Will probably switch back to batches of 10 or so, but multi-processing or some other concurrent dispatch/processing.
* start of hrmars.com ingest supportBryan Newbold2019-11-142-2/+7
|
* treat failure to get terminal capture as a SavePageNowErrorBryan Newbold2019-11-131-1/+1
|
* citation_pdf_url with host-relative URLsBryan Newbold2019-11-131-1/+3
|