import json import datetime import requests from flask import abort, flash from fatcat_covid19.webface import app def do_search(index, request, limit=30, offset=0, deep_page_limit=2000): # Sanity checks if limit > 100: limit = 100 if offset < 0: offset = 0 if offset > deep_page_limit: # Avoid deep paging problem. offset = deep_page_limit request["size"] = int(limit) request["from"] = int(offset) # print(request) resp = requests.get("%s/%s/_search" % (app.config['ELASTICSEARCH_BACKEND'], index), json=request) if resp.status_code == 400: print("elasticsearch 400: " + str(resp.content)) #flash("Search query failed to parse; you might need to use quotes.

{}".format(resp.content)) abort(resp.status_code) elif resp.status_code != 200: print("elasticsearch non-200 status code: " + str(resp.status_code)) print(resp.content) abort(resp.status_code) content = resp.json() #print(json.dumps(content, indent=2)) results = [] for h in content['hits']['hits']: r = h['_source'] r['_highlights'] = [] highlights = h.get('highlight', {}) for k in highlights: r['_highlights'] += highlights[k] results.append(r) print(json.dumps(results, indent=2)) for h in results: # Handle surrogate strings that elasticsearch returns sometimes, # probably due to mangled data processing in some pipeline. # "Crimes against Unicode"; production workaround for key in h: if type(h[key]) is str: h[key] = h[key].encode('utf8', 'ignore').decode('utf8') return {"count_returned": len(results), "count_found": content['hits']['total'], "results": results, "offset": offset, "deep_page_limit": deep_page_limit} def do_fulltext_search(q, limit=30, offset=0): #print("Search hit: " + q) if limit > 100: # Sanity check limit = 100 # Convert raw DOIs to DOI queries if len(q.split()) == 1 and q.startswith("10.") and q.count("/") >= 1: q = 'doi:"{}"'.format(q) search_request = { "query": { "query_string": { "query": q, "default_operator": "AND", "analyze_wildcard": True, "lenient": True, "fields": [ "everything", "abstract", "fulltext.body", "fulltext.annex", ], }, }, "highlight" : { "number_of_fragments" : 3, "fragment_size" : 150, "fields" : { "abstract": { }, "fulltext.body": { }, "fulltext.annex": { }, #"everything": { "number_of_fragments" : 3 }, #"fulltext.abstract": { "number_of_fragments" : 3 }, #"fulltext.body": { "number_of_fragments" : 3 }, #"fulltext.annex": { "number_of_fragments" : 3 }, }, }, } resp = do_search(app.config['ELASTICSEARCH_FULLTEXT_INDEX'], search_request, offset=offset) for h in resp['results']: # Ensure 'contrib_names' is a list, not a single string if type(h['contrib_names']) is not list: h['contrib_names'] = [h['contrib_names'], ] h['contrib_names'] = [name.encode('utf8', 'ignore').decode('utf8') for name in h['contrib_names']] resp["query"] = { "q": q } resp["limit"] = limit return resp