Show the extract

This commit is contained in:
Daoud Clarke 2021-12-19 20:48:28 +00:00
parent 16121d2b19
commit 4cbed29c08
3 changed files with 38 additions and 9 deletions

16
analyse/inspect_index.py Normal file
View file

@ -0,0 +1,16 @@
from index import TinyIndex, Document, NUM_PAGES, PAGE_SIZE
from paths import INDEX_PATH
def run():
tiny_index = TinyIndex(Document, INDEX_PATH, NUM_PAGES, PAGE_SIZE)
for i in range(100):
items = tiny_index.retrieve('eggless')
# items = tiny_index.convert_items(page)
if items:
print("Items", items)
break
if __name__ == '__main__':
run()

View file

@ -27,15 +27,16 @@ def create(tiny_index: TinyIndex):
formatted_results = []
for result in results:
pattern = get_query_regex(terms)
title = result.title
matches = re.finditer(pattern, title, re.IGNORECASE)
all_spans = [0] + sum((list(m.span()) for m in matches), []) + [len(title)]
title_and_extract = f"{result.title} - {result.extract}"
matches = re.finditer(pattern, title_and_extract, re.IGNORECASE)
all_spans = [0] + sum((list(m.span()) for m in matches), []) + [len(title_and_extract)]
formatted_result = []
title_length = len(result.title)
for i in range(len(all_spans) - 1):
is_bold = i % 2 == 1
start = all_spans[i]
end = all_spans[i + 1]
formatted_result.append({'value': title[start:end], 'is_bold': is_bold})
formatted_result.append({'value': title_and_extract[start:end], 'is_bold': is_bold})
formatted_results.append({'title': formatted_result, 'url': result.url})
logger.info("Return results: %r", formatted_results)
@ -46,19 +47,22 @@ def create(tiny_index: TinyIndex):
pattern = '|'.join(term_patterns)
return pattern
def score_result(terms, r):
def score_result(terms, result: Document):
print("Score result", result)
result_string = f"{result.title} {result.extract}"
query_regex = get_query_regex(terms)
matches = re.findall(query_regex, r, flags=re.IGNORECASE)
matches = re.findall(query_regex, result_string, flags=re.IGNORECASE)
match_strings = {x.lower() for x in matches}
match_length = sum(len(x) for x in match_strings)
num_words = len(re.findall(r'\b\w+\b', r))
num_words = len(re.findall(r'\b\w+\b', result_string))
total_possible_match_length = sum(len(x) for x in terms)
return (match_length + 1./num_words) / (total_possible_match_length + 1)
def order_results(terms: list[str], results: list[Document]):
results_and_scores = [(score_result(terms, result.title), result) for result in results]
results_and_scores = [(score_result(terms, result), result) for result in results]
ordered_results = sorted(results_and_scores, key=itemgetter(0), reverse=True)
print("Ordered results", ordered_results)
filtered_results = [result for score, result in ordered_results if score > SCORE_THRESHOLD]
# ordered_results = sorted(results, key=lambda result: score_result(terms, result.title), reverse=True)
# print("Order results", query, ordered_results, sep='\n')
@ -75,15 +79,23 @@ def create(tiny_index: TinyIndex):
# print("Results", results)
return [q, results]
# TODO: why does 'leek and potato soup' result not get returned for 'potato soup' query?
def get_results(q):
terms = [x.lower() for x in q.replace('.', ' ').split()]
# completed = complete_term(terms[-1])
# terms = terms[:-1] + [completed]
pages = []
seen_items = set()
for term in terms:
items = tiny_index.retrieve(term)
print("Items", items)
if items is not None:
pages += [item for item in items if term in item.title.lower()]
for item in items:
if term in item.title.lower() or term in item.extract.lower():
if item.title not in seen_items:
pages.append(item)
seen_items.add(item.title)
ordered_results = order_results(terms, pages)
return ordered_results, terms

View file

@ -188,6 +188,7 @@ def get_pages(nlp, titles_urls_and_extracts) -> Iterable[TokenizedDocument]:
prepared_url = prepare_url_for_tokenizing(unquote(url))
url_tokens = tokenize(nlp, prepared_url)
extract_tokens = tokenize(nlp, extract)
print("Extract tokens", extract_tokens)
tokens = title_tokens | url_tokens | extract_tokens
yield TokenizedDocument(tokens=list(tokens), url=url, title=title_cleaned, extract=extract)