mirror of
https://github.com/khoj-ai/khoj.git
synced 2024-11-23 23:48:56 +01:00
Modularize script, provide cmdline control, improve results rendering
This commit is contained in:
parent
eb03f57917
commit
ad7e90bec3
1 changed files with 95 additions and 47 deletions
142
asymmetric.py
142
asymmetric.py
|
@ -7,71 +7,119 @@ import gzip
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import torch
|
import torch
|
||||||
|
import argparse
|
||||||
|
import pathlib
|
||||||
|
|
||||||
# We use the Bi-Encoder to encode all passages, so that we can use it with sematic search
|
|
||||||
model_name = 'msmarco-MiniLM-L-6-v3'
|
|
||||||
bi_encoder = SentenceTransformer(model_name)
|
|
||||||
top_k = 100 # Number of passages we want to retrieve with the bi-encoder
|
|
||||||
|
|
||||||
# The bi-encoder will retrieve 100 documents.
|
def initialize_model():
|
||||||
# We use a cross-encoder, to re-rank the results list to improve the quality
|
"Initialize model for assymetric semantic search. That is, where query smaller than results"
|
||||||
cross_encoder = CrossEncoder('cross-encoder/ms-marco-MiniLM-L-6-v2')
|
bi_encoder = SentenceTransformer('msmarco-MiniLM-L-6-v3') # The bi-encoder encodes all entries to use for semantic search
|
||||||
|
top_k = 100 # Number of entries we want to retrieve with the bi-encoder
|
||||||
|
cross_encoder = CrossEncoder('cross-encoder/ms-marco-MiniLM-L-6-v2') # The cross-encoder re-ranks the results to improve quality
|
||||||
|
return bi_encoder, cross_encoder, top_k
|
||||||
|
|
||||||
# We split these articles into paragraphs and encode them with the bi-encoder
|
|
||||||
notes_filepath = 'Notes.jsonl.gz'
|
|
||||||
|
|
||||||
passages = []
|
def extract_entries(notesfile, verbose=False):
|
||||||
with gzip.open(notes_filepath, 'rt', encoding='utf8') as fIn:
|
"Load entries from compressed jsonl"
|
||||||
for line in fIn:
|
entries = []
|
||||||
data = json.loads(line.strip())
|
with gzip.open(str(notesfile.expanduser()), 'rt', encoding='utf8') as jsonl:
|
||||||
passages.extend([f'{data["Title"]}\n{data["Body"] if "Body" in data else ""}'])
|
for line in jsonl:
|
||||||
|
note = json.loads(line.strip())
|
||||||
|
note_string = f'{note["Title"]}\n{note["Body"] if "Body" in note else ""}'
|
||||||
|
entries.extend([note_string])
|
||||||
|
|
||||||
print(f"Passages: {len(passages)}")
|
if verbose:
|
||||||
|
print(f"Loaded {len(entries)} entries from {notesfile}")
|
||||||
|
|
||||||
embeddings_filename = 'notes_embeddings.pt'
|
return entries
|
||||||
# Load pre-computed embeddings from file if exists
|
|
||||||
if os.path.exists(embeddings_filename):
|
|
||||||
corpus_embeddings = torch.load(embeddings_filename)
|
|
||||||
else: # Else compute the corpus_embeddings from scratch, which can take a while
|
|
||||||
corpus_embeddings = bi_encoder.encode(passages, convert_to_tensor=True, show_progress_bar=True)
|
|
||||||
torch.save(corpus_embeddings, 'notes_embeddings.pt')
|
|
||||||
|
|
||||||
# This function will search all notes for passages that answer the query
|
|
||||||
def search(query):
|
|
||||||
print("Input question:", query)
|
|
||||||
|
|
||||||
##### Sematic Search #####
|
def compute_embeddings(entries, bi_encoder, embeddings_file, verbose=False):
|
||||||
# Encode the query using the bi-encoder and find potentially relevant passages
|
"Compute (and Save) Embeddings or Load Pre-Computed Embeddings"
|
||||||
|
# Load pre-computed embeddings from file if exists
|
||||||
|
if embeddings_file.exists():
|
||||||
|
corpus_embeddings = torch.load(str(embeddings_file.expanduser()))
|
||||||
|
if verbose:
|
||||||
|
print(f"Loaded embeddings from {embeddings_file}")
|
||||||
|
|
||||||
|
else: # Else compute the corpus_embeddings from scratch, which can take a while
|
||||||
|
corpus_embeddings = bi_encoder.encode(entries, convert_to_tensor=True, show_progress_bar=True)
|
||||||
|
torch.save(corpus_embeddings, str(embeddings_file.expanduser()))
|
||||||
|
if verbose:
|
||||||
|
print(f"Computed embeddings and save them to {embeddings_file}")
|
||||||
|
|
||||||
|
return corpus_embeddings
|
||||||
|
|
||||||
|
|
||||||
|
def query_notes(query, corpus_embeddings, entries, bi_encoder, cross_encoder, topk=100):
|
||||||
|
"Search all notes for entries that answer the query"
|
||||||
|
# Encode the query using the bi-encoder
|
||||||
question_embedding = bi_encoder.encode(query, convert_to_tensor=True)
|
question_embedding = bi_encoder.encode(query, convert_to_tensor=True)
|
||||||
#question_embedding = question_embedding.cuda()
|
|
||||||
|
# Find relevant entries for the query
|
||||||
hits = util.semantic_search(question_embedding, corpus_embeddings, top_k=top_k)
|
hits = util.semantic_search(question_embedding, corpus_embeddings, top_k=top_k)
|
||||||
hits = hits[0] # Get the hits for the first query
|
hits = hits[0] # Get the hits for the first query
|
||||||
|
|
||||||
##### Re-Ranking #####
|
# Score all retrieved entries using the cross-encoder
|
||||||
# Now, score all retrieved passages with the cross_encoder
|
cross_inp = [[query, entries[hit['corpus_id']]] for hit in hits]
|
||||||
cross_inp = [[query, passages[hit['corpus_id']]] for hit in hits]
|
|
||||||
cross_scores = cross_encoder.predict(cross_inp)
|
cross_scores = cross_encoder.predict(cross_inp)
|
||||||
|
|
||||||
# Sort results by the cross-encoder scores
|
# Store cross-encoder scores in results dictionary for ranking
|
||||||
for idx in range(len(cross_scores)):
|
for idx in range(len(cross_scores)):
|
||||||
hits[idx]['cross-score'] = cross_scores[idx]
|
hits[idx]['cross-score'] = cross_scores[idx]
|
||||||
|
|
||||||
# Output of top-5 hits from bi-encoder
|
# Order results by cross encoder score followed by biencoder score
|
||||||
print("\n-------------------------\n")
|
hits.sort(key=lambda x: x['score'], reverse=True) # sort by biencoder score
|
||||||
print("Top-3 Bi-Encoder Retrieval hits")
|
hits.sort(key=lambda x: x['cross-score'], reverse=True) # sort by cross encoder score
|
||||||
hits = sorted(hits, key=lambda x: x['score'], reverse=True)
|
return hits
|
||||||
for hit in hits[0:3]:
|
|
||||||
print("\t{:.3f}\t{}".format(hit['score'], passages[hit['corpus_id']].replace("\n", " ")))
|
|
||||||
|
|
||||||
# Output of top-5 hits from re-ranker
|
|
||||||
|
def render_results(hits, entries, count=5, display_biencoder_results=False):
|
||||||
|
"Render the Results returned by Search for the Query"
|
||||||
|
if display_biencoder_results:
|
||||||
|
# Output of top hits from bi-encoder
|
||||||
|
print("\n-------------------------\n")
|
||||||
|
print(f"Top-{count} Bi-Encoder Retrieval hits")
|
||||||
|
hits = sorted(hits, key=lambda x: x['score'], reverse=True)
|
||||||
|
for hit in hits[0:count]:
|
||||||
|
print(f"Score: {hit['score']:.3f}\n------------\n{entries[hit['corpus_id']]}")
|
||||||
|
|
||||||
|
# Output of top hits from re-ranker
|
||||||
print("\n-------------------------\n")
|
print("\n-------------------------\n")
|
||||||
print("Top-3 Cross-Encoder Re-ranker hits")
|
print(f"Top-{count} Cross-Encoder Re-ranker hits")
|
||||||
hits = sorted(hits, key=lambda x: x['cross-score'], reverse=True)
|
hits = sorted(hits, key=lambda x: x['cross-score'], reverse=True)
|
||||||
for hit in hits[0:3]:
|
for hit in hits[0:count]:
|
||||||
print("\t{:.3f}\t{}".format(hit['cross-score'], passages[hit['corpus_id']].replace("\n", " ")))
|
print(f"CrossScore: {hit['cross-score']:.3f}\n-----------------\n{entries[hit['corpus_id']]}")
|
||||||
|
|
||||||
while True:
|
|
||||||
user_query = input("Enter your query: ")
|
if __name__ == '__main__':
|
||||||
if user_query == "exit":
|
# Setup Argument Parser
|
||||||
exit(0)
|
parser = argparse.ArgumentParser(description="Map Org-Mode notes into JSONL format")
|
||||||
search(query = user_query)
|
parser.add_argument('--jsonl-file', '-j', required=True, type=pathlib.Path, help="Input file for compressed JSONL formatted notes to compute embeddings from")
|
||||||
|
parser.add_argument('--embeddings-file', '-e', type=pathlib.Path, help="File to save/load model embeddings to/from. Default: ./embeddings.pt")
|
||||||
|
parser.add_argument('--results-count', '-n', default=5, type=int, help="Number of results to render. Default: 5")
|
||||||
|
parser.add_argument('--interactive', action='store_true', default=False, help="Interactive mode allows user to run queries on the model. Default: true")
|
||||||
|
parser.add_argument('--verbose', action='store_true', default=False, help="Show verbose conversion logs. Default: false")
|
||||||
|
args = parser.parse_args()
|
||||||
|
|
||||||
|
# Initialize Model
|
||||||
|
bi_encoder, cross_encoder, top_k = initialize_model()
|
||||||
|
|
||||||
|
# Extract Entries
|
||||||
|
entries = extract_entries(args.jsonl_file, args.verbose)
|
||||||
|
|
||||||
|
# Compute or Load Embeddings
|
||||||
|
corpus_embeddings = compute_embeddings(entries, bi_encoder, args.embeddings_file, args.verbose)
|
||||||
|
|
||||||
|
# Run User Queries on Entries in Interactive Mode
|
||||||
|
while args.interactive:
|
||||||
|
# get query from user
|
||||||
|
user_query = input("Enter your query: ")
|
||||||
|
if user_query == "exit":
|
||||||
|
exit(0)
|
||||||
|
|
||||||
|
# query notes
|
||||||
|
hits = query_notes(user_query, corpus_embeddings, entries, bi_encoder, cross_encoder, top_k)
|
||||||
|
|
||||||
|
# render results
|
||||||
|
render_results(hits, entries, count=args.results_count)
|
||||||
|
|
Loading…
Reference in a new issue