Setup File Filter for Markdown and Ledger content types

- Pass file associated with entries in markdown, beancount to json converters
- Add File, Word, Date Filters to Ledger, Markdown Types
  - Word, Date Filters were accidently removed from the above types yesterday
  - File Filter is the only filter that newly got added
This commit is contained in:
Debanjum Singh Solanky 2022-09-06 15:27:31 +03:00
parent 94cf3e97f3
commit 490157cafa
5 changed files with 62 additions and 28 deletions

View file

@ -42,30 +42,62 @@ def configure_server(args, required=False):
def configure_search(model: SearchModels, config: FullConfig, regenerate: bool, t: SearchType = None): def configure_search(model: SearchModels, config: FullConfig, regenerate: bool, t: SearchType = None):
# Initialize Org Notes Search # Initialize Org Notes Search
if (t == SearchType.Org or t == None) and config.content_type.org: if (t == SearchType.Org or t == None) and config.content_type.org:
filter_directory = resolve_absolute_path(config.content_type.org.compressed_jsonl.parent)
filters = [DateFilter(), WordFilter(filter_directory, search_type=SearchType.Org), FileFilter()]
# Extract Entries, Generate Notes Embeddings # Extract Entries, Generate Notes Embeddings
model.orgmode_search = text_search.setup(org_to_jsonl, config.content_type.org, search_config=config.search_type.asymmetric, regenerate=regenerate, filters=filters) model.orgmode_search = text_search.setup(
org_to_jsonl,
config.content_type.org,
search_config=config.search_type.asymmetric,
regenerate=regenerate,
filters=[
DateFilter(),
WordFilter(config.content_type.org.compressed_jsonl.parent, SearchType.Org),
FileFilter(),
])
# Initialize Org Music Search # Initialize Org Music Search
if (t == SearchType.Music or t == None) and config.content_type.music: if (t == SearchType.Music or t == None) and config.content_type.music:
# Extract Entries, Generate Music Embeddings # Extract Entries, Generate Music Embeddings
model.music_search = text_search.setup(org_to_jsonl, config.content_type.music, search_config=config.search_type.asymmetric, regenerate=regenerate) model.music_search = text_search.setup(
org_to_jsonl,
config.content_type.music,
search_config=config.search_type.asymmetric,
regenerate=regenerate)
# Initialize Markdown Search # Initialize Markdown Search
if (t == SearchType.Markdown or t == None) and config.content_type.markdown: if (t == SearchType.Markdown or t == None) and config.content_type.markdown:
# Extract Entries, Generate Markdown Embeddings # Extract Entries, Generate Markdown Embeddings
model.markdown_search = text_search.setup(markdown_to_jsonl, config.content_type.markdown, search_config=config.search_type.asymmetric, regenerate=regenerate) model.markdown_search = text_search.setup(
markdown_to_jsonl,
config.content_type.markdown,
search_config=config.search_type.asymmetric,
regenerate=regenerate,
filters=[
DateFilter(),
WordFilter(config.content_type.markdown.compressed_jsonl.parent, SearchType.Markdown),
FileFilter(),
])
# Initialize Ledger Search # Initialize Ledger Search
if (t == SearchType.Ledger or t == None) and config.content_type.ledger: if (t == SearchType.Ledger or t == None) and config.content_type.ledger:
# Extract Entries, Generate Ledger Embeddings # Extract Entries, Generate Ledger Embeddings
model.ledger_search = text_search.setup(beancount_to_jsonl, config.content_type.ledger, search_config=config.search_type.symmetric, regenerate=regenerate) model.ledger_search = text_search.setup(
beancount_to_jsonl,
config.content_type.ledger,
search_config=config.search_type.symmetric,
regenerate=regenerate,
filters=[
DateFilter(),
WordFilter(config.content_type.ledger.compressed_jsonl.parent, SearchType.Ledger),
FileFilter(),
])
# Initialize Image Search # Initialize Image Search
if (t == SearchType.Image or t == None) and config.content_type.image: if (t == SearchType.Image or t == None) and config.content_type.image:
# Extract Entries, Generate Image Embeddings # Extract Entries, Generate Image Embeddings
model.image_search = image_search.setup(config.content_type.image, search_config=config.search_type.image, regenerate=regenerate) model.image_search = image_search.setup(
config.content_type.image,
search_config=config.search_type.image,
regenerate=regenerate)
return model return model

View file

@ -28,10 +28,10 @@ def beancount_to_jsonl(beancount_files, beancount_file_filter, output_file):
beancount_files = get_beancount_files(beancount_files, beancount_file_filter) beancount_files = get_beancount_files(beancount_files, beancount_file_filter)
# Extract Entries from specified Beancount files # Extract Entries from specified Beancount files
entries = extract_beancount_entries(beancount_files) entries, transaction_to_file_map = extract_beancount_entries(beancount_files)
# Process Each Entry from All Notes Files # Process Each Entry from All Notes Files
jsonl_data = convert_beancount_entries_to_jsonl(entries) jsonl_data = convert_beancount_entries_to_jsonl(entries, transaction_to_file_map)
# Compress JSONL formatted Data # Compress JSONL formatted Data
if output_file.suffix == ".gz": if output_file.suffix == ".gz":
@ -74,22 +74,24 @@ def extract_beancount_entries(beancount_files):
empty_newline = f'^[{empty_escape_sequences}]*$' empty_newline = f'^[{empty_escape_sequences}]*$'
entries = [] entries = []
transaction_to_file_map = []
for beancount_file in beancount_files: for beancount_file in beancount_files:
with open(beancount_file) as f: with open(beancount_file) as f:
ledger_content = f.read() ledger_content = f.read()
entries.extend([entry.strip(empty_escape_sequences) transactions_per_file = [entry.strip(empty_escape_sequences)
for entry for entry
in re.split(empty_newline, ledger_content, flags=re.MULTILINE) in re.split(empty_newline, ledger_content, flags=re.MULTILINE)
if re.match(transaction_regex, entry)]) if re.match(transaction_regex, entry)]
transaction_to_file_map += [beancount_file]*len(transactions_per_file)
return entries entries.extend(transactions_per_file)
return entries, transaction_to_file_map
def convert_beancount_entries_to_jsonl(entries): def convert_beancount_entries_to_jsonl(entries, transaction_to_file_map):
"Convert each Beancount transaction to JSON and collate as JSONL" "Convert each Beancount transaction to JSON and collate as JSONL"
jsonl = '' jsonl = ''
for entry in entries: for entry_id, entry in enumerate(entries):
entry_dict = {'compiled': entry, 'raw': entry} entry_dict = {'compiled': entry, 'raw': entry, 'file': f'{transaction_to_file_map[entry_id]}'}
# Convert Dictionary to JSON and Append to JSONL string # Convert Dictionary to JSON and Append to JSONL string
jsonl += f'{json.dumps(entry_dict, ensure_ascii=False)}\n' jsonl += f'{json.dumps(entry_dict, ensure_ascii=False)}\n'

View file

@ -28,10 +28,10 @@ def markdown_to_jsonl(markdown_files, markdown_file_filter, output_file):
markdown_files = get_markdown_files(markdown_files, markdown_file_filter) markdown_files = get_markdown_files(markdown_files, markdown_file_filter)
# Extract Entries from specified Markdown files # Extract Entries from specified Markdown files
entries = extract_markdown_entries(markdown_files) entries, entry_to_file_map = extract_markdown_entries(markdown_files)
# Process Each Entry from All Notes Files # Process Each Entry from All Notes Files
jsonl_data = convert_markdown_entries_to_jsonl(entries) jsonl_data = convert_markdown_entries_to_jsonl(entries, entry_to_file_map)
# Compress JSONL formatted Data # Compress JSONL formatted Data
if output_file.suffix == ".gz": if output_file.suffix == ".gz":
@ -74,21 +74,24 @@ def extract_markdown_entries(markdown_files):
markdown_heading_regex = r'^#' markdown_heading_regex = r'^#'
entries = [] entries = []
entry_to_file_map = []
for markdown_file in markdown_files: for markdown_file in markdown_files:
with open(markdown_file) as f: with open(markdown_file) as f:
markdown_content = f.read() markdown_content = f.read()
entries.extend([f'#{entry.strip(empty_escape_sequences)}' markdown_entries_per_file = [f'#{entry.strip(empty_escape_sequences)}'
for entry for entry
in re.split(markdown_heading_regex, markdown_content, flags=re.MULTILINE)]) in re.split(markdown_heading_regex, markdown_content, flags=re.MULTILINE)]
entry_to_file_map += [markdown_file]*len(markdown_entries_per_file)
entries.extend(markdown_entries_per_file)
return entries return entries, entry_to_file_map
def convert_markdown_entries_to_jsonl(entries): def convert_markdown_entries_to_jsonl(entries, entry_to_file_map):
"Convert each Markdown entries to JSON and collate as JSONL" "Convert each Markdown entries to JSON and collate as JSONL"
jsonl = '' jsonl = ''
for entry in entries: for entry_id, entry in enumerate(entries):
entry_dict = {'compiled': entry, 'raw': entry} entry_dict = {'compiled': entry, 'raw': entry, 'file': f'{entry_to_file_map[entry_id]}'}
# Convert Dictionary to JSON and Append to JSONL string # Convert Dictionary to JSON and Append to JSONL string
jsonl += f'{json.dumps(entry_dict, ensure_ascii=False)}\n' jsonl += f'{json.dumps(entry_dict, ensure_ascii=False)}\n'

View file

@ -45,7 +45,7 @@ class DateFilter(BaseFilter):
continue continue
self.date_to_entry_ids[date_in_entry].add(id) self.date_to_entry_ids[date_in_entry].add(id)
end = time.time() end = time.time()
logger.debug(f"Created file filter index: {end - start} seconds") logger.debug(f"Created date filter index: {end - start} seconds")
def can_filter(self, raw_query): def can_filter(self, raw_query):

View file

@ -4,9 +4,6 @@ import time
import pickle import pickle
import logging import logging
# External Packages
import torch
# Internal Packages # Internal Packages
from src.search_filter.base_filter import BaseFilter from src.search_filter.base_filter import BaseFilter
from src.utils.helpers import LRU, resolve_absolute_path from src.utils.helpers import LRU, resolve_absolute_path