search-assistant / read_write_index.py
arabellastrange's picture
added fallback documents, trying to make gpt responses longer
428be9a
raw
history blame
892 Bytes
import logging
import os
from llama_index.core import SimpleDirectoryReader, VectorStoreIndex, StorageContext, load_index_from_storage
logger = logging.getLogger(__name__)
DOCUMENT_PATH = 'search_data/'
# remember to delete stored vectors when new documents are added to the data so the storage is recreated
def read_write_index(path):
if not os.path.exists(path):
documents = SimpleDirectoryReader(DOCUMENT_PATH).load_data()
logger.info(f'Indexing documents in {DOCUMENT_PATH}...')
index = VectorStoreIndex.from_documents(documents)
index.storage_context.persist(persist_dir=path)
logger.info(f'{len(documents)} documents indexed.')
else:
logger.info(f'Loading index from {path}...')
storage_context = StorageContext.from_defaults(persist_dir=path)
index = load_index_from_storage(storage_context)
return index