Merge pull request #949 from jmorganca/matt/fixPrivateGPT
fix: private gpt example was broken due to changes in chroma
This commit is contained in:
commit
ad88799411
|
@ -6,7 +6,6 @@ PERSIST_DIRECTORY = os.environ.get('PERSIST_DIRECTORY', 'db')
|
||||||
|
|
||||||
# Define the Chroma settings
|
# Define the Chroma settings
|
||||||
CHROMA_SETTINGS = Settings(
|
CHROMA_SETTINGS = Settings(
|
||||||
chroma_db_impl='duckdb+parquet',
|
|
||||||
persist_directory=PERSIST_DIRECTORY,
|
persist_directory=PERSIST_DIRECTORY,
|
||||||
anonymized_telemetry=False
|
anonymized_telemetry=False
|
||||||
)
|
)
|
||||||
|
|
|
@ -150,7 +150,7 @@ def main():
|
||||||
print("Creating new vectorstore")
|
print("Creating new vectorstore")
|
||||||
texts = process_documents()
|
texts = process_documents()
|
||||||
print(f"Creating embeddings. May take some minutes...")
|
print(f"Creating embeddings. May take some minutes...")
|
||||||
db = Chroma.from_documents(texts, embeddings, persist_directory=persist_directory, client_settings=CHROMA_SETTINGS)
|
db = Chroma.from_documents(texts, embeddings, persist_directory=persist_directory)
|
||||||
db.persist()
|
db.persist()
|
||||||
db = None
|
db = None
|
||||||
|
|
||||||
|
|
|
@ -4,6 +4,7 @@ from langchain.embeddings import HuggingFaceEmbeddings
|
||||||
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
|
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
|
||||||
from langchain.vectorstores import Chroma
|
from langchain.vectorstores import Chroma
|
||||||
from langchain.llms import Ollama
|
from langchain.llms import Ollama
|
||||||
|
import chromadb
|
||||||
import os
|
import os
|
||||||
import argparse
|
import argparse
|
||||||
import time
|
import time
|
||||||
|
@ -22,7 +23,9 @@ def main():
|
||||||
# Parse the command line arguments
|
# Parse the command line arguments
|
||||||
args = parse_arguments()
|
args = parse_arguments()
|
||||||
embeddings = HuggingFaceEmbeddings(model_name=embeddings_model_name)
|
embeddings = HuggingFaceEmbeddings(model_name=embeddings_model_name)
|
||||||
db = Chroma(persist_directory=persist_directory, embedding_function=embeddings, client_settings=CHROMA_SETTINGS)
|
|
||||||
|
db = Chroma(persist_directory=persist_directory, embedding_function=embeddings)
|
||||||
|
|
||||||
retriever = db.as_retriever(search_kwargs={"k": target_source_chunks})
|
retriever = db.as_retriever(search_kwargs={"k": target_source_chunks})
|
||||||
# activate/deactivate the streaming StdOut callback for LLMs
|
# activate/deactivate the streaming StdOut callback for LLMs
|
||||||
callbacks = [] if args.mute_stream else [StreamingStdOutCallbackHandler()]
|
callbacks = [] if args.mute_stream else [StreamingStdOutCallbackHandler()]
|
||||||
|
|
File diff suppressed because it is too large
Load diff
Loading…
Reference in a new issue