diff --git a/application/app.py b/application/app.py index 8419cf52..1185b1a6 100644 --- a/application/app.py +++ b/application/app.py @@ -33,11 +33,6 @@ import celeryconfig # os.environ["LANGCHAIN_HANDLER"] = "langchain" -if os.getenv("EMBEDDINGS_NAME") is not None: - embeddings_choice = os.getenv("EMBEDDINGS_NAME") -else: - embeddings_choice = "openai_text-embedding-ada-002" - if settings.LLM_NAME == "manifest": from manifest import Manifest from langchain.llms.manifest import ManifestWrapper @@ -119,7 +114,7 @@ def ingest(self, directory, formats, name_job, filename, user): @app.route("/") def home(): return render_template("index.html", api_key_set=api_key_set, llm_choice=settings.LLM_NAME, - embeddings_choice=embeddings_choice) + embeddings_choice=settings.EMBEDDINGS_NAME) @app.route("/api/answer", methods=["POST"]) @@ -156,13 +151,13 @@ def api_answer(): # vectorstore = "outputs/inputs/" # loading the index and the store and the prompt template # Note if you have used other embeddings than OpenAI, you need to change the embeddings - if embeddings_choice == "openai_text-embedding-ada-002": + if settings.EMBEDDINGS_NAME == "openai_text-embedding-ada-002": docsearch = FAISS.load_local(vectorstore, OpenAIEmbeddings(openai_api_key=embeddings_key)) - elif embeddings_choice == "huggingface_sentence-transformers/all-mpnet-base-v2": + elif settings.EMBEDDINGS_NAME == "huggingface_sentence-transformers/all-mpnet-base-v2": docsearch = FAISS.load_local(vectorstore, HuggingFaceHubEmbeddings()) - elif embeddings_choice == "huggingface_hkunlp/instructor-large": + elif settings.EMBEDDINGS_NAME == "huggingface_hkunlp/instructor-large": docsearch = FAISS.load_local(vectorstore, HuggingFaceInstructEmbeddings()) - elif embeddings_choice == "cohere_medium": + elif settings.EMBEDDINGS_NAME == "cohere_medium": docsearch = FAISS.load_local(vectorstore, CohereEmbeddings(cohere_api_key=embeddings_key)) # create a prompt template @@ -312,7 +307,7 @@ def combined_json(): "fullName": 'default', "date": 'default', "docLink": 'default', - "model": embeddings_choice, + "model": settings.EMBEDDINGS_NAME, "location": "local" }] # structure: name, language, version, description, fullName, date, docLink @@ -326,7 +321,7 @@ def combined_json(): "fullName": index['name'], "date": index['date'], "docLink": index['location'], - "model": embeddings_choice, + "model": settings.EMBEDDINGS_NAME, "location": "local" }) @@ -417,7 +412,7 @@ def upload_index_files(): "language": job_name, "location": save_dir, "date": datetime.datetime.now().strftime("%d/%m/%Y %H:%M:%S"), - "model": embeddings_choice, + "model": settings.EMBEDDINGS_NAME, "type": "local" }) return {"status": 'ok'} diff --git a/application/core/settings.py b/application/core/settings.py index bb5063f5..de1cbabb 100644 --- a/application/core/settings.py +++ b/application/core/settings.py @@ -4,6 +4,7 @@ from pathlib import Path class Settings(BaseSettings): LLM_NAME: str = "openai_chat" + EMBEDDINGS_NAME: str = "openai_text-embedding-ada-002" openai_token: str