Spaces:
Running
Running
Upload app.py with huggingface_hub
Browse files
app.py
CHANGED
@@ -29,11 +29,11 @@ from langchain.agents import create_tool_calling_agent, AgentExecutor
|
|
29 |
from langchain_core.prompts import ChatPromptTemplate
|
30 |
|
31 |
# LangChain OpenAI imports
|
32 |
-
from langchain_openai import AzureOpenAIEmbeddings, AzureChatOpenAI # OpenAI embeddings and models
|
33 |
-
from langchain.embeddings.openai import OpenAIEmbeddings # OpenAI embeddings for text vectors
|
34 |
# Added
|
35 |
from langchain_openai import OpenAIEmbeddings, ChatOpenAI
|
36 |
-
from langchain_openai import ChatOpenAI
|
37 |
|
38 |
# LlamaParse & LlamaIndex imports
|
39 |
from llama_parse import LlamaParse # Document parsing library
|
@@ -159,7 +159,6 @@ def expand_query(state):
|
|
159 |
vector_store = Chroma(
|
160 |
collection_name="nutritional_hypotheticals",
|
161 |
persist_directory="./nutritional_db",
|
162 |
-
#persist_directory="./nutritional_db2",
|
163 |
embedding_function=embedding_model
|
164 |
)
|
165 |
print("vector_store initialized...")
|
@@ -584,8 +583,8 @@ class NutritionBot:
|
|
584 |
#)
|
585 |
|
586 |
self.client = ChatOpenAI(
|
587 |
-
openai_api_base=endpoint,
|
588 |
-
openai_api_key=
|
589 |
model="gpt-4o",
|
590 |
streaming=False, # Explicitly disabling streaming
|
591 |
temperature=0
|
|
|
29 |
from langchain_core.prompts import ChatPromptTemplate
|
30 |
|
31 |
# LangChain OpenAI imports
|
32 |
+
#from langchain_openai import AzureOpenAIEmbeddings, AzureChatOpenAI # OpenAI embeddings and models
|
33 |
+
#from langchain.embeddings.openai import OpenAIEmbeddings # OpenAI embeddings for text vectors
|
34 |
# Added
|
35 |
from langchain_openai import OpenAIEmbeddings, ChatOpenAI
|
36 |
+
#from langchain_openai import ChatOpenAI
|
37 |
|
38 |
# LlamaParse & LlamaIndex imports
|
39 |
from llama_parse import LlamaParse # Document parsing library
|
|
|
159 |
vector_store = Chroma(
|
160 |
collection_name="nutritional_hypotheticals",
|
161 |
persist_directory="./nutritional_db",
|
|
|
162 |
embedding_function=embedding_model
|
163 |
)
|
164 |
print("vector_store initialized...")
|
|
|
583 |
#)
|
584 |
|
585 |
self.client = ChatOpenAI(
|
586 |
+
#openai_api_base=endpoint,
|
587 |
+
openai_api_key=my_api_key,
|
588 |
model="gpt-4o",
|
589 |
streaming=False, # Explicitly disabling streaming
|
590 |
temperature=0
|