Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -8,7 +8,7 @@ from llama_index.core import (
|
|
8 |
from llama_index.core.memory import ChatMemoryBuffer
|
9 |
from llama_index.llms.huggingface import HuggingFaceLLM
|
10 |
from llama_index.embeddings.langchain import LangchainEmbedding
|
11 |
-
from
|
12 |
|
13 |
# ---------- Constants ----------
|
14 |
SYSTEM_PROMPT = """
|
@@ -53,8 +53,6 @@ def get_chat_engine():
|
|
53 |
generate_kwargs={"temperature": 0.2, "do_sample": True},
|
54 |
device_map="auto",
|
55 |
model_kwargs={
|
56 |
-
"torch_dtype": torch.float16,
|
57 |
-
"load_in_4bit": True,
|
58 |
"use_auth_token": os.environ["HF_TOKEN"]
|
59 |
},
|
60 |
system_prompt=SYSTEM_PROMPT,
|
|
|
8 |
from llama_index.core.memory import ChatMemoryBuffer
|
9 |
from llama_index.llms.huggingface import HuggingFaceLLM
|
10 |
from llama_index.embeddings.langchain import LangchainEmbedding
|
11 |
+
from langchain_huggingface import HuggingFaceEmbeddings
|
12 |
|
13 |
# ---------- Constants ----------
|
14 |
SYSTEM_PROMPT = """
|
|
|
53 |
generate_kwargs={"temperature": 0.2, "do_sample": True},
|
54 |
device_map="auto",
|
55 |
model_kwargs={
|
|
|
|
|
56 |
"use_auth_token": os.environ["HF_TOKEN"]
|
57 |
},
|
58 |
system_prompt=SYSTEM_PROMPT,
|