Aleksandr Maiorov commited on
Commit
1f1b386
·
1 Parent(s): 459539b

- добавлен llama-index

Files changed (1) hide show
  1. app.py +9 -1
app.py CHANGED
@@ -18,14 +18,18 @@ logger = logging.getLogger(__name__)
18
  logger.info("Запускаемся... 🥳🥳🥳")
19
 
20
  app = FastAPI()
 
21
  model_url = "https://huggingface.co/Qwen/Qwen2.5-7B-Instruct-GGUF/resolve/main/qwen2.5-7b-instruct-q3_k_m.gguf"
 
22
  SYSTEM_PROMPT = ''
 
23
  tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-7B-Instruct")
 
24
  set_global_tokenizer(
25
  AutoTokenizer.from_pretrained("Qwen/Qwen2.5-7B-Instruct").encode
26
  )
27
 
28
- embed_model = HuggingFaceEmbedding(model_name="BAAI/bge-small-en-v1.5")
29
 
30
  documents = SimpleDirectoryReader("./data/").load_data()
31
 
@@ -94,8 +98,12 @@ async def set_system_prompt(text: str):
94
  async def predict(text: str):
95
  # Генерация ответа с помощью модели
96
  logger.info('post/predict')
 
 
97
  # response = llm.complete(text)
98
  response = query_engine.query(text)
 
 
99
  # text_response = generate_response(response)
100
  # return {"response": text_response}
101
  return {"response": response}
 
18
  logger.info("Запускаемся... 🥳🥳🥳")
19
 
20
  app = FastAPI()
21
+
22
  model_url = "https://huggingface.co/Qwen/Qwen2.5-7B-Instruct-GGUF/resolve/main/qwen2.5-7b-instruct-q3_k_m.gguf"
23
+
24
  SYSTEM_PROMPT = ''
25
+
26
  tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-7B-Instruct")
27
+
28
  set_global_tokenizer(
29
  AutoTokenizer.from_pretrained("Qwen/Qwen2.5-7B-Instruct").encode
30
  )
31
 
32
+ embed_model = HuggingFaceEmbedding(model_name="sentence-transformers/all-mpnet-base-v2")
33
 
34
  documents = SimpleDirectoryReader("./data/").load_data()
35
 
 
98
  async def predict(text: str):
99
  # Генерация ответа с помощью модели
100
  logger.info('post/predict')
101
+ logger.info('ЗАПРОС:')
102
+ logger.info(text)
103
  # response = llm.complete(text)
104
  response = query_engine.query(text)
105
+ logger.info('ОТВЕТ:')
106
+ logger.info(response)
107
  # text_response = generate_response(response)
108
  # return {"response": text_response}
109
  return {"response": response}