Spaces:
Sleeping
Sleeping
some other debug code added
Browse files
app.py
CHANGED
@@ -11,12 +11,12 @@ model = model
|
|
11 |
|
12 |
|
13 |
def predict(message, history):
|
14 |
-
print(history)
|
15 |
history_transformer_format = history + [{"role": "user", "content": message},
|
16 |
{"role": "assistant", "content": ""}]
|
17 |
|
18 |
model_inputs = tokenizer.apply_chat_template(history_transformer_format, return_tensors="pt")
|
19 |
-
streamer = TextIteratorStreamer(tokenizer, timeout=
|
20 |
generate_kwargs = dict(
|
21 |
streamer=streamer,
|
22 |
max_new_tokens=1024,
|
@@ -32,6 +32,7 @@ def predict(message, history):
|
|
32 |
|
33 |
partial_message = ""
|
34 |
for new_token in streamer:
|
|
|
35 |
if new_token != '<':
|
36 |
partial_message += new_token
|
37 |
yield partial_message
|
|
|
11 |
|
12 |
|
13 |
def predict(message, history):
|
14 |
+
# print(history) [[вопрос1, ответ1], [вопрос2, ответ2]...]
|
15 |
history_transformer_format = history + [{"role": "user", "content": message},
|
16 |
{"role": "assistant", "content": ""}]
|
17 |
|
18 |
model_inputs = tokenizer.apply_chat_template(history_transformer_format, return_tensors="pt")
|
19 |
+
streamer = TextIteratorStreamer(tokenizer, timeout=20., skip_prompt=True, skip_special_tokens=True)
|
20 |
generate_kwargs = dict(
|
21 |
streamer=streamer,
|
22 |
max_new_tokens=1024,
|
|
|
32 |
|
33 |
partial_message = ""
|
34 |
for new_token in streamer:
|
35 |
+
print(new_token)
|
36 |
if new_token != '<':
|
37 |
partial_message += new_token
|
38 |
yield partial_message
|