Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -47,7 +47,7 @@ class HfModelWrapper:
|
|
47 |
input_text = self.build_prompt(user_input, history)
|
48 |
input_ids = self.tokenizer.encode(input_text, return_tensors="pt").to("cuda")
|
49 |
|
50 |
-
streamer = TextIteratorStreamer(self.tokenizer, skip_prompt=True
|
51 |
|
52 |
gen_kwargs = {
|
53 |
"inputs": input_ids,
|
@@ -57,7 +57,7 @@ class HfModelWrapper:
|
|
57 |
"temperature": 0.1,
|
58 |
"top_p": 0.9,
|
59 |
#"repetition_penalty": 1.1,
|
60 |
-
"min_length":
|
61 |
}
|
62 |
|
63 |
thread = Thread(target=self.model.generate, kwargs=gen_kwargs)
|
|
|
47 |
input_text = self.build_prompt(user_input, history)
|
48 |
input_ids = self.tokenizer.encode(input_text, return_tensors="pt").to("cuda")
|
49 |
|
50 |
+
streamer = TextIteratorStreamer(self.tokenizer, skip_prompt=True)#skip_special_tokens=True)
|
51 |
|
52 |
gen_kwargs = {
|
53 |
"inputs": input_ids,
|
|
|
57 |
"temperature": 0.1,
|
58 |
"top_p": 0.9,
|
59 |
#"repetition_penalty": 1.1,
|
60 |
+
"min_length": 10,
|
61 |
}
|
62 |
|
63 |
thread = Thread(target=self.model.generate, kwargs=gen_kwargs)
|