Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -189,17 +189,16 @@ async def respond_stream(message: str, chat_history: List[Dict[str, str]], max_t
|
|
189 |
messages.append({"role": "user", "content": message})
|
190 |
|
191 |
try:
|
192 |
-
stream = hf_client.
|
193 |
-
"CohereForAI/c4ai-command-r-plus-08-2024",
|
194 |
messages,
|
195 |
-
|
196 |
temperature=temperature,
|
197 |
top_p=top_p,
|
198 |
stream=True
|
199 |
)
|
200 |
|
201 |
full_response = ""
|
202 |
-
for response in stream:
|
203 |
if response.token.special:
|
204 |
continue
|
205 |
full_response += response.token.text
|
@@ -320,7 +319,7 @@ def create_ui():
|
|
320 |
msg = gr.Textbox(label="๋ฉ์์ง")
|
321 |
|
322 |
max_tokens = gr.Slider(minimum=1, maximum=8000, value=4000, label="Max Tokens", visible=False)
|
323 |
-
temperature = gr.Slider(minimum=0, maximum=1, value=0.
|
324 |
top_p = gr.Slider(minimum=0, maximum=1, value=0.9, label="Top P", visible=False)
|
325 |
|
326 |
examples = [
|
@@ -331,20 +330,20 @@ def create_ui():
|
|
331 |
["ํนํ ์ถ์์ ํ์ฉํ ๊ธฐ์ ๋ฐ ๋น์ฆ๋์ค๋ชจ๋ธ ์ธก๋ฉด์ ํฌํจํ์ฌ ํนํ ์ถ์์ ๊ตฌ์ฑ์ ๋ง๊ฒ ํ์ ์ ์ธ ์ฐฝ์ ๋ฐ๋ช
๋ด์ฉ์ ์ค์ฌ์ผ๋ก 4000ํ ํฐ ์ด์ ์์ฑํ๋ผ."],
|
332 |
["๊ณ์ ์ด์ด์ ๋ต๋ณํ๋ผ"],
|
333 |
]
|
334 |
-
|
335 |
gr.Examples(examples, inputs=msg)
|
336 |
-
|
337 |
def respond_wrapper(message, chat_history, max_tokens, temperature, top_p):
|
338 |
-
|
|
|
339 |
chat_history.append({"role": "user", "content": message})
|
340 |
chat_history.append({"role": "assistant", "content": bot_message})
|
341 |
return "", chat_history
|
342 |
-
|
343 |
-
msg.submit(respond_wrapper, [msg, chatbot, max_tokens, temperature, top_p], [msg, chatbot])
|
344 |
|
|
|
345 |
|
|
|
346 |
|
347 |
-
|
348 |
|
349 |
|
350 |
space_id_state = gr.State()
|
|
|
189 |
messages.append({"role": "user", "content": message})
|
190 |
|
191 |
try:
|
192 |
+
stream = hf_client.chat_completion(
|
|
|
193 |
messages,
|
194 |
+
max_tokens=max_tokens,
|
195 |
temperature=temperature,
|
196 |
top_p=top_p,
|
197 |
stream=True
|
198 |
)
|
199 |
|
200 |
full_response = ""
|
201 |
+
async for response in stream:
|
202 |
if response.token.special:
|
203 |
continue
|
204 |
full_response += response.token.text
|
|
|
319 |
msg = gr.Textbox(label="๋ฉ์์ง")
|
320 |
|
321 |
max_tokens = gr.Slider(minimum=1, maximum=8000, value=4000, label="Max Tokens", visible=False)
|
322 |
+
temperature = gr.Slider(minimum=0, maximum=1, value=0.3, label="Temperature", visible=False)
|
323 |
top_p = gr.Slider(minimum=0, maximum=1, value=0.9, label="Top P", visible=False)
|
324 |
|
325 |
examples = [
|
|
|
330 |
["ํนํ ์ถ์์ ํ์ฉํ ๊ธฐ์ ๋ฐ ๋น์ฆ๋์ค๋ชจ๋ธ ์ธก๋ฉด์ ํฌํจํ์ฌ ํนํ ์ถ์์ ๊ตฌ์ฑ์ ๋ง๊ฒ ํ์ ์ ์ธ ์ฐฝ์ ๋ฐ๋ช
๋ด์ฉ์ ์ค์ฌ์ผ๋ก 4000ํ ํฐ ์ด์ ์์ฑํ๋ผ."],
|
331 |
["๊ณ์ ์ด์ด์ ๋ต๋ณํ๋ผ"],
|
332 |
]
|
333 |
+
|
334 |
gr.Examples(examples, inputs=msg)
|
335 |
+
|
336 |
def respond_wrapper(message, chat_history, max_tokens, temperature, top_p):
|
337 |
+
generator = respond_stream(message, chat_history, max_tokens, temperature, top_p)
|
338 |
+
bot_message = asyncio.run(generator.__anext__())
|
339 |
chat_history.append({"role": "user", "content": message})
|
340 |
chat_history.append({"role": "assistant", "content": bot_message})
|
341 |
return "", chat_history
|
|
|
|
|
342 |
|
343 |
+
msg.submit(respond_wrapper, [msg, chatbot, max_tokens, temperature, top_p], [msg, chatbot])
|
344 |
|
345 |
+
|
346 |
|
|
|
347 |
|
348 |
|
349 |
space_id_state = gr.State()
|