ginipick commited on
Commit
e13feb1
ยท
verified ยท
1 Parent(s): 9857813

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -11
app.py CHANGED
@@ -189,17 +189,16 @@ async def respond_stream(message: str, chat_history: List[Dict[str, str]], max_t
189
  messages.append({"role": "user", "content": message})
190
 
191
  try:
192
- stream = hf_client.text_generation(
193
- "CohereForAI/c4ai-command-r-plus-08-2024",
194
  messages,
195
- max_new_tokens=max_tokens,
196
  temperature=temperature,
197
  top_p=top_p,
198
  stream=True
199
  )
200
 
201
  full_response = ""
202
- for response in stream:
203
  if response.token.special:
204
  continue
205
  full_response += response.token.text
@@ -320,7 +319,7 @@ def create_ui():
320
  msg = gr.Textbox(label="๋ฉ”์‹œ์ง€")
321
 
322
  max_tokens = gr.Slider(minimum=1, maximum=8000, value=4000, label="Max Tokens", visible=False)
323
- temperature = gr.Slider(minimum=0, maximum=1, value=0.7, label="Temperature", visible=False)
324
  top_p = gr.Slider(minimum=0, maximum=1, value=0.9, label="Top P", visible=False)
325
 
326
  examples = [
@@ -331,20 +330,20 @@ def create_ui():
331
  ["ํŠนํ—ˆ ์ถœ์›์— ํ™œ์šฉํ•  ๊ธฐ์ˆ  ๋ฐ ๋น„์ฆˆ๋‹ˆ์Šค๋ชจ๋ธ ์ธก๋ฉด์„ ํฌํ•จํ•˜์—ฌ ํŠนํ—ˆ ์ถœ์›์„œ ๊ตฌ์„ฑ์— ๋งž๊ฒŒ ํ˜์‹ ์ ์ธ ์ฐฝ์˜ ๋ฐœ๋ช… ๋‚ด์šฉ์„ ์ค‘์‹ฌ์œผ๋กœ 4000ํ† ํฐ ์ด์ƒ ์ž‘์„ฑํ•˜๋ผ."],
332
  ["๊ณ„์† ์ด์–ด์„œ ๋‹ต๋ณ€ํ•˜๋ผ"],
333
  ]
334
-
335
  gr.Examples(examples, inputs=msg)
336
-
337
  def respond_wrapper(message, chat_history, max_tokens, temperature, top_p):
338
- bot_message = asyncio.run(respond_stream(message, chat_history, max_tokens, temperature, top_p).__anext__())
 
339
  chat_history.append({"role": "user", "content": message})
340
  chat_history.append({"role": "assistant", "content": bot_message})
341
  return "", chat_history
342
-
343
- msg.submit(respond_wrapper, [msg, chatbot, max_tokens, temperature, top_p], [msg, chatbot])
344
 
 
345
 
 
346
 
347
-
348
 
349
 
350
  space_id_state = gr.State()
 
189
  messages.append({"role": "user", "content": message})
190
 
191
  try:
192
+ stream = hf_client.chat_completion(
 
193
  messages,
194
+ max_tokens=max_tokens,
195
  temperature=temperature,
196
  top_p=top_p,
197
  stream=True
198
  )
199
 
200
  full_response = ""
201
+ async for response in stream:
202
  if response.token.special:
203
  continue
204
  full_response += response.token.text
 
319
  msg = gr.Textbox(label="๋ฉ”์‹œ์ง€")
320
 
321
  max_tokens = gr.Slider(minimum=1, maximum=8000, value=4000, label="Max Tokens", visible=False)
322
+ temperature = gr.Slider(minimum=0, maximum=1, value=0.3, label="Temperature", visible=False)
323
  top_p = gr.Slider(minimum=0, maximum=1, value=0.9, label="Top P", visible=False)
324
 
325
  examples = [
 
330
  ["ํŠนํ—ˆ ์ถœ์›์— ํ™œ์šฉํ•  ๊ธฐ์ˆ  ๋ฐ ๋น„์ฆˆ๋‹ˆ์Šค๋ชจ๋ธ ์ธก๋ฉด์„ ํฌํ•จํ•˜์—ฌ ํŠนํ—ˆ ์ถœ์›์„œ ๊ตฌ์„ฑ์— ๋งž๊ฒŒ ํ˜์‹ ์ ์ธ ์ฐฝ์˜ ๋ฐœ๋ช… ๋‚ด์šฉ์„ ์ค‘์‹ฌ์œผ๋กœ 4000ํ† ํฐ ์ด์ƒ ์ž‘์„ฑํ•˜๋ผ."],
331
  ["๊ณ„์† ์ด์–ด์„œ ๋‹ต๋ณ€ํ•˜๋ผ"],
332
  ]
333
+
334
  gr.Examples(examples, inputs=msg)
335
+
336
  def respond_wrapper(message, chat_history, max_tokens, temperature, top_p):
337
+ generator = respond_stream(message, chat_history, max_tokens, temperature, top_p)
338
+ bot_message = asyncio.run(generator.__anext__())
339
  chat_history.append({"role": "user", "content": message})
340
  chat_history.append({"role": "assistant", "content": bot_message})
341
  return "", chat_history
 
 
342
 
343
+ msg.submit(respond_wrapper, [msg, chatbot, max_tokens, temperature, top_p], [msg, chatbot])
344
 
345
+
346
 
 
347
 
348
 
349
  space_id_state = gr.State()