Spaces:
Build error
Build error
null and void
commited on
Update app.py
Browse files
app.py
CHANGED
@@ -82,16 +82,13 @@ class ConversationManager:
|
|
82 |
manager = ConversationManager()
|
83 |
|
84 |
def get_model(dropdown, custom):
|
85 |
-
|
|
|
86 |
|
87 |
def chat(model1, model2, user_input, history, inserted_response=""):
|
88 |
-
|
89 |
-
|
90 |
-
|
91 |
-
manager.load_model(model2)
|
92 |
-
except Exception as e:
|
93 |
-
return f"Error loading models: {str(e)}", ""
|
94 |
-
|
95 |
if not manager.conversation:
|
96 |
manager.initial_prompt = user_input
|
97 |
manager.clear_conversation()
|
@@ -131,6 +128,7 @@ def chat(model1, model2, user_input, history, inserted_response=""):
|
|
131 |
current_model_index = (current_model_index + 1) % 2
|
132 |
|
133 |
return history, "Conversation completed."
|
|
|
134 |
|
135 |
models = [model1, model2]
|
136 |
current_model_index = 0 if manager.current_model in ["User", "Model 2"] else 1
|
@@ -224,10 +222,10 @@ with gr.Blocks() as demo:
|
|
224 |
|
225 |
with gr.Row():
|
226 |
with gr.Column(scale=1):
|
227 |
-
model1_dropdown = gr.Dropdown(open_source_models, label="Model 1")
|
228 |
model1_custom = gr.Textbox(label="Custom Model 1")
|
229 |
with gr.Column(scale=1):
|
230 |
-
model2_dropdown = gr.Dropdown(open_source_models, label="Model 2")
|
231 |
model2_custom = gr.Textbox(label="Custom Model 2")
|
232 |
|
233 |
user_input = gr.Textbox(label="Initial prompt", lines=2)
|
@@ -250,6 +248,7 @@ with gr.Blocks() as demo:
|
|
250 |
|
251 |
user_satisfaction_input = gr.Textbox(label="Are you satisfied with the result? (Yes/No)", visible=False)
|
252 |
|
|
|
253 |
gr.Markdown("""
|
254 |
## Button Descriptions
|
255 |
- **Pause**: Temporarily stops the conversation. The current model will finish its response.
|
@@ -269,8 +268,8 @@ with gr.Blocks() as demo:
|
|
269 |
chat_output = start_btn.click(
|
270 |
chat,
|
271 |
inputs=[
|
272 |
-
|
273 |
-
|
274 |
user_input,
|
275 |
chat_history
|
276 |
],
|
@@ -290,8 +289,8 @@ with gr.Blocks() as demo:
|
|
290 |
).then(
|
291 |
chat,
|
292 |
inputs=[
|
293 |
-
|
294 |
-
|
295 |
user_input,
|
296 |
chat_history
|
297 |
],
|
@@ -302,8 +301,8 @@ with gr.Blocks() as demo:
|
|
302 |
resume_btn.click(
|
303 |
chat,
|
304 |
inputs=[
|
305 |
-
|
306 |
-
|
307 |
user_input,
|
308 |
chat_history,
|
309 |
inserted_response
|
@@ -315,8 +314,8 @@ with gr.Blocks() as demo:
|
|
315 |
restart_btn.click(
|
316 |
restart_conversation,
|
317 |
inputs=[
|
318 |
-
|
319 |
-
|
320 |
user_input
|
321 |
],
|
322 |
outputs=[chat_history, current_response]
|
@@ -326,3 +325,4 @@ with gr.Blocks() as demo:
|
|
326 |
|
327 |
if __name__ == "__main__":
|
328 |
demo.launch()
|
|
|
|
82 |
manager = ConversationManager()
|
83 |
|
84 |
def get_model(dropdown, custom):
|
85 |
+
model = custom if custom.strip() else dropdown
|
86 |
+
return (model, model) # Return a tuple (label, value)
|
87 |
|
88 |
def chat(model1, model2, user_input, history, inserted_response=""):
|
89 |
+
model1 = get_model(model1, model1_custom.value)[0]
|
90 |
+
model2 = get_model(model2, model2_custom.value)[0]
|
91 |
+
|
|
|
|
|
|
|
|
|
92 |
if not manager.conversation:
|
93 |
manager.initial_prompt = user_input
|
94 |
manager.clear_conversation()
|
|
|
128 |
current_model_index = (current_model_index + 1) % 2
|
129 |
|
130 |
return history, "Conversation completed."
|
131 |
+
|
132 |
|
133 |
models = [model1, model2]
|
134 |
current_model_index = 0 if manager.current_model in ["User", "Model 2"] else 1
|
|
|
222 |
|
223 |
with gr.Row():
|
224 |
with gr.Column(scale=1):
|
225 |
+
model1_dropdown = gr.Dropdown(choices=open_source_models, label="Model 1")
|
226 |
model1_custom = gr.Textbox(label="Custom Model 1")
|
227 |
with gr.Column(scale=1):
|
228 |
+
model2_dropdown = gr.Dropdown(choices=open_source_models, label="Model 2")
|
229 |
model2_custom = gr.Textbox(label="Custom Model 2")
|
230 |
|
231 |
user_input = gr.Textbox(label="Initial prompt", lines=2)
|
|
|
248 |
|
249 |
user_satisfaction_input = gr.Textbox(label="Are you satisfied with the result? (Yes/No)", visible=False)
|
250 |
|
251 |
+
|
252 |
gr.Markdown("""
|
253 |
## Button Descriptions
|
254 |
- **Pause**: Temporarily stops the conversation. The current model will finish its response.
|
|
|
268 |
chat_output = start_btn.click(
|
269 |
chat,
|
270 |
inputs=[
|
271 |
+
model1_dropdown,
|
272 |
+
model2_dropdown,
|
273 |
user_input,
|
274 |
chat_history
|
275 |
],
|
|
|
289 |
).then(
|
290 |
chat,
|
291 |
inputs=[
|
292 |
+
model1_dropdown,
|
293 |
+
model2_dropdown,
|
294 |
user_input,
|
295 |
chat_history
|
296 |
],
|
|
|
301 |
resume_btn.click(
|
302 |
chat,
|
303 |
inputs=[
|
304 |
+
model1_dropdown,
|
305 |
+
model2_dropdown,
|
306 |
user_input,
|
307 |
chat_history,
|
308 |
inserted_response
|
|
|
314 |
restart_btn.click(
|
315 |
restart_conversation,
|
316 |
inputs=[
|
317 |
+
model1_dropdown,
|
318 |
+
model2_dropdown,
|
319 |
user_input
|
320 |
],
|
321 |
outputs=[chat_history, current_response]
|
|
|
325 |
|
326 |
if __name__ == "__main__":
|
327 |
demo.launch()
|
328 |
+
|