Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -58,7 +58,7 @@ def stream_together_model(model_name, user_prompt, add_thinking_delay=False):
|
|
58 |
if add_thinking_delay:
|
59 |
# Add 4 second delay and yield thinking message
|
60 |
yield "🤔 Thinking..."
|
61 |
-
time.sleep(
|
62 |
# Clear the thinking message
|
63 |
yield ""
|
64 |
|
@@ -93,7 +93,7 @@ def stream_model_c(user_prompt, enable_thinking=True):
|
|
93 |
{"role": "user", "content": user_prompt}
|
94 |
],
|
95 |
"stream": True,
|
96 |
-
"max_tokens":
|
97 |
"enable_thinking": enable_thinking # Add thinking mode parameter
|
98 |
}
|
99 |
|
@@ -328,8 +328,7 @@ h1 {
|
|
328 |
with gr.Blocks(css=custom_css, theme=gr.themes.Base()) as demo:
|
329 |
gr.HTML("""
|
330 |
<div class="container">
|
331 |
-
<h1>Palmyra
|
332 |
-
<p class="subtitle">Compare responses from DeepSeek, Llama, and Baseten models</p>
|
333 |
</div>
|
334 |
""")
|
335 |
|
|
|
58 |
if add_thinking_delay:
|
59 |
# Add 4 second delay and yield thinking message
|
60 |
yield "🤔 Thinking..."
|
61 |
+
time.sleep(8)
|
62 |
# Clear the thinking message
|
63 |
yield ""
|
64 |
|
|
|
93 |
{"role": "user", "content": user_prompt}
|
94 |
],
|
95 |
"stream": True,
|
96 |
+
"max_tokens": 14096,
|
97 |
"enable_thinking": enable_thinking # Add thinking mode parameter
|
98 |
}
|
99 |
|
|
|
328 |
with gr.Blocks(css=custom_css, theme=gr.themes.Base()) as demo:
|
329 |
gr.HTML("""
|
330 |
<div class="container">
|
331 |
+
<h1>Palmyra-x5</h1>
|
|
|
332 |
</div>
|
333 |
""")
|
334 |
|