Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -68,7 +68,7 @@ def analyze_chunk(chunk):
|
|
68 |
|
69 |
try:
|
70 |
completion = client.chat.completions.create(
|
71 |
-
model="llama-3.
|
72 |
messages=messages,
|
73 |
temperature=0.7,
|
74 |
max_tokens=4096,
|
@@ -104,7 +104,7 @@ def analyze_document(file):
|
|
104 |
"Make the final output concise and coherent."
|
105 |
)
|
106 |
final_summary = client.chat.completions.create(
|
107 |
-
model="llama-3.
|
108 |
messages=[
|
109 |
{"role": "system", "content": consolidated_summary_prompt},
|
110 |
{"role": "user", "content": "\n\n".join(all_insights)}
|
|
|
68 |
|
69 |
try:
|
70 |
completion = client.chat.completions.create(
|
71 |
+
model="llama-3.3-70b-versatile",
|
72 |
messages=messages,
|
73 |
temperature=0.7,
|
74 |
max_tokens=4096,
|
|
|
104 |
"Make the final output concise and coherent."
|
105 |
)
|
106 |
final_summary = client.chat.completions.create(
|
107 |
+
model="llama-3.3-70b-versatile",
|
108 |
messages=[
|
109 |
{"role": "system", "content": consolidated_summary_prompt},
|
110 |
{"role": "user", "content": "\n\n".join(all_insights)}
|