Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -121,47 +121,46 @@ with gr.Blocks() as demo:
|
|
121 |
if assistant_msg is not None:
|
122 |
chat_data.append({"role": "assistant", "content": assistant_msg})
|
123 |
|
124 |
-
if refusal_condition(history[-1][0]):
|
125 |
-
history[-1][1] = '抱歉,我無法回答您這個問題'
|
126 |
-
return history
|
127 |
-
|
128 |
message = tokenizer.apply_chat_template(chat_data, tokenize=False)
|
129 |
message = message[3:] # remove SOT token
|
130 |
|
131 |
-
|
132 |
-
|
133 |
-
"model": MODEL_NAME,
|
134 |
-
"prompt": str(message),
|
135 |
-
"temperature": float(temperature) + 0.01,
|
136 |
-
"n": 1,
|
137 |
-
"max_tokens": int(max_new_tokens),
|
138 |
-
"stop": "",
|
139 |
-
"top_p": float(top_p),
|
140 |
-
"logprobs": 0,
|
141 |
-
"echo": False,
|
142 |
-
"presence_penalty": PRESENCE_PENALTY,
|
143 |
-
"frequency_penalty": FREQUENCY_PENALTY,
|
144 |
-
"stream": True,
|
145 |
-
}
|
146 |
-
|
147 |
-
with requests.post(API_URL, headers=HEADERS, data=json.dumps(data), stream=True) as r:
|
148 |
-
for response in r.iter_lines():
|
149 |
-
if len(response) > 0:
|
150 |
-
text = response.decode()
|
151 |
-
if text != "data: [DONE]":
|
152 |
-
if text.startswith("data: "):
|
153 |
-
text = text[5:]
|
154 |
-
delta = json.loads(text)["choices"][0]["text"]
|
155 |
-
|
156 |
-
if history[-1][1] is None:
|
157 |
-
history[-1][1] = delta
|
158 |
-
else:
|
159 |
-
history[-1][1] += delta
|
160 |
-
yield history
|
161 |
-
if history[-1][1].endswith('</s>'):
|
162 |
-
history[-1][1] = history[-1][1][:-4]
|
163 |
yield history
|
164 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
165 |
print('== Record ==\nQuery: {query}\nResponse: {response}'.format(query=repr(message), response=repr(history[-1][1])))
|
166 |
|
167 |
msg.submit(user, [msg, chatbot], [msg, chatbot], queue=False).then(
|
|
|
121 |
if assistant_msg is not None:
|
122 |
chat_data.append({"role": "assistant", "content": assistant_msg})
|
123 |
|
|
|
|
|
|
|
|
|
124 |
message = tokenizer.apply_chat_template(chat_data, tokenize=False)
|
125 |
message = message[3:] # remove SOT token
|
126 |
|
127 |
+
if refusal_condition(history[-1][0]):
|
128 |
+
history[-1][1] = '抱歉,我無法回答您這個問題'
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
129 |
yield history
|
130 |
+
else:
|
131 |
+
data = {
|
132 |
+
"model": MODEL_NAME,
|
133 |
+
"prompt": str(message),
|
134 |
+
"temperature": float(temperature) + 0.01,
|
135 |
+
"n": 1,
|
136 |
+
"max_tokens": int(max_new_tokens),
|
137 |
+
"stop": "",
|
138 |
+
"top_p": float(top_p),
|
139 |
+
"logprobs": 0,
|
140 |
+
"echo": False,
|
141 |
+
"presence_penalty": PRESENCE_PENALTY,
|
142 |
+
"frequency_penalty": FREQUENCY_PENALTY,
|
143 |
+
"stream": True,
|
144 |
+
}
|
145 |
+
|
146 |
+
with requests.post(API_URL, headers=HEADERS, data=json.dumps(data), stream=True) as r:
|
147 |
+
for response in r.iter_lines():
|
148 |
+
if len(response) > 0:
|
149 |
+
text = response.decode()
|
150 |
+
if text != "data: [DONE]":
|
151 |
+
if text.startswith("data: "):
|
152 |
+
text = text[5:]
|
153 |
+
delta = json.loads(text)["choices"][0]["text"]
|
154 |
+
|
155 |
+
if history[-1][1] is None:
|
156 |
+
history[-1][1] = delta
|
157 |
+
else:
|
158 |
+
history[-1][1] += delta
|
159 |
+
yield history
|
160 |
+
if history[-1][1].endswith('</s>'):
|
161 |
+
history[-1][1] = history[-1][1][:-4]
|
162 |
+
yield history
|
163 |
+
|
164 |
print('== Record ==\nQuery: {query}\nResponse: {response}'.format(query=repr(message), response=repr(history[-1][1])))
|
165 |
|
166 |
msg.submit(user, [msg, chatbot], [msg, chatbot], queue=False).then(
|