Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -271,9 +271,7 @@ def predict(
|
|
271 |
response = strip_stop_words(full_response, stop_words)
|
272 |
conversation.update_last_message(response)
|
273 |
gradio_chatbot_output[-1][1] = response
|
274 |
-
yield gradio_chatbot_output, to_gradio_history(
|
275 |
-
conversation
|
276 |
-
), "Generating..."
|
277 |
|
278 |
print("flushed result to gradio")
|
279 |
torch.cuda.empty_cache()
|
@@ -284,7 +282,7 @@ def predict(
|
|
284 |
f"temperature: {temperature}, top_p: {top_p}, repetition_penalty: {repetition_penalty}, max_length_tokens: {max_length_tokens}"
|
285 |
)
|
286 |
|
287 |
-
yield gradio_chatbot_output
|
288 |
|
289 |
|
290 |
def retry(
|
|
|
271 |
response = strip_stop_words(full_response, stop_words)
|
272 |
conversation.update_last_message(response)
|
273 |
gradio_chatbot_output[-1][1] = response
|
274 |
+
yield gradio_chatbot_output, #to_gradio_history(conversation), "Generating..."
|
|
|
|
|
275 |
|
276 |
print("flushed result to gradio")
|
277 |
torch.cuda.empty_cache()
|
|
|
282 |
f"temperature: {temperature}, top_p: {top_p}, repetition_penalty: {repetition_penalty}, max_length_tokens: {max_length_tokens}"
|
283 |
)
|
284 |
|
285 |
+
yield gradio_chatbot_output,, "Generate: Success",# to_gradio_history(conversation)
|
286 |
|
287 |
|
288 |
def retry(
|