Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -359,13 +359,13 @@ def generate(
|
|
359 |
formatted_prompt = format_prompt(formatted_prompt, history, max_history_turns=5) # Truncated history
|
360 |
logging.info(f"Formatted Prompt: {formatted_prompt}")
|
361 |
stream = client.text_generation(formatted_prompt, temperature=temperature, max_new_tokens=max_new_tokens, top_p=top_p, repetition_penalty=repetition_penalty, stream=True, details=True, return_full_text=False)
|
362 |
-
|
363 |
for response in stream:
|
364 |
resp += response.token.text
|
|
|
365 |
|
366 |
if VERBOSE:
|
367 |
-
logging.info(LOG_RESPONSE.format(
|
368 |
-
return formatted_prompt # Return resp here
|
369 |
|
370 |
def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
|
371 |
"""Generates text in chunks to avoid token limit errors."""
|
|
|
359 |
formatted_prompt = format_prompt(formatted_prompt, history, max_history_turns=5) # Truncated history
|
360 |
logging.info(f"Formatted Prompt: {formatted_prompt}")
|
361 |
stream = client.text_generation(formatted_prompt, temperature=temperature, max_new_tokens=max_new_tokens, top_p=top_p, repetition_penalty=repetition_penalty, stream=True, details=True, return_full_text=False)
|
362 |
+
resp = ""
|
363 |
for response in stream:
|
364 |
resp += response.token.text
|
365 |
+
yield resp # This allows for streaming the response
|
366 |
|
367 |
if VERBOSE:
|
368 |
+
logging.info(LOG_RESPONSE.format(resp)) # Pass resp to format
|
|
|
369 |
|
370 |
def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
|
371 |
"""Generates text in chunks to avoid token limit errors."""
|