Spaces:
Running
Running
Commit
·
ca7a52b
1
Parent(s):
36c21dd
Update main.py
Browse files
main.py
CHANGED
@@ -37,11 +37,12 @@ def generate(prompt, history, temperature=0.2, max_new_tokens=30000, top_p=0.95,
|
|
37 |
seed=42,
|
38 |
)
|
39 |
formatted_prompt = format_prompt(prompt, history)
|
40 |
-
|
|
|
41 |
|
|
|
42 |
# Accumula l'output in una lista
|
43 |
-
output_list = []
|
44 |
-
for response in stream:
|
45 |
-
|
46 |
-
|
47 |
-
return iter(output_list) # Restituisci la lista come un iteratore
|
|
|
37 |
seed=42,
|
38 |
)
|
39 |
formatted_prompt = format_prompt(prompt, history)
|
40 |
+
output = client.text_generation(formatted_prompt, **generate_kwargs, stream=False, details=True)
|
41 |
+
return output
|
42 |
|
43 |
+
#stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
|
44 |
# Accumula l'output in una lista
|
45 |
+
#output_list = []
|
46 |
+
#for response in stream:
|
47 |
+
# output_list.append(response.token.text)
|
48 |
+
#return iter(output_list) # Restituisci la lista come un iteratore
|
|