Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -28,7 +28,7 @@ def generate(
|
|
| 28 |
elif model_name == "Felladrin/Llama-160M-Chat-v1":
|
| 29 |
outputs = pipe(prompt, max_new_tokens=250, use_cache=True, penalty_alpha=0.5, top_k=4, repetition_penalty=1.01)
|
| 30 |
elif model_name == "Felladrin/TinyMistral-248M-Chat-v4":
|
| 31 |
-
outputs = pipe(prompt, max_new_tokens=250, use_cache=True,
|
| 32 |
else:
|
| 33 |
outputs = pipe(prompt, max_new_tokens=250, do_sample=True, temperature=0.65, top_k=35, top_p=0.55, repetition_penalty=1.176)
|
| 34 |
|
|
|
|
| 28 |
elif model_name == "Felladrin/Llama-160M-Chat-v1":
|
| 29 |
outputs = pipe(prompt, max_new_tokens=250, use_cache=True, penalty_alpha=0.5, top_k=4, repetition_penalty=1.01)
|
| 30 |
elif model_name == "Felladrin/TinyMistral-248M-Chat-v4":
|
| 31 |
+
outputs = pipe(prompt, max_new_tokens=250, use_cache=True, penalty_alpha=0.5, top_k=5, repetition_penalty=1.0)
|
| 32 |
else:
|
| 33 |
outputs = pipe(prompt, max_new_tokens=250, do_sample=True, temperature=0.65, top_k=35, top_p=0.55, repetition_penalty=1.176)
|
| 34 |
|