cleaned up
Browse files- README.md +2 -4
- generate.py +2 -4
README.md
CHANGED
|
@@ -122,14 +122,12 @@ def generate_text(prompt, max_new_tokens=128, temperature=0.7, top_p=0.9, top_k=
|
|
| 122 |
new_tokens = outputs[0][input_ids.shape[1]:].tolist()
|
| 123 |
generated = tokenizer.decode(new_tokens)
|
| 124 |
|
| 125 |
-
print(f"
|
| 126 |
print(f"\nTokens generated: {len(new_tokens)}")
|
| 127 |
|
| 128 |
if __name__ == "__main__":
|
| 129 |
prompt = "In a small village nestled between mountains,"
|
| 130 |
-
|
| 131 |
-
# Test different generation settings
|
| 132 |
-
print(f"PROMPT: {prompt}")
|
| 133 |
generate_text(prompt)
|
| 134 |
```
|
| 135 |
|
|
|
|
| 122 |
new_tokens = outputs[0][input_ids.shape[1]:].tolist()
|
| 123 |
generated = tokenizer.decode(new_tokens)
|
| 124 |
|
| 125 |
+
print(f"{prompt}{generated}")
|
| 126 |
print(f"\nTokens generated: {len(new_tokens)}")
|
| 127 |
|
| 128 |
if __name__ == "__main__":
|
| 129 |
prompt = "In a small village nestled between mountains,"
|
| 130 |
+
print(f"PROMPT: {prompt}\n--")
|
|
|
|
|
|
|
| 131 |
generate_text(prompt)
|
| 132 |
```
|
| 133 |
|
generate.py
CHANGED
|
@@ -43,12 +43,10 @@ def generate_text(prompt, max_new_tokens=128, temperature=0.7, top_p=0.9, top_k=
|
|
| 43 |
new_tokens = outputs[0][input_ids.shape[1]:].tolist()
|
| 44 |
generated = tokenizer.decode(new_tokens)
|
| 45 |
|
| 46 |
-
print(f"
|
| 47 |
print(f"\nTokens generated: {len(new_tokens)}")
|
| 48 |
|
| 49 |
if __name__ == "__main__":
|
| 50 |
prompt = "In a small village nestled between mountains,"
|
| 51 |
-
|
| 52 |
-
# Test different generation settings
|
| 53 |
-
print(f"PROMPT: {prompt}")
|
| 54 |
generate_text(prompt)
|
|
|
|
| 43 |
new_tokens = outputs[0][input_ids.shape[1]:].tolist()
|
| 44 |
generated = tokenizer.decode(new_tokens)
|
| 45 |
|
| 46 |
+
print(f"{prompt}{generated}")
|
| 47 |
print(f"\nTokens generated: {len(new_tokens)}")
|
| 48 |
|
| 49 |
if __name__ == "__main__":
|
| 50 |
prompt = "In a small village nestled between mountains,"
|
| 51 |
+
print(f"PROMPT: {prompt}\n--")
|
|
|
|
|
|
|
| 52 |
generate_text(prompt)
|