Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -204,12 +204,12 @@ def dotranslate():
|
|
| 204 |
|
| 205 |
# Tokenize the input text
|
| 206 |
encoded = tokenizer(txt, return_tensors="pt")
|
| 207 |
-
|
| 208 |
-
|
| 209 |
-
|
| 210 |
-
|
| 211 |
-
|
| 212 |
-
|
| 213 |
|
| 214 |
# Decode the generated tokens
|
| 215 |
translation = tokenizer.decode(generated_tokens[0], skip_special_tokens=True)
|
|
|
|
| 204 |
|
| 205 |
# Tokenize the input text
|
| 206 |
encoded = tokenizer(txt, return_tensors="pt")
|
| 207 |
+
with torch.no_grad():
|
| 208 |
+
# Generate translation
|
| 209 |
+
generated_tokens = model.generate(
|
| 210 |
+
**encoded,
|
| 211 |
+
forced_bos_token_id=tokenizer.get_lang_id(dest)
|
| 212 |
+
)
|
| 213 |
|
| 214 |
# Decode the generated tokens
|
| 215 |
translation = tokenizer.decode(generated_tokens[0], skip_special_tokens=True)
|