Snapshot
Browse files
app.py
CHANGED
@@ -107,11 +107,11 @@ model, tokenizer = load_model_and_tokenizer(model_name, device)
|
|
107 |
#%%
|
108 |
|
109 |
input_text = "He asked me to prostrate myself before the king, but I rifused."
|
110 |
-
|
111 |
|
112 |
#%%
|
113 |
|
114 |
-
token_probs: list[tuple[int, float]] = calculate_log_probabilities(model, tokenizer,
|
115 |
|
116 |
#%%
|
117 |
|
|
|
107 |
#%%
|
108 |
|
109 |
input_text = "He asked me to prostrate myself before the king, but I rifused."
|
110 |
+
inputs: BatchEncoding = tokenize(input_text, tokenizer, device)
|
111 |
|
112 |
#%%
|
113 |
|
114 |
+
token_probs: list[tuple[int, float]] = calculate_log_probabilities(model, tokenizer, inputs)
|
115 |
|
116 |
#%%
|
117 |
|