Snapshot
Browse files
app.py
CHANGED
|
@@ -107,11 +107,11 @@ model, tokenizer = load_model_and_tokenizer(model_name, device)
|
|
| 107 |
#%%
|
| 108 |
|
| 109 |
input_text = "He asked me to prostrate myself before the king, but I rifused."
|
| 110 |
-
|
| 111 |
|
| 112 |
#%%
|
| 113 |
|
| 114 |
-
token_probs: list[tuple[int, float]] = calculate_log_probabilities(model, tokenizer,
|
| 115 |
|
| 116 |
#%%
|
| 117 |
|
|
|
|
| 107 |
#%%
|
| 108 |
|
| 109 |
input_text = "He asked me to prostrate myself before the king, but I rifused."
|
| 110 |
+
inputs: BatchEncoding = tokenize(input_text, tokenizer, device)
|
| 111 |
|
| 112 |
#%%
|
| 113 |
|
| 114 |
+
token_probs: list[tuple[int, float]] = calculate_log_probabilities(model, tokenizer, inputs)
|
| 115 |
|
| 116 |
#%%
|
| 117 |
|