Update app.py
Browse files
app.py
CHANGED
@@ -27,14 +27,15 @@ model = pipe
|
|
27 |
# Evaluate the model
|
28 |
# model.eval()
|
29 |
#print("model.eval ", model.eval())
|
30 |
-
|
31 |
-
audio = batch["audio"]
|
32 |
-
input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
|
33 |
-
batch["reference"] = processor.tokenizer._normalize(batch['text'])
|
34 |
-
|
35 |
-
|
36 |
-
with torch.no_grad():
|
37 |
-
|
|
|
38 |
transcription = processor.decode(predicted_ids)
|
39 |
batch["prediction"] = processor.tokenizer._normalize(transcription)
|
40 |
return batch
|
|
|
27 |
# Evaluate the model
|
28 |
# model.eval()
|
29 |
#print("model.eval ", model.eval())
|
30 |
+
def map_to_pred(batch):
|
31 |
+
audio = batch["audio"]
|
32 |
+
input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
|
33 |
+
batch["reference"] = processor.tokenizer._normalize(batch['text'])
|
34 |
+
|
35 |
+
|
36 |
+
with torch.no_grad():
|
37 |
+
predicted_ids = model.generate(input_features.to("cuda"))[0]
|
38 |
+
|
39 |
transcription = processor.decode(predicted_ids)
|
40 |
batch["prediction"] = processor.tokenizer._normalize(transcription)
|
41 |
return batch
|