Update app.py
Browse files
app.py
CHANGED
@@ -40,10 +40,18 @@ def transcribe(audio_path):
|
|
40 |
# predicted_ids = torch.argmax(logits, dim=-1)
|
41 |
# transcription = processor.batch_decode(predicted_ids)
|
42 |
|
43 |
-
transcription = model(audio_path,batch_size=1000, generate_kwargs={"task": "transcribe"}, return_timestamps=True)["text"]
|
44 |
|
45 |
# result = pipe(sample)
|
46 |
# 返回转录结果
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
return transcription
|
48 |
|
49 |
|
|
|
40 |
# predicted_ids = torch.argmax(logits, dim=-1)
|
41 |
# transcription = processor.batch_decode(predicted_ids)
|
42 |
|
43 |
+
# transcription = model(audio_path,batch_size=1000, generate_kwargs={"task": "transcribe"}, return_timestamps=True)["text"]
|
44 |
|
45 |
# result = pipe(sample)
|
46 |
# 返回转录结果
|
47 |
+
# return transcription
|
48 |
+
|
49 |
+
|
50 |
+
#------
|
51 |
+
audio_input = processor(audio_path, return_tensors="pt", sampling_rate=16000).input_values
|
52 |
+
with torch.no_grad():
|
53 |
+
predicted_ids = model.generate(audio_input)
|
54 |
+
transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)[0]
|
55 |
return transcription
|
56 |
|
57 |
|