Replies: 1 comment
-
I think you are using the master version of openai/whisper which changed the behavior of We prepared the same change in #225 but I wanted to wait for OpenAI to cut a new version before merging it. If you use this branch you should get the same transcription as openai-whisper. However, note that faster-whisper has a more robust way to deal with silence with the argument |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I've been working with an audio clip in spanish that contains speech along large silence segments. Faster-whisper transcription is failing, if the 'no_speech_threshold' is set to 0.6, its default value. I've tested, large-v2 and medium obtaining a similar behavior. However, if you increment it to 1.0, the transcription is correct.
I've compared the results with Whisper and the transcription is correct using the default values.
If anyone wants to play with it, you could download the audio clip from here:
https://www.dropbox.com/s/1r4cjvdkqcc0nr1/test_audio.wav?dl=0
Beta Was this translation helpful? Give feedback.
All reactions