Replies: 2 comments 1 reply
-
Read there -> Some problems with large-v3
That's normal.
Yes, test with same r160.3 version. |
Beta Was this translation helpful? Give feedback.
1 reply
-
Looks like a general consensus is that Whisper-v3 Hallucinations on Real World Data Whisper's original large-v3 tread: openai/whisper#1762 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I needed to create subtitles for a 90 min documentary so thought it would be a decent test to see if the new v3-large model had any significant improvements over the v2-large.
This post is not about an issue though, just my findings.
I upgraded to
Whisper-Faster_r160.3_windows
and then ran these commands against the same video file.whisper-faster "%vidname%" -l=en -m=large-v2 -ct=int8_float32 -o=source --verbose=true -bs=1
using the v2-large model
and
whisper-faster "%vidname%" -l=en -m=large-v3 -ct=int8_float32 -o=source --verbose=true -bs=1
using the v3-large model
I compared the two outputs in Subtitle Edit and there were plenty of differences, too many to list here.
Is this a fair comparison or should I have used the old binary (r160) with the v2-large model and then the new binary with the v3-large model?
Just remembered I can change the verbose command to just -v
Beta Was this translation helpful? Give feedback.
All reactions