You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've compared the transcription speed on AMD Ryzen 5950X CPU with and without batching. faster-whisper is running inside a single docker container on a 5 minutes mp4 file. The turbo model is cached locally.
v1.1.0 is few seconds slower without batching and batching doesn't improve the speed.
Results:
Version
Precision
Beam size
Time
v1.0.3 (cpu_threads=4)
int8
1
55s
v1.1.0 (cpu_threads=4)
int8
1
1m2s
v1.1.0 (cpu_threads=4, batch_size=4)
int8
1
55s
Without batching:
fromfaster_whisperimportWhisperModelmodel_size="turbo"model=WhisperModel(model_size, device="cpu", compute_type="int8", cpu_threads=4)
segments, info=model.transcribe("test.mp4", beam_size=1, vad_filter=True, task="transcribe")
print("Detected language '%s' with probability %f"% (info.language, info.language_probability))
forsegmentinsegments:
print("[%.2fs -> %.2fs] %s"% (segment.start, segment.end, segment.text))
With batching
fromfaster_whisperimportWhisperModel, BatchedInferencePipelinemodel_size="turbo"model=WhisperModel(model_size, device="cpu", compute_type="int8", cpu_threads=4)
batched_model=BatchedInferencePipeline(model=model)
segments, info=batched_model.transcribe("test.mp4", beam_size=1, vad_filter=True, task="transcribe", batch_size=4)
print("Detected language '%s' with probability %f"% (info.language, info.language_probability))
forsegmentinsegments:
print("[%.2fs -> %.2fs] %s"% (segment.start, segment.end, segment.text))
The text was updated successfully, but these errors were encountered:
Feel free to reference this repository — it's designed for users with CPU-only. With an Intel i3-12300 CPU and Whisper-Large-V3-Turbo, it takes just 15 minutes to generate subtitles for a 2-hour movie. Additionally, the faster SenseVoiceSmall model can transcribe the same movie in just 7 minutes. This tool also includes a VAD and denoiser to improve subtitles accuracy.
I've compared the transcription speed on AMD Ryzen 5950X CPU with and without batching. faster-whisper is running inside a single docker container on a 5 minutes mp4 file. The turbo model is cached locally.
v1.1.0 is few seconds slower without batching and batching doesn't improve the speed.
Results:
cpu_threads=4
)cpu_threads=4
)cpu_threads=4, batch_size=4
)Without batching:
With batching
The text was updated successfully, but these errors were encountered: