Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This is a cool project. I’ve been very happy with whisper as an alternative to otter; it works better and solves real problems for me.

I feel compelled to point out whisper.cpp. It may be cheaper for the author but is relevant for others.

I was running whisper on a gtx 1070 to get decent performance; it was terribly slow on M1 Mac. Whisper.cpp has comparable performance to the 1070 while running on M1 CPU. It is easy to build and run and well documented.

https://github.com/ggerganov/whisper.cpp

I hope this doesn’t come off the wrong way, I love this project and I’m glad to see the technology democratized. Easily accessible high-quality transcription will be a game changer for many people and organizations.



Thanks for sharing! I was looking for a M1 solution weeks ago snd couldn‘t find any working one. Will try that one now! Looking around for servers with GPUs etc. resulted in stopping me at playing around with it as i got overwhelmed with options.


How long would whisper.cpp take to transcribe 2 hours of audio on M1?


Not sure about M1, but on the Macbook Pro 14" with an M1 Max using 8 threads I transcribed a 44 minute podcast in 16 minutes. So about 3x "real time" speed.


What model are you using? I guess large, as my M1 Max takes about 1.4 min for a 4 min file (35% of total time)?


Yep, large model.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: