hi as above, tx 4 reading.
I don’t think you’d want that website. Whisper is fairly efficient (even an old GTX can do pretty well at 4x-8x real-time speed), but a website like that would still require pretty expensive cloud GPUs. It’s really not possible to imagine that a website like that would not be data mining you and selling all your audio to advertisers to pay off investors.
Better to buy a GPU and do it yourself. (Good news: it takes like 30 seconds to install)
I run it on fairly short audio files using a i5-3470.
It’s faster than real time I’m pretty sure. But I rarely use it. Adobe has their own thing that’s in beta and it is a lot better.
It’s part of their podcast platform.
tx 4 comment. had no idea gpu processing existed. I found this https://github.com/Const-me/Whisper and works perfectly. hopefully nothing nasty on my win pc 🤞tx 👍
For iOS and MacOS there is Hello Transcribe, a privacy respecting app which uses Whisper
Google Colab can run Whisper, but it can most likely also run on an old, cheap laptop that you have.
second this. use collab. but add ‘!’ to their commands since collab runs shell. also use T4 as ur runtime to support floating point. gl. running whisper locally is way slow, especially without a dedicated gpu.
Just use buzz.