WhisperD: linux voice-to-text using OpenAI whisper-1 transcription

5 points by icholy


I wrote this as an exercise to learn how to use ioctl & input devices, but I like how it turned out! It does have a hard dependency on pipewire though.

acatton

From the title, I thought this was using whisper with whisper.cpp.

But I just misunderstood, it looks this is just streaming the microphone to OpenAI servers. This doesn’t diminish the work of gluing everything together.

That could be a cool next step though, to make it local only since whisper.cpp has some go bindings: https://github.com/ggml-org/whisper.cpp/tree/master/bindings/go

david_chisnall

The ‘suggest tags’ button seems to have gone, but this should be using the vibecoding tag, not ai. The description of the ai tag says:

Developing artificial intelligence, machine learning. Tag AI usage only with vibecoding.