I tried this in Fedora/Wayland previously, and it seems to work in most applications. It uses “VOSK” models which the GUI can download automatically - you just pick your language and desired model size when setting it up.
When I was exploring this a few months ago, I noticed speech recognition models have moved on quite recently (e.g. https://github.com/openai/whisper which can be run locally) but didn’t see anything integrating it into an input-method like the above.
https://github.com/Manish7093/IBus-Speech-To-Text
I tried this in Fedora/Wayland previously, and it seems to work in most applications. It uses “VOSK” models which the GUI can download automatically - you just pick your language and desired model size when setting it up.
When I was exploring this a few months ago, I noticed speech recognition models have moved on quite recently (e.g. https://github.com/openai/whisper which can be run locally) but didn’t see anything integrating it into an input-method like the above.