This site will have limited functionality while we undergo maintenance to improve your experience. If an article doesn't solve your issue and you want to ask a question, we have our support community waiting to help you at @FirefoxSupport on Twitter and/r/firefox on Reddit.

Search Support

Avoid support scams. We will never ask you to call or text a phone number or share personal information. Please report suspicious activity using the “Report Abuse” option.

Learn More

What is the current state of Web Speech API?

  • Amul benn tontu
  • 0 am na jafe-jafe bii
more options

I'm curious about the current state of Web Speech API. This is a W3C specification that was implemented in both Firefox and Chrome and includes Automatic Speech Recognition.

There is a demo page at https://mdn.github.io/dom-examples/web-speech-api/speech-color-changer/

To use this in Firefox, I have to change two settings in about:config to True: media.webspeech.recognition.enable media.webspeech.recognition.force_enable

There used to be a setting that could be used to set a specific endpoint for the actual STT processing, but that doesn't show up in my list of settings anymore. I guess I could add it, but have no idea if it would be respected by the browser or what the endpoint should look like. Anyway, whatever endpoint Firefox is currently using does not seem to be working anymore, so I just get an "Error occurred in recognition: network" error.

In Chrome it works, but most likely is sending my raw audio to Google Cloud STT to do the actual transcribing. I'd like to be able to control the endpoint so I can use a local service connected to VOSK or Whisper or some other locally hosted STT engine.

I'm thinking about implementing something using WebRTC, but if anyone is working along similar lines, I'd like to join that project rather than creating my own.

Is anyone working on this?

I'm curious about the current state of Web Speech API. This is a W3C specification that was implemented in both Firefox and Chrome and includes Automatic Speech Recognition. There is a demo page at https://mdn.github.io/dom-examples/web-speech-api/speech-color-changer/ To use this in Firefox, I have to change two settings in about:config to True: media.webspeech.recognition.enable media.webspeech.recognition.force_enable There used to be a setting that could be used to set a specific endpoint for the actual STT processing, but that doesn't show up in my list of settings anymore. I guess I could add it, but have no idea if it would be respected by the browser or what the endpoint should look like. Anyway, whatever endpoint Firefox is currently using does not seem to be working anymore, so I just get an "Error occurred in recognition: network" error. In Chrome it works, but most likely is sending my raw audio to Google Cloud STT to do the actual transcribing. I'd like to be able to control the endpoint so I can use a local service connected to VOSK or Whisper or some other locally hosted STT engine. I'm thinking about implementing something using WebRTC, but if anyone is working along similar lines, I'd like to join that project rather than creating my own. Is anyone working on this?

You must log in to your account to reply to posts. Please start a new question, if you do not have an account yet.