Mozilla Voice STT in the Wild!

At Bangor University we’re using DeepSpeech (and Welsh CommonVoice) within a Welsh language digital assistant project. Based on Flutter, our app for Android and iOS can respond to simple questions regarding weather, news, time, Welsh language wikipedia and Welsh language music on Spotify, thanks to a hosted DeepSpeech server.

We’re also evaluating DeepSpeech’s recent work on transfer learning for larger domains such as dictation and captioning. Results so far have been very exciting for a lesser resourced language like Welsh. It would be awesome to see transfer learning supported in main releases of DeepSpeech.

Thank you so much Mozilla for DeepSpeech!!

7 Likes

My name is Dan, and I’m working on a voice + motion control system called Jaxcore. I’m using DeepSpeech to add speech recognition to control computers / home theaters / and smart home devices.

The upcoming open source desktop app will have DeepSpeech built-in and voice commands for doing things like controlling your computer mouse, typing on your computer, controlling media players.

You’ll be able to write web games that have speech recognition (using only client-side Javascript):

The speech recognition libraries are all modular and can be used individually:

3 Likes

I use Deepspeech as a local STT engine for mycroft.ai. It’s called via the deepspeech-server tool, so I can have multiple devices access it. Runs on a desktop cpu quite nicely. The audio is saved from mycroft, which I can use for fine tuning down the road sometime.

There’s a fine-tuned model and I use some filtering to help. Accuracy is good, latency is good.

3 Likes

This one? https://github.com/MainRo/deepspeech-server

1 Like

That’s the one. Used that for a while now.

2 Likes

Our company Iara Health provides a system to aid radiologists in the writing of medical reports in the portuguese (BR) language. All our system is built over DeepSpeech, running locally on the user’s computer.

In the video above, you can see our portal being able to recognize commands (like loading a template), and handle punctuation, acronym and abbreviations. Our system eases the work of radiologists, making them produce more in less times.

We want to thank Mozilla for DeepSpeech.

11 Likes

We’re using DeepSpeech in tarteel.io to recognize Quran recitation and correct people’s mistakes on a word-by-word level!
The Quran is the Muslim’s holy book and we are instructed to recite it with "Tarteel, which translates best to “slow measured rhythmic tones”.
Muslims who try to memorize the Quran sometimes struggle to find an instructor someone to correct them. The Tarteel platform provides a “Quran Companion” they can use to recite to when they don’t have any to correct them.

8 Likes

Hello.
I’m French, autodidact…

I’m a robotician (passion only), and I work on social interactions between robots and humains.
Vocal interactions are imperatives…
Thanks to Deepspeech.

I made a tuto to help each other create it own model.:wink:

4 Likes

I’m a computer scientist student intrigued and interested by data science, who want to deploy a spanish model Speech-To-Text that can be easily integrated, easy to use and can be flexible. I actually find out that in some cases my DeepSpeech spanish model can outperform Google and IBM Watson Speech-To-Text models in real situations with just 450 hours for train dev and test.

4 Likes

We are Vivoce, a startup using DeepSpeech to detect pronunciation errors and help users improve their accents for language learning.

5 Likes

I hope to use Deep Speech for African Languages soon in the future

2 Likes

What languages are you interested in?

We use Mozilla DeepSpeech for voicemail transcription in FusionPBX via our DeepSpeech Frontend and some code we upstreamed into FusionPBX to add support for custom STT providers. Our users find transcriptions quite useful, with Mozilla DeepSpeech serving them with transcriptions since August 2018!

I would love to collaborate with @Gabriel_Guedj and others to build tuned models and deeper integrations in the telephony space. Feel free to reach out, I am @dan:whomst.online in #machinelearning:mozilla.org on Matrix.

Mozilla DeepSpeech is awesome, I really appreciate all the hard work @kdavis, @reuben, @lissyx, and other contributors have put in over the years to build this!

5 Likes

A student at Dalarna University, Sweden, trying to use deepspeech to train a model for the Somali language

2 Likes

Hi victornoriega7,
Could you share your deepspeech spanish model? I am quit far of getting so many hours of transcribed data in spanish to train my own spanish model.
Thanks
ana

I am building an online video consultation tool for primary care (eg. GPs, family medicine) and using DeepSpeech to enable transcription of the meeting. I am currently working on integrating DeepSpeech with Jitsi (an open-source videoconferencing tool).

Potentially once the integration works, the transcription generate by DeepSpeech can be used to run ML algorithms over it to generate suggestions for doctor.

1 Like

Maybe you could talk to @bernardohenz? He might have advice on adapting DeepSpeech to the medical domain.

I am building a small personal companion that has the following features:

  1. It does not require Wifi or internet connections.
  2. It will be powered with less than 5 volts.
  3. It will auto recharge with light.
  4. If you talk to it, it will talk back to you intelligently with a voice.
  5. The companion can talk to you an hour a day for 15 years, with no overlap.
  6. The companion will have an extensive memory and will learn from you.
  7. It will be no larger than an apple or a large deck of cards.
  8. The per-unit cost will be less than $40 USD.

I am doing this because I can do it and I want to create something interesting.

3 Likes

I am trying to build a model that understands german dialects.

3 Likes