Our company Iara Health provides a system to aid radiologists in the writing of medical reports in the portuguese (BR) language. All our system is built over DeepSpeech, running locally on the user’s computer.
In the video above, you can see our portal being able to recognize commands (like loading a template), and handle punctuation, acronym and abbreviations. Our system eases the work of radiologists, making them produce more in less times.
We want to thank Mozilla for DeepSpeech.
We’re using DeepSpeech in tarteel.io to recognize Quran recitation and correct people’s mistakes on a word-by-word level!
The Quran is the Muslim’s holy book and we are instructed to recite it with "Tarteel, which translates best to “slow measured rhythmic tones”.
Muslims who try to memorize the Quran sometimes struggle to find an instructor someone to correct them. The Tarteel platform provides a “Quran Companion” they can use to recite to when they don’t have any to correct them.
I’m French, autodidact…
I’m a robotician (passion only), and I work on social interactions between robots and humains.
Vocal interactions are imperatives…
Thanks to Deepspeech.
I made a tuto to help each other create it own model.
I’m a computer scientist student intrigued and interested by data science, who want to deploy a spanish model Speech-To-Text that can be easily integrated, easy to use and can be flexible. I actually find out that in some cases my DeepSpeech spanish model can outperform Google and IBM Watson Speech-To-Text models in real situations with just 450 hours for train dev and test.
We are Vivoce, a startup using DeepSpeech to detect pronunciation errors and help users improve their accents for language learning.
I hope to use Deep Speech for African Languages soon in the future
What languages are you interested in?
We use Mozilla DeepSpeech for voicemail transcription in FusionPBX via our DeepSpeech Frontend and some code we upstreamed into FusionPBX to add support for custom STT providers. Our users find transcriptions quite useful, with Mozilla DeepSpeech serving them with transcriptions since August 2018!
I would love to collaborate with @Gabriel_Guedj and others to build tuned models and deeper integrations in the telephony space. Feel free to reach out, I am @dan:whomst.online in #machinelearning:mozilla.org on Matrix.
A student at Dalarna University, Sweden, trying to use deepspeech to train a model for the Somali language
Could you share your deepspeech spanish model? I am quit far of getting so many hours of transcribed data in spanish to train my own spanish model.
I am building an online video consultation tool for primary care (eg. GPs, family medicine) and using DeepSpeech to enable transcription of the meeting. I am currently working on integrating DeepSpeech with Jitsi (an open-source videoconferencing tool).
Potentially once the integration works, the transcription generate by DeepSpeech can be used to run ML algorithms over it to generate suggestions for doctor.
Maybe you could talk to @bernardohenz? He might have advice on adapting DeepSpeech to the medical domain.
I am building a small personal companion that has the following features:
- It does not require Wifi or internet connections.
- It will be powered with less than 5 volts.
- It will auto recharge with light.
- If you talk to it, it will talk back to you intelligently with a voice.
- The companion can talk to you an hour a day for 15 years, with no overlap.
- The companion will have an extensive memory and will learn from you.
- It will be no larger than an apple or a large deck of cards.
- The per-unit cost will be less than $40 USD.
I am doing this because I can do it and I want to create something interesting.
I am trying to build a model that understands german dialects.
Tēnā koutou kātoa!
Te Hiku Media is a Māori organization based in Aotearoa (New Zealand). Our purpose is to preserve and promote te reo Māori, the indigenous language of Aotearoa.
We’ve been using DeepSpeech since May 2018. We found it worked pretty well for te reo Māori, which is an oral language that was phonetically transcribed in the 19th century. We have an API running with a w.e.r. about 10%, and we use this API to help us speed up the transcription of native speaker (L1) recordings.
We’re running DeepSpeech within a Docker on a p2.xlarge DeepLearning Ubuntu AMI in AWS. This is behind a Load Balancer and an Auto Scaling Group which allows us to bid for those old p2 instances at a relatively affordable cost (we spend about USD$1000/month to keep the API available 24/7). We use FastAPI to load and run DS in python. We’ve got a Django instance, koreromaori.io, between this API and what the end-user sees (there’s reasons for this) but we’re in the process of figuring out how to more efficiently deploy DeepSpeech. Keen to hear what others are doing.
For many reasons, some of which you can learn about in this Greater Than Code podcast and this Te Pūtahi podcast, we’ve built our own Django based web-app to collect te reo Māori data, koreromaori.com [corpora]. We started this around the same time as the Common Voice project and because my experience was in Django, it made more sense for us to work on corpora. Of course in hindsight there are many more reasons why using your own platform for data collection can be useful. For example, all the data is available through an API which helps us when it comes time to train models. We also label data specifically to our context, such as whether a speaker is “native” (L1 vs. L2) or whether pronunciation or intonation is correct. Finally, for indigenous languages, it’s often more appropriate for the data to remain with the community rather than being put in the public domain.
Since we were able to train a DS model early on, we use this model to help us “machine review” data. We also built our own transcription tool, Kaituhi, to help us with transcribing our audio archives. It’s kind of like the BBC React Transcript Editor which I found out about AFTER we started work on Kaituhi. We use koreromaori.io to provide automated transcriptions for Kaituhi, and we’re hoping to add word level confidences to the transcriptions to speed up the review process (the confidences are in the DS api, they’re just not exposed yet in our API).
The main reason why our initiative to build STT for a language that nearly went extinct was successful is because of the community around the language. We cannot forget the hard work done by so many during the 20th century to make te reo Māori (and other indigenous languages) a living language once again. I think Mozilla’s done a good job with building a community around Common Voice. If you’re someone working on language tools for non mainstream languages, building trust with the right community is critical to solving the data problem. Also understanding that there’s a level of respect and responsibility that comes with access to data is important.
We wouldn’t be where we are today in terms of the technology if it wasn’t for DeepSpeech. So a big thank you to Mozilla and the DeepSpeech team and all of you who are an active part of the DS and common voice community!
Having posted earlier in this thread I just wanted to update on details of the scripts we’ve developed at Bangor University that bring together the various features of DeepSpeech, along with CommonVoice data, and provides a complete solution for producing models and scorers for Welsh language speech recognition. They may be of interest to any other users of DeepSpeech that are working with a similarly lesser resourced language to Welsh.
- are based on DeepSpeech 0.7.4
- make use of DeepSpeech’s Dockerfiles (so setup and installation is easier).
- train with CommonVoice data
- utilize transfer learning
- with some additional test sets and corpora, produce optimized scorers/language models for various applications
- exports models with metadata
The initial README describes how to get started.
We’d like to share also the models that are produced from these scripts which can be found at https://github.com/techiaith/docker-deepspeech-cy/releases/tag/20.06
At the moment these models are used in two prototype applications which the Welsh speaking community can install and try, namely a Windows/C# based transcriber and an Android/iOS voice assistant app called Macsen. Source code for these applications using DeepSpeech can also be found on GitHub.
We are immensly grateful to Mozilla for creating the Common Voice and DeepSpeech projects.
Hello, we are polish comapny that produces software for logistics purposes. Our R&D team is working on “hands free” system/device for workers in warehouses. We have Android app that is displayed on AR glasses and DeepSpeech will be responsible for ASR (not full speech; only couple hundred of commands) to control logistic processes by speech (barcodes and numbers dictation, commands like: confirm, cancel, print, go back, go up, go down etc.). For now we are mainly focusing on polish speech recognition.
I wanted to say that DeepSpeech is awsome and to DS team: your work allows many researches and comapnies around the world make great, new tech. Thank You for that!! Also DS community is very helpful I am very greatful for Your work and involvement. Thank you once again.
Hi, I did build a Voice Assistant for multiple languages which is called Jaco-Assistant.
It can run completely offline and the code is open source.
There’s also a skill store where you can share and download skills. Jaco automatically generates domain specific language models out of the installed skills to achieve high detection accuracy.
In a benchmark Jaco did perform better in understanding speech commands than solutions from Amazon and Google (but it has some problems with loud background noise):
The assistant uses a highly modular and containerbased architecture which also allows it to easily reuse parts of it for other projects. The skills run in containers too, which has the great benefit that developers are not restricted if they have special requirements for their own skills.
Currently Jaco understands English and German and runs on linux computers. It is extendable to further languages and I hope I can get it to run on Raspi in the next time too.
You can find the project here:
At this time it’s working and usable, but far from perfect and I’m happy about everyone wanting to contribute to the project by translating it to other languages, solving issues or creating cool new skills:)
Definitely in the wild, this one. I am building a sexbot companion as a personal project. In-head raspberry pi captures 4 channels of audio, does some basic noise reduction, classification, segmentation, and sends signal over wifi to be processed by DeepSpeech on a desktop gpu, which returns the result. I have that much working. Currently collecting more data to train the models. It’s mostly gibberish now, but I expect in a few more months and a lot more GBs of wav files the words will flow. Thank you for making DeepSpeech.