Good morning,
I have a basic question about using DeepSpeech.
I have the following project. I am setting up an Ubuntu 20.04 LTS server (headless) and want to install DeepSpeech there.
The task of the server should be to send audio files (vmtl ogg, opus) from a program via network to the server and let it process them. I do not know which language the audio files contain. As return I want to have the recognized text. I want to continue working with this in the program.
How can I implement this? The program I use to deliver the data has a Python interface.
Do I have to download different language models, if so where can I find them?
Greetings from Berlin
Bruno