i successfully installed deepspeech via python on my windows 10. i tested the pretrained english models with the provided sample audio and it worked properly.
so i downloaded the pretrained models (pbmm and scorer) for german from jaco-assistant gitlab repo. i used the same command as mentioned in the original mozilla
deepspeech github repo, i just changed the used pbmm and scorer, looking like this:
deepspeech --model output_graph_de.pbmm --scorer kenlm_de.scorer --audio random.wav
the wav data was a short 30 sec audio from youtube, form a german cartoon. i converted it to wav with 16bits and 16k sample rate. the inference output of 30sec of spoken audio were only 5 random german words which had nothing to do with the audio file.
did i do something wrong? im sorry if this is a noob question.
EDIT: well the audio was maybe a bit too long. since i read that too long audio files wont work properly, so i recorded my own voice (saying something like: hello my friends) and converted it to the right format. now the inference is… blank. lol. is there any step i needed to do before i use different pre trained pbmm and scorer models?