I’ve been playing with Deep Speech the past couple of days with the task of transcribing University lectures, that were presented in English by German lecturers. I tried to use the models provided with the 0.6.1 release (pb, pbmn and tflite). So far the transcriptions have been far from accurate. What is the best way to tackle this issue?
Can I adjust some of the parameters to try to get better results?
Should a train a new model based on germans speaking English?
Where can I find additional information about the following parameters:
–beam_width BEAM_WIDTH Beam width for the CTC decoder
–lm_alpha LM_ALPHA Language model weight (lm_alpha)
–lm_beta LM_BETA Word insertion bonus (lm_beta)
If I should create a new model, where could I get the data? Is it possible to filter only german speaking English from the common voice initiative?
Thank you in advance!