Hopefully this is an easy / quick to answer question
Regarding the language model included with the 0.5.1 release could someone from the team confirm that it was trained with the data / process here: DeepSpeech/data/lm at master 路 mozilla/DeepSpeech 路 GitHub ?
I just wanted to be sure as I鈥檓 looking at extending the LM with some particular text for my application (eg names not present in LibriSpeech) and wanted to know I was starting from the correct base.
@kdavis There was also talk of using only the top 10k - 50k words here - has that been implemented yet or is it still a work in progress? Seemed like it had potential.