Assigning weights to certain words while training DeepSpeech Model

Hi, higher batch size results in OOM errors for me the maximum I could go was unfortunately, 1. :confused: Just curious how would a higher batch size lead to better results?

For the learning rate, I was using 0.00005 but it led to overfitting and hence early stopping at epoch 25 so I was thinking that I should lower the learning rate?

Yep, will try a dropout rate of 0.3.