The model is overfitting, validation error falls max to 70 and the training error continues to decrease. I am thinking of changing the hidden layers. from 2048 to 1800. I have played with the learning rate/dropout and it does not make a difference.
I am training on a p2.xlarge on the Common Voice Dataset.
lissyx
((slow to reply) [NOT PROVIDING SUPPORT])
2
lissyx
((slow to reply) [NOT PROVIDING SUPPORT])
4
That’s not going to be enough data to get you anything really useful. I’d say stop the training at the epoch where you start drifting between training and validation loss ? I don’t really understand what you are asking here.
I am trying to build a Speech to Text model using DeepSpeech. I tried training on the Common Voice dataset from scratch but the accuracy is not good. Are you saying I need more data? Should I train further on the pre-trained model?