I’m a bit confused with the dropout parameter, in the 0.9.0 release description it is said that the dropout value used for training is 0.4, but there are 5 more dropout parameters values.
In the release checkpoint flags file, theses values a described :
– dropout_rate=0.05
–dropout_rate2=0.05
–dropout_rate3=0.05
–dropout_rate4=0.0
–dropout_rate5=0.0
–dropout_rate6=0.05
Does it mean that you used :
– dropout_rate=0.4
–dropout_rate2=0.4
–dropout_rate3=0.4
–dropout_rate4=0.0
–dropout_rate5=0.0
–dropout_rate6=0.4
last time, I made a optuna script for dropout parameter optimisation only for the first value (–dropout_rate), but i’m not sure i’m right.
Does it make sense to look for differents dropout values for each layer?
@lissyx, @reuben this is indeed a bit strange. Do you know why the flags.txt in the checkpoint for English lists these values, is this from a continued training with lower learning rate?
It would be great to know if high droputs like 0.6 have a meaningful impact for DeepSpeech models. I have only tried up to .5 and that made our model worse.