Hi, higher batch size results in OOM errors for me the maximum I could go was unfortunately, 1. Just curious how would a higher batch size lead to better results?
For the learning rate, I was using 0.00005 but it led to overfitting and hence early stopping at epoch 25 so I was thinking that I should lower the learning rate?
Yep, will try a dropout rate of 0.3.