I wanted to train the provided deepspeech 0.3 checkpoints with my own recorded audio files.
For now, I want to train only the last few layers (like fully connected layers and last few RNN layers) while freezing the rest of layers. What changes are needed in code to do that.
Also, for an advice, Is it better to train only the last few layers or the entire network, provided your dataset is just another accent of English language itself with few extra words that were not earlier in vocabulary.
(3) train with DeepSpeech.py, making sure the paths match up to your local data. I’d try replacing the top layer and maybe the second layer. Going down deeper probably won’t make sense for just accent-transfer.
Are the above two parameters correct and enough, for the purpose of transfer learning by unfreezing only the last two layers
And, the model seems to have converged quickly and is performing well with my test dataset. Though, I’ve a feeling that it has over-fitted, because it works well with similar sentences that I’ve added to vocabulary (though from a completely new speaker) , but doesn’t work well if the same speaker talks entirely new sentences.
Any suggestion on what I can do to improve this.
If we look into Deepspeech’s architecture, it has 3 FC layers, then one BiLSTM layer followed by one FC layer.
So, If I am training last two layers using the Transfer-learning branch, that @josh_meyer has mentioned, would that modify the weights of even the BiLSTM layer ('coz that’s the last-but-one layer as per the arch.)
Also @josh_meyer@lissyx@kdavis,
Can you suggest any better training method such that it retains the American accent but performs well on other accents too.
Problem I am facing is that, If I train the last one/two layers with just Indian accent, it’s forgetting the previous accents.
Hey @rpratesh how has transfer learning worked? I’m stuck using checkpoint and training henceforth. That is causing some problems. Looking forward to hearing from you.
Sir, I am trying to train the existing model from released checkpoints-0.4.1. I have my own data and want to continue training from these checkpoints.
But my training shows that it is resuming from different epoch number like 10232 for one dataset, 45000 for another dataset, 1832 for another dataset
(actually it should have started from 31 as the model is trained for 30 epochs as written on releases page)
Any idea about this sir?
And also how to use your transfer learning repository? can you please elaborate?