However, I’m planning on training only on the last FC layer and the RNN layer after unfreezing them for a few epoch and then train on the complete layers for the next few epochs.
Is there a possibility to do the same, i.e. to unfreeze the last layers and then the remaining layers?
Assuming the data feeding your model is mono, 16-bit PCM @ 16kHZ and represents appropriate audio/transcription pairs. Up-sampling may cause issues.
If you run this, it should create three files / epoch and three logging files in your checkpoint_dir directory.
at each epoch:
-model.ckpt-#####.data-#####-of-#####
-model.ckpt-#####.index
-model.ckpt-#####.meta
for logging:
-checkpoint
-events.out.tfevents…
-graph.pbtxt
after you run your 3 epochs, you’ll see that the checkpoint file lists all the unfrozen graph file names you can use to restart training. use the one where loss was lowest. You should then be able to run your script again, with a lower learning rate, to fine tune your most recent checkpoint. (That worked for me anyway)
I believe there is a warning for it but make sure you empty the folder of checkpoint files you don’t want to use.