Must be doing something stupid here. Trying to run one of the released multispeaker models:
when I checkout 2136433 and try a test with synthesize.py I still get a bunch of layer size mistmatches.
size mismatch for decoder.attention_rnn.weight_ih: copying a param with shape torch.Size([4096, 1024]) from checkpoint, the shape in current model is torch.Size([4096, 896])
(seems to be a delta of 128)…tried a few other commits around the time of the model and no luck. Can anyone point out my mistake?
EDIT: I also noticed that the link to Universal WaveGrad on the wiki doesn’t point to a google drive link like the rest of the vocoder models. Would like to try it if possible.