I’m training tacotron 2 with the default config (with a few audio config tweaks) on one french speaker of the m-ailabs dataset (50h).
I noticed after the 50K iter dip in loss (due to the default gradual training parameters) that while the loss went down and the alignment score went up, the alignment started to make repetitions. Mostly on short sentences but I noticed artefacts on longer sentences as well.
I tried training for a longer time (150k iter) hoping that it would disappear but no luck.
Has anyone encountered the same issues ?