I’d like to train a TTS model in an Indigenous language that uses an orthography that doesn’t have a supported phone set. Would I be able to change the training data from the language’s orthography into it’s IPA representation and train it that way? Example below:
if you have IPA representation directly in the dataset, then just use character based training but give IPA characters as the char list (not phoneme char list).
// PHONEMES
"phoneme_cache_path": "mozilla_us_phonemes_2_1", // phoneme computation is slow, therefore, it caches results in the given folder.
"use_phonemes": false, // use phonemes instead of raw characters. It is suggested for better pronounciation.
"phoneme_language": "en-us", // depending on your target language, pick one from https://github.com/bootphon/phonemizer#languages