I've been searching around for a clear example on what to run in order to get a forced alignment output for a bunch of wav files that I have. I'm expecting the output to be something that aligns phonemes with start and end times. Been trying to follow along just setting up a training enviorment, but I've yet to come across something that really explains the setup necessary for forced alignment. This article is the closest thing I could find: https://cmusphinx.github.io/wiki/pocketsphinx_pronunciation_evaluation/
I've been searching around for a clear example on what to run in order to get a forced alignment output for a bunch of wav files that I have. I'm expecting the output to be something that aligns phonemes with start and end times. Been trying to follow along just setting up a training enviorment, but I've yet to come across something that really explains the setup necessary for forced alignment. This article is the closest thing I could find:
https://cmusphinx.github.io/wiki/pocketsphinx_pronunciation_evaluation/
I have my entire project up on this github repo:
https://github.com/arcman7/cmu_sphinx
You can check https://montreal-forced-aligner.readthedocs.io/en/latest/
The solution here being not to use cmusphinx?