TY - JOUR
T1 - Articulatory speech synthesis based upon fluid dynamic principles
AU - Huang, Jun
AU - Levinson, Stephen
AU - Davis, Donald
AU - Slimon, Scott
PY - 2002
Y1 - 2002
N2 - In this paper, an articulatory speech synthesizer based on fluid dynamic principles is presented. The key idea is to devise a refined speech production model based on the most fundamental physics of the human vocal apparatus. Our articulatory synthesizer essentially contains two parts: a vocal fold model which represents the excitation source and a vocal tract model, which describes the positions of articulators. First, we propose a combined minimum error and minimum jerk criterion to estimated the moving vocal tract shapes during speech production. Second, we propose a nonlinear mechanical model to generate the vocal fold excitation signals. Finally, a computational fluid dynamics (CFD) approach is used to solve the Reynolds-averaged Navier-Stokes (RANS) equations, which are the governing equations of speech production inside vocal apparatus. Experimental results show that our system can synthesize intelligible continuous speech sentences while naturally handling the co-articulation effects during speech production.
AB - In this paper, an articulatory speech synthesizer based on fluid dynamic principles is presented. The key idea is to devise a refined speech production model based on the most fundamental physics of the human vocal apparatus. Our articulatory synthesizer essentially contains two parts: a vocal fold model which represents the excitation source and a vocal tract model, which describes the positions of articulators. First, we propose a combined minimum error and minimum jerk criterion to estimated the moving vocal tract shapes during speech production. Second, we propose a nonlinear mechanical model to generate the vocal fold excitation signals. Finally, a computational fluid dynamics (CFD) approach is used to solve the Reynolds-averaged Navier-Stokes (RANS) equations, which are the governing equations of speech production inside vocal apparatus. Experimental results show that our system can synthesize intelligible continuous speech sentences while naturally handling the co-articulation effects during speech production.
UR - http://www.scopus.com/inward/record.url?scp=0036294868&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=0036294868&partnerID=8YFLogxK
U2 - 10.1109/ICASSP.2002.5743750
DO - 10.1109/ICASSP.2002.5743750
M3 - Article
AN - SCOPUS:0036294868
SN - 1520-6149
VL - 1
SP - 445
EP - 448
JO - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
JF - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
ER -