A modular architecture for articulatory synthesis from gestural specification

This paper proposes a modular architecture for articulatory synthesis from a gestural specification comprising relatively simple models for the vocal tract, the glottis, aero-acoustics, and articulatory control. The vocal tract module combines a midsagittal statistical analysis articulatory model, d...

Full description

Bibliographic Details
Main Authors: Alexander, R. (Author), Narayanan, S. (Author), Sorensen, T. (Author), Toutios, A. (Author)
Format: Article
Language:English
Published: Acoustical Society of America 2019
Subjects:
Online Access:View Fulltext in Publisher
Description
Summary:This paper proposes a modular architecture for articulatory synthesis from a gestural specification comprising relatively simple models for the vocal tract, the glottis, aero-acoustics, and articulatory control. The vocal tract module combines a midsagittal statistical analysis articulatory model, derived by factor analysis of air-tissue boundaries in real-time magnetic resonance imaging data, with an α β model for converting midsagittal section to area function specifications. The aero-acoustics and glottis models were based on a software implementation of classic work by Maeda. The articulatory control module uses dynamical systems, which implement articulatory gestures, to animate the statistical articulatory model, inspired by the task dynamics model. Results on synthesizing vowel-consonant-vowel sequences with plosive consonants, using models that were built on data from, and simulate the behavior of, two different speakers are presented. © 2019 Acoustical Society of America.
ISBN:00014966 (ISSN)
DOI:10.1121/1.5139413