Information Services banner Edinburgh Research Archive The University of Edinburgh crest

Edinburgh Research Archive >
Centre for Speech Technology Research >
CSTR publications >

Please use this identifier to cite or link to this item: http://hdl.handle.net/1842/3832

This item has been viewed 7 times in the last year. View Statistics

Files in This Item:

File Description SizeFormat
IS081086.PDF250.19 kBAdobe PDFView/Open
Title: Glottal Spectral Separation for Parametric Speech Synthesis
Authors: Renals, Steve
Yamagishi, Junichi
Richmond, Korin
Cabral, Joao P
Issue Date: 2008
Journal Title: In Proc. Interspeech, pages 1829-1832, Brisbane, Australia, September 2008.
Abstract: This paper presents a method to control the characteristics of synthetic speech flexibly by integrating articulatory features into a Hidden Markov Model (HMM)-based parametric speech synthesis system. In contrast to model adaptation and interpolation approaches for speaking style control, this method is driven by phonetic knowledge, and target speech samples are not required. The joint distribution of parallel acoustic and articulatory features considering cross-stream feature dependency is estimated. At synthesis time, acoustic and articulatory features are generated simultaneously based on the maximum-likelihood criterion. The synthetic speech can be controlled flexibly by modifying the generated articulatory features according to arbitrary phonetic rules in the parameter generation process. Our experiments show that the proposed method is effective in both changing the overall character of synthesized speech and in controlling the quality of a specific vowel.
URI: http://hdl.handle.net/1842/3832
Appears in Collections:CSTR publications

Items in ERA are protected by copyright, with all rights reserved, unless otherwise indicated.

 

Valid XHTML 1.0! Unless explicitly stated otherwise, all material is copyright © The University of Edinburgh 2013, and/or the original authors. Privacy and Cookies Policy