This paper describes a speaker-adaptive HMM-based speech synthesis system. The new system, called ldquoHTS-2007,rdquo employs speaker adaptation (CSMAPLR+MAP), feature-space adaptive training, mixed-gender modeling, and full-covariance modeling using CSMAPLR transforms, in addition to several other techniques that have proved effective in our previous systems. Subjective evaluation results show that the new system generates significantly better quality synthetic speech than speaker-dependent approaches with realistic amounts of speech data, and that it bears comparison with speaker-dependent approaches even when large amounts of speech data are available. In addition, a comparison study with several speech synthesis techniques shows the new system is very robust: It is able to build voices from less-than-ideal speech data and synthesize good-quality speech even for out-of-domain sentences.
%0 Journal Article
%1 Yamagishi2009a
%A Yamagishi, Junichi
%A Nose, Takashi
%A Zen, Heiga
%A Ling, Zhen-Hua
%A Toda, Tomoki
%A Tokuda, Keiichi
%A King, Simon
%A Renals, Steve
%D 2009
%J IEEE Transactions on Audio, Speech, and Language Processing
%K (artificial HTS;HMM-based Markov Speech Synthesis System, adaptation;speech analysis;Speech analysis;full-covariance approach;text-to-speech conversion hidden intelligence);speech model;speaker-dependent modeling;mixed-gender modeling;robust models;High models;learning reactor;Councils;Hidden science;Continuous-stirred science;Nose;Robustness;Speech speaker-adaptive speech superconductors;Information synthesis;Average synthesis;Computer synthesis;speaker synthesis;text synthesis;voice tank temperature voice;HMM
%N 6
%P 1208-1230
%R 10.1109/TASL.2009.2016394
%T Robust Speaker-Adaptive HMM-Based Text-to-Speech Synthesis
%V 17
%X This paper describes a speaker-adaptive HMM-based speech synthesis system. The new system, called ldquoHTS-2007,rdquo employs speaker adaptation (CSMAPLR+MAP), feature-space adaptive training, mixed-gender modeling, and full-covariance modeling using CSMAPLR transforms, in addition to several other techniques that have proved effective in our previous systems. Subjective evaluation results show that the new system generates significantly better quality synthetic speech than speaker-dependent approaches with realistic amounts of speech data, and that it bears comparison with speaker-dependent approaches even when large amounts of speech data are available. In addition, a comparison study with several speech synthesis techniques shows the new system is very robust: It is able to build voices from less-than-ideal speech data and synthesize good-quality speech even for out-of-domain sentences.
@article{Yamagishi2009a,
abstract = {This paper describes a speaker-adaptive HMM-based speech synthesis system. The new system, called ldquoHTS-2007,rdquo employs speaker adaptation (CSMAPLR+MAP), feature-space adaptive training, mixed-gender modeling, and full-covariance modeling using CSMAPLR transforms, in addition to several other techniques that have proved effective in our previous systems. Subjective evaluation results show that the new system generates significantly better quality synthetic speech than speaker-dependent approaches with realistic amounts of speech data, and that it bears comparison with speaker-dependent approaches even when large amounts of speech data are available. In addition, a comparison study with several speech synthesis techniques shows the new system is very robust: It is able to build voices from less-than-ideal speech data and synthesize good-quality speech even for out-of-domain sentences.},
added-at = {2021-02-01T10:51:23.000+0100},
author = {Yamagishi, Junichi and Nose, Takashi and Zen, Heiga and Ling, Zhen-Hua and Toda, Tomoki and Tokuda, Keiichi and King, Simon and Renals, Steve},
biburl = {https://www.bibsonomy.org/bibtex/29c1acaffc4254bf751efdb56e13e3867/m-toman},
doi = {10.1109/TASL.2009.2016394},
file = {:pdfs/yamagishi_transasp_2009.pdf:PDF},
interhash = {7987f8368ebd9e2c2db975036f36fd33},
intrahash = {9c1acaffc4254bf751efdb56e13e3867},
issn = {1558-7916},
journal = {IEEE Transactions on Audio, Speech, and Language Processing},
keywords = {(artificial HTS;HMM-based Markov Speech Synthesis System, adaptation;speech analysis;Speech analysis;full-covariance approach;text-to-speech conversion hidden intelligence);speech model;speaker-dependent modeling;mixed-gender modeling;robust models;High models;learning reactor;Councils;Hidden science;Continuous-stirred science;Nose;Robustness;Speech speaker-adaptive speech superconductors;Information synthesis;Average synthesis;Computer synthesis;speaker synthesis;text synthesis;voice tank temperature voice;HMM},
month = Aug,
number = 6,
owner = {schabus},
pages = {1208-1230},
timestamp = {2021-02-01T10:51:23.000+0100},
title = {Robust Speaker-Adaptive HMM-Based Text-to-Speech Synthesis},
volume = 17,
year = 2009
}