Unsupervised Speaker Adaptation for Speech-to-Speech Translation System
スポンサーリンク
概要
- 論文の詳細を見る
In the EMIME project, we are developing a mobile device that performs personalized speech-to-speech translation such that a user's spoken input in one language is used to produce spoken output in another language, while continuing to sound like the user's voice. We integrate two techniques, unsupervised adaptation for HMM-based TTS using a word-based large-vocabulary continuous speech recognizer and cross-lingual speaker adaptation for HMM-based TTS, into a single architecture. Thus, an unsupervised cross-lingual speaker adaptation system can be developed. Listening tests show very promising results, demonstrating that adapted voices sound similar to the target speaker and that differences between supervised and unsupervised cross-lingual speaker adaptation are small.
- 一般社団法人情報処理学会の論文
- 2009-12-14
著者
-
Keiichi Tokuda
Nagoya Institute Of Technology Japan
-
Keiichiro Oura
Nagoya Institute Of Technology Japan
-
Junichi Yamagishi
The Centre for Speech Technology Research, UK
-
Mirjam Wester
The Centre for Speech Technology Research, UK
-
Simon King
The Centre for Speech Technology Research, UK
-
Junichi Yamagishi
The Centre For Speech Technology Research Uk
-
Simon King
The Centre For Speech Technology Research Uk
-
Mirjam Wester
The Centre For Speech Technology Research Uk