[1]
|
Manuel Sam Ribeiro, Oliver Watts, and Junichi Yamagishi.
Learning word vector representations based on acoustic counts.
In Proceedings of Interspeech, Stockholm, Sweden, August 2017.
[ bib |
.PDF |
Abstract ]
|
[2]
|
Michael Pucher, Bettina Zillinger, Markus Toman, Dietmar Schabus, Cassia
Valentini-Botinhao, Junichi Yamagishi, Erich Schmid, and Thomas Woltron.
Influence of speaker familiarity on blind and visually impaired
children and young adults perception of synthetic voices.
Computer Speech and Language, 46:179-195, June 2017.
[ bib |
DOI |
Abstract ]
|
[3]
|
Cassia Valentini-Botinhao and Junichi Yamagishi.
Speech intelligibility in cars: the effect of speaking style, noise
and listener age.
In Interspeech, 2017.
[ bib |
.pdf |
Abstract ]
|
[4]
|
Jaime Lorenzo-Trueba, Cassia Valentini-Botinhao, Gustav Henter, and Junichi
Yamagishi.
Misperceptions of the emotional content of natural and vocoded speech
in a car.
In Interspeech, 2017.
[ bib |
.pdf |
Abstract ]
|
[5]
|
Manuel Sam Ribeiro, Oliver Watts, and Junichi Yamagishi.
Syllable-level representations of suprasegmental features for
DNN-based text-to-speech synthesis.
In Proceedings of Interspeech, San Francisco, United States,
September 2016.
[ bib |
.PDF |
Abstract ]
|
[6]
|
Manuel Sam Ribeiro, Oliver Watts, and Junichi Yamagishi.
Parallel and cascaded deep neural networks for text-to-speech
synthesis.
In 9th ISCA Workshop on Speech Synthesis (SSW9), Sunnyvale,
United States, September 2016.
[ bib |
.pdf |
Abstract ]
|
[7]
|
Cassia Valentini-Botinhao, Xin Wang, Shinji Takaki, and Junichi Yamagishi.
Speech enhancement for a noise-robust text-to-speech synthesis system
using deep recurrent neural networks.
In Interspeech, pages 352-356. ISCA, September 2016.
[ bib |
DOI |
.pdf |
Abstract ]
|
[8]
|
Cassia Valentini-Botinhao, Xin Wang, Shinji Takaki, and Junichi Yamagishi.
Investigating RNN-based speech enhancement methods for noise-robust
text-to-speech.
In Proceedings of 9th ISCA Speech Synthesis Workshop, pages
159-165, September 2016.
[ bib |
.pdf |
Abstract ]
|
[9]
|
Fernando Villavicencio, Junichi Yamagishi, Jordi Bonada, and Felipe Espic.
Applying spectral normalisation and efficient envelope estimation and
statistical transformation for the voice conversion challenge 2016.
In Interspeech, pages 1657-61, San Francisco, USA, September
2016.
[ bib |
DOI |
http |
.PDF |
Abstract ]
|
[10]
|
Takenori Yoshimura, Gustav Eje Henter, Oliver Watts, Mirjam Wester, Junichi
Yamagishi, and Keiichi Tokuda.
A hierarchical predictor of synthetic speech naturalness using neural
networks.
In Interspeech 2016, pages 342-346. International Speech
Communication Association, September 2016.
[ bib |
DOI |
.PDF |
Abstract ]
|
[11]
|
Rasmus Dall, Sandrine Brognaux, Korin Richmond, Cassia Valentini-Botinhao,
Gustav Eje Henter, Julia Hirschberg, and Junichi Yamagishi.
Testing the consistency assumption: pronunciation variant forced
alignment in read and spontaneous speech synthesis.
In Proc. IEEE International Conference on Acoustics, Speech, and
Signal Processing (ICASSP), pages 5155-5159, March 2016.
[ bib |
.pdf |
Abstract ]
|
[12]
|
Qiong Hu, Junichi Yamagishi, Korin Richmond, Kartick Subramanian, and Yannis
Stylianou.
Initial investigation of speech synthesis based on complex-valued
neural networks.
In Proc. IEEE International Conference on Acoustics, Speech, and
Signal Processing (ICASSP), pages 5630-5634, March 2016.
[ bib |
.pdf |
Abstract ]
|
[13]
|
Manuel Sam Ribeiro, Oliver Watts, Junichi Yamagishi, and Robert A. J. Clark.
Wavelet-based decomposition of f0 as a secondary task for DNN-based
speech synthesis with multi-task learning.
In IEEE International Conference on Acoustics, Speech and Signal
Processing (ICASSP), Shanghai, China, March 2016.
[ bib |
.pdf |
Abstract ]
|
[14]
|
Adriana Stan, Yoshitaka Mamiya, Junichi Yamagishi, Peter Bell, Oliver Watts,
Rob Clark, and Simon King.
ALISA: An automatic lightly supervised speech segmentation and
alignment tool.
Computer Speech and Language, 35:116-133, 2016.
[ bib |
DOI |
http |
.pdf |
Abstract ]
|
[15]
|
Tomoki Toda, Ling-Hui Chen, Daisuke Saito, Fernando Villavicencio, Mirjam
Wester, Zhizheng Wu, and Junichi Yamagishi.
The voice conversion challenge 2016.
In Proc. Interspeech, 2016.
[ bib |
.pdf |
Abstract ]
|
[16]
|
Mirjam Wester, Zhizheng Wu, and Junichi Yamagishi.
Analysis of the voice conversion challenge 2016 evaluation results.
In Proc. Interspeech, 2016.
[ bib |
.pdf |
Abstract ]
|
[17]
|
Mirjam Wester, Zhizheng Wu, and Junichi Yamagishi.
Multidimensional scaling of systems in the voice conversion challenge
2016.
In Proc. Speech Synthesis Workshop 9, Sunnyvale, CA., 2016.
[ bib |
.pdf |
Abstract ]
|
[18]
|
Thomas Merritt, Robert A J Clark, Zhizheng Wu, Junichi Yamagishi, and Simon
King.
Deep neural network-guided unit selection synthesis.
In Proc. ICASSP, 2016.
[ bib |
.pdf |
Abstract ]
|
[19]
|
Cassia Valentini-Botinhao, Markus Toman, Michael Pucher, Dietmar Schabus, and
Junichi Yamagishi.
Intelligibility of time-compressed synthetic speech: Compression
method and speaking style.
Speech Communication, October 2015.
[ bib |
DOI |
Abstract ]
|
[20]
|
M. Pucher, M. Toman, D. Schabus, C. Valentini-Botinhao, J. Yamagishi,
B. Zillinger, and E Schmid.
Influence of speaker familiarity on blind and visually impaired
children's perception of synthetic voices in audio games.
In Proc. Interspeech, Dresden, Germany, September 2015.
[ bib |
.pdf |
Abstract ]
|
[21]
|
Thomas Merritt, Junichi Yamagishi, Zhizheng Wu, Oliver Watts, and Simon King.
Deep neural network context embeddings for model selection in
rich-context HMM synthesis.
In Proc. Interspeech, Dresden, September 2015.
[ bib |
.pdf |
Abstract ]
|
[22]
|
Manuel Sam Ribeiro, Junichi Yamagishi, and Robert A. J. Clark.
A perceptual investigation of wavelet-based decomposition of f0 for
text-to-speech synthesis.
In Proc. Interspeech, Dresden, Germany, September 2015.
[ bib |
.pdf |
Abstract ]
|
[23]
|
Mirjam Wester, Zhizheng Wu, and Junichi Yamagishi.
Human vs machine spoofing detection on wideband and narrowband data.
In Proc. Interspeech, Dresden, September 2015.
[ bib |
.pdf |
Abstract ]
|
[24]
|
Qiong Hu, Zhizheng Wu, Korin Richmond, Junichi Yamagishi, Yannis Stylianou, and
Ranniery Maia.
Fusion of multiple parameterisations for DNN-based sinusoidal
speech synthesis with multi-task learning.
In Proc. Interspeech, Dresden, Germany, September 2015.
[ bib |
.pdf |
Abstract ]
|
[25]
|
Qiong Hu, Yannis Stylianou, Ranniery Maia, Korin Richmond, and Junichi
Yamagishi.
Methods for applying dynamic sinusoidal models to statistical
parametric speech synthesis.
In Proc. ICASSP, Brisbane, Austrilia, April 2015.
[ bib |
.pdf |
Abstract ]
|
[26]
|
Ling-Hui Chen, T. Raitio, C. Valentini-Botinhao, Z. Ling, and J. Yamagishi.
A deep generative architecture for postfiltering in statistical
parametric speech synthesis.
Audio, Speech, and Language Processing, IEEE/ACM Transactions
on, 23(11):2003-2014, 2015.
[ bib |
DOI |
Abstract ]
|
[27]
|
Zhizheng Wu, Tomi Kinnunen, Nicholas Evans, Junichi Yamagishi, Cemal Hanilci,
Md Sahidullah, and Aleksandr Sizov.
ASVspoof 2015: the first automatic speaker verification spoofing
and countermeasures challenge.
In Interspeech, 2015.
[ bib |
.pdf ]
|
[28]
|
Zhizheng Wu, Ali Khodabakhsh, Cenk Demiroglu, Junichi Yamagishi, Daisuke Saito,
Tomoki Toda, and Simon King.
SAS: A speaker verification spoofing database containing diverse
attacks.
In Proceedings of the IEEE International Conference on
Acoustics, Speech, and Signal Processing (ICASSP), 2015.
[ bib |
.pdf ]
|
[29]
|
Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Rubén San-Segundo, Javier
Ferreiros, Junichi Yamagishi, and Juan M. Montero.
Emotion transplantation through adaptation in hmm-based speech
synthesis.
Computer Speech & Language, 34(1):292 - 307, 2015.
[ bib |
DOI |
http |
Abstract ]
|
[30]
|
Korin Richmond, Zhen-Hua Ling, and Junichi Yamagishi.
The use of articulatory movement data in speech synthesis
applications: An overview - application of articulatory movements using
machine learning algorithms [invited review].
Acoustical Science and Technology, 36(6):467-477, 2015.
[ bib |
DOI ]
|
[31]
|
Korin Richmond, Junichi Yamagishi, and Zhen-Hua Ling.
Applications of articulatory movements based on machine learning.
Journal of the Acoustical Society of Japan, 70(10):539-545,
2015.
[ bib ]
|
[32]
|
Cassia Valentini-Botinhao, Junichi Yamagishi, and Simon King.
Intelligibility enhancement of speech in noise.
In Proceedings of the Institute of Acoustics, volume 36 Pt. 2,
pages 96-103, Birmingham, UK, October 2014.
[ bib |
.pdf |
Abstract ]
|
[33]
|
Qiong Hu, Yannis Stylianou, Ranniery Maia, Korin Richmond, Junichi Yamagishi,
and Javier Latorre.
An investigation of the application of dynamic sinusoidal models to
statistical parametric speech synthesis.
In Proc. Interspeech, pages 780-784, Singapore, September
2014.
[ bib |
.pdf |
Abstract ]
|
[34]
|
L.-H. Chen, T. Raitio, C. Valentini-Botinhao, J. Yamagishi, and Z.-H. Ling.
DNN-Based Stochastic Postfilter for HMM-Based Speech Synthesis.
In Proc. Interspeech, pages 1954-1958, Singapore, September
2014.
[ bib |
.pdf |
Abstract ]
|
[35]
|
C. Valentini-Botinhao, M. Toman, M. Pucher, D. Schabus, and J. Yamagishi.
Intelligibility Analysis of Fast Synthesized Speech.
In Proc. Interspeech, pages 2922-2926, Singapore, September
2014.
[ bib |
.pdf |
Abstract ]
|
[36]
|
Nicholas W D Evans, Tomi Kinnunen, Junichi Yamagishi,
Zhizheng Wu, Federico Alegre, and Phillip De Leon.
Speaker recognition anti-spoofing.
Book Chapter in "Handbook of Biometric Anti-spoofing",
Springer, S. Marcel, S. Li and M. Nixon, Eds., 2014, June
2014.
[ bib |
DOI |
.pdf |
Abstract ]
|
[37]
|
Oliver Watts, Siva Gangireddy, Junichi Yamagishi, Simon King, Steve Renals,
Adriana Stan, and Mircea Giurgiu.
Neural net word representations for phrase-break prediction without a
part of speech tagger.
In Proc. ICASSP, pages 2618-2622, Florence, Italy, May 2014.
[ bib |
.pdf |
Abstract ]
|
[38]
|
Rasmus Dall, Junichi Yamagishi, and Simon King.
Rating naturalness in speech synthesis: The effect of style and
expectation.
In Proc. Speech Prosody, May 2014.
[ bib |
.pdf |
Abstract ]
|
[39]
|
Qiong Hu, Yannis Stylianou, Korin Richmond, Ranniery Maia, Junichi Yamagishi,
and Javier Latorre.
A fixed dimension and perceptually based dynamic sinusoidal model of
speech.
In Proceedings of the IEEE International Conference on
Acoustics, Speech, and Signal Processing (ICASSP), pages 6311-6315,
Florence, Italy, May 2014.
[ bib |
.pdf |
Abstract ]
|
[40]
|
L. Saheer, J. Yamagishi, P.N. Garner, and J. Dines.
Combining vocal tract length normalization with hierarchical linear
transformations.
Selected Topics in Signal Processing, IEEE Journal of,
8(2):262-272, April 2014.
[ bib |
DOI ]
|
[41]
|
J.P. Cabral, K. Richmond, J. Yamagishi, and S. Renals.
Glottal spectral separation for speech synthesis.
Selected Topics in Signal Processing, IEEE Journal of,
8(2):195-208, April 2014.
[ bib |
DOI |
.pdf |
Abstract ]
|
[42]
|
C. Valentini-Botinhao, J. Yamagishi, S. King, and R. Maia.
Intelligibility enhancement of HMM-generated speech in additive
noise by modifying mel cepstral coefficients to increase the glimpse
proportion.
Computer Speech and Language, 28(2):665-686, 2014.
[ bib |
DOI |
.pdf |
Abstract ]
|
[43]
|
Moses Ekpenyong, Eno-Abasi Urua, Oliver Watts, Simon King, and Junichi
Yamagishi.
Statistical parametric speech synthesis for Ibibio.
Speech Communication, 56:243-251, January 2014.
[ bib |
DOI |
http |
.pdf |
Abstract ]
|
[44]
|
P. Lanchantin, M. J. F. Gales, S. King, and J. Yamagishi.
Multiple-average-voice-based speech synthesis.
In Proc. ICASSP, 2014.
[ bib |
Abstract ]
|
[45]
|
Maria Luisa Garcia Lecumberri, Roberto Barra-Chicote, Rubén Pérez
Ramón, Junichi Yamagishi, and Martin Cooke.
Generating segmental foreign accent.
In Fifteenth Annual Conference of the International Speech
Communication Association, 2014.
[ bib |
.pdf |
Abstract ]
|
[46]
|
C. Valentini-Botinhao, J. Yamagishi, S. King, and Y. Stylianou.
Combining perceptually-motivated spectral shaping with loudness and
duration modification for intelligibility enhancement of HMM-based synthetic
speech in noise.
In Proc. Interspeech, Lyon, France, August 2013.
[ bib |
.pdf ]
|
[47]
|
Cassia Valentini-Botinhao, Mirjam Wester, Junichi Yamagishi, and Simon King.
Using neighbourhood density and selective SNR boosting to increase
the intelligibility of synthetic speech in noise.
In 8th ISCA Workshop on Speech Synthesis, pages 133-138,
Barcelona, Spain, August 2013.
[ bib |
.pdf |
Abstract ]
|
[48]
|
Nicholas W D Evans, Tomi Kinnunen, and Junichi Yamagishi.
Spoofing and countermeasures for automatic speaker verification.
In Interspeech 2013, 14th Annual Conference of the
International Speech Communication Association, August 25-29, 2013,
Lyon, France, Lyon, FRANCE, August 2013.
[ bib |
.pdf ]
|
[49]
|
Maria Astrinaki, Alexis Moinet, Junichi Yamagishi, Korin Richmond, Zhen-Hua
Ling, Simon King, and Thierry Dutoit.
Mage - reactive articulatory feature control of HMM-based
parametric speech synthesis.
In 8th ISCA Workshop on Speech Synthesis, pages 227-231,
Barcelona, Spain, August 2013.
[ bib |
.pdf ]
|
[50]
|
Qiong Hu, Korin Richmond, Junichi Yamagishi, and Javier Latorre.
An experimental comparison of multiple vocoder types.
In 8th ISCA Workshop on Speech Synthesis, pages 155-160,
Barcelona, Spain, August 2013.
[ bib |
.pdf |
Abstract ]
|
[51]
|
Adriana Stan, Peter Bell, Junichi Yamagishi, and Simon King.
Lightly supervised discriminative training of grapheme models for
improved sentence-level alignment of speech and text data.
In Proc. Interspeech, Lyon, France, August 2013.
[ bib |
.pdf |
Abstract ]
|
[52]
|
Yoshitaka Mamiya, Adriana Stan, Junichi Yamagishi, Peter Bell, Oliver Watts,
Robert Clark, and Simon King.
Using adaptation to improve speech transcription alignment in noisy
and reverberant environments.
In 8th ISCA Workshop on Speech Synthesis, pages 61-66,
Barcelona, Spain, August 2013.
[ bib |
.pdf |
Abstract ]
|
[53]
|
Oliver Watts, Adriana Stan, Rob Clark, Yoshitaka Mamiya, Mircea Giurgiu,
Junichi Yamagishi, and Simon King.
Unsupervised and lightly-supervised learning for rapid construction
of TTS systems in multiple languages from 'found' data: evaluation and
analysis.
In 8th ISCA Workshop on Speech Synthesis, pages 121-126,
Barcelona, Spain, August 2013.
[ bib |
.pdf |
Abstract ]
|
[54]
|
Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Junichi Yamagishi, Oliver Watts,
and Juan M. Montero.
Towards speaking style transplantation in speech synthesis.
In 8th ISCA Workshop on Speech Synthesis, pages 179-183,
Barcelona, Spain, August 2013.
[ bib |
.pdf |
Abstract ]
|
[55]
|
Adriana Stan, Oliver Watts, Yoshitaka Mamiya, Mircea Giurgiu, Rob Clark,
Junichi Yamagishi, and Simon King.
TUNDRA: A Multilingual Corpus of Found Data for TTS Research Created
with Light Supervision.
In Proc. Interspeech, Lyon, France, August 2013.
[ bib |
.pdf |
Abstract ]
|
[56]
|
Korin Richmond, Zhenhua Ling, Junichi Yamagishi, and Benigno Uría.
On the evaluation of inversion mapping performance in the acoustic
domain.
In Proc. Interspeech, pages 1012-1016, Lyon, France, August
2013.
[ bib |
.pdf |
Abstract ]
|
[57]
|
Maria Astrinaki, Alexis Moinet, Junichi Yamagishi, Korin Richmond, Zhen-Hua
Ling, Simon King, and Thierry Dutoit.
Mage - HMM-based speech synthesis reactively controlled by the
articulators.
In 8th ISCA Workshop on Speech Synthesis, page 243, Barcelona,
Spain, August 2013.
[ bib |
.pdf |
Abstract ]
|
[58]
|
Keiichi Tokuda, Yoshihiko Nankaku, Tomoki Today, Heiga Zen, Junichi Yamagishi,
and Keiichiro Oura.
Speech synthesis based on hidden markov models.
Proceedings of the IEEE, 101(6), June 2013.
(in press).
[ bib |
Abstract ]
|
[59]
|
C. Valentini-Botinhao, E. Godoy, Y. Stylianou, B. Sauert, S. King, and
J. Yamagishi.
Improving intelligibility in noise of HMM-generated speech via
noise-dependent and -independent methods.
In Proc. ICASSP, Vancouver, Canada, May 2013.
[ bib |
.pdf ]
|
[60]
|
John Dines, Hui Liang, Lakshmi Saheer, Matthew Gibson, William Byrne, Keiichiro
Oura, Keiichi Tokuda, Junichi Yamagishi, Simon King, Mirjam Wester, Teemu
Hirsimäki, Reima Karhila, and Mikko Kurimo.
Personalising speech-to-speech translation: Unsupervised
cross-lingual speaker adaptation for HMM-based speech synthesis.
Computer Speech and Language, 27(2):420-437, February 2013.
[ bib |
DOI |
http |
Abstract ]
|
[61]
|
Z. Ling, K. Richmond, and J. Yamagishi.
Articulatory control of HMM-based parametric speech synthesis using
feature-space-switched multiple regression.
Audio, Speech, and Language Processing, IEEE Transactions on,
21(1):207-219, 2013.
[ bib |
DOI |
.pdf |
Abstract ]
|
[62]
|
Sarah Creer, Stuart Cunningham, Phil Green, and Junichi Yamagishi.
Building personalised synthetic voices for individuals with severe
speech impairment.
Computer Speech & Language, 27(6):1178 - 1193, 2013.
<ce:title>Special Issue on Speech and Language Processing for
Assistive Technology</ce:title>.
[ bib |
DOI |
http ]
|
[63]
|
Yoshitaka Mamiya, Junichi Yamagishi, Oliver Watts, Robert A.J. Clark, Simon
King, and Adriana Stan.
Lightly supervised gmm vad to use audiobook for speech synthesiser.
In Proc. ICASSP, 2013.
[ bib |
.pdf |
Abstract ]
|
[64]
|
P. L. De Leon, M. Pucher, J. Yamagishi, I. Hernaez, and I. Saratxaga.
Evaluation of speaker verification security and detection of
HMM-based synthetic speech.
Audio, Speech, and Language Processing, IEEE Transactions on,
20(8):2280-2290, October 2012.
[ bib |
DOI |
Abstract ]
|
[65]
|
Zhen-Hua Ling, Korin Richmond, and Junichi Yamagishi.
Vowel creation by articulatory control in HMM-based parametric
speech synthesis.
In Proc. Interspeech, Portland, Oregon, USA, September 2012.
[ bib |
.pdf |
Abstract ]
|
[66]
|
Phillip L. De Leon, Bryan Stewart, and Junichi Yamagishi.
Synthetic speech discrimination using pitch pattern statistics
derived from image analysis.
In Proc. Interspeech, Portland, Oregon, USA, September 2012.
[ bib |
Abstract ]
|
[67]
|
J. Lorenzo, B. Martinez, R. Barra-Chicote, V. Lopez–Ludena, J. Ferreiros,
J. Yamagishi, and J.M. Montero.
Towards an unsupervised speaking style voice building framework:
Multi–style speaker diarization.
In Proc. Interspeech, Portland, Oregon, USA, September 2012.
[ bib |
Abstract ]
|
[68]
|
Rasmus Dall, Christophe Veaux, Junichi Yamagishi, and Simon King.
Analysis of speaker clustering techniques for HMM-based speech
synthesis.
In Proc. Interspeech, September 2012.
[ bib |
.pdf |
Abstract ]
|
[69]
|
Jaime Lorenzo-Trueba, Roberto Barra-Chicote, Tuomo Raitio, Nicolas Obin, Paavo
Alku, Junichi Yamagishi, and Juan M Montero.
Towards glottal source controllability in expressive speech
synthesis.
In Proc. Interspeech, Portland, Oregon, USA, September 2012.
[ bib |
Abstract ]
|
[70]
|
C. Valentini-Botinhao, J. Yamagishi, and S. King.
Evaluating speech intelligibility enhancement for HMM-based
synthetic speech in noise.
In Proc. Sapa Workshop, Portland, USA, September 2012.
[ bib |
.pdf |
Abstract ]
|
[71]
|
C. Valentini-Botinhao, J. Yamagishi, and S. King.
Mel cepstral coefficient modification based on the Glimpse
Proportion measure for improving the intelligibility of HMM-generated
synthetic speech in noise.
In Proc. Interspeech, Portland, USA, September 2012.
[ bib |
Abstract ]
|
[72]
|
Zhenhua Ling, Korin Richmond, and Junichi Yamagishi.
Vowel creation by articulatory control in HMM-based parametric
speech synthesis.
In Proc. The Listening Talker Workshop, page 72, Edinburgh, UK,
May 2012.
[ bib |
.pdf ]
|
[73]
|
C. Valentini-Botinhao, J. Yamagishi, and S. King.
Using an intelligibility measure to create noise robust cepstral
coefficients for HMM-based speech synthesis.
In Proc. LISTA Workshop, Edinburgh, UK, May 2012.
[ bib |
.pdf ]
|
[74]
|
C. Valentini-Botinhao, R. Maia, J. Yamagishi, S. King, and H. Zen.
Cepstral analysis based on the Glimpse proportion measure for
improving the intelligibility of HMM-based synthetic speech in noise.
In Proc. ICASSP, pages 3997-4000, Kyoto, Japan, March 2012.
[ bib |
DOI |
.pdf |
Abstract ]
|
[75]
|
L. Saheer, J. Yamagishi, P.N. Garner, and J. Dines.
Combining vocal tract length normalization with hierarchial linear
transformations.
In Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE
International Conference on, pages 4493 -4496, March 2012.
[ bib |
DOI |
Abstract ]
|
[76]
|
Chen-Yu Yang, G. Brown, Liang Lu, J. Yamagishi, and S. King.
Noise-robust whispered speech recognition using a non-audible-murmur
microphone with vts compensation.
In Chinese Spoken Language Processing (ISCSLP), 2012 8th
International Symposium on, pages 220-223, 2012.
[ bib |
DOI |
Abstract ]
|
[77]
|
Jaime Lorenzo-Trueba, Oliver Watts, Roberto Barra-Chicote, Junichi Yamagishi,
Simon King, and Juan M Montero.
Simple4all proposals for the albayzin evaluations in speech
synthesis.
In Proc. Iberspeech 2012, 2012.
[ bib |
.pdf |
Abstract ]
|
[78]
|
Sebastian Andersson, Junichi Yamagishi, and Robert A.J. Clark.
Synthesis and evaluation of conversational characteristics in
HMM-based speech synthesis.
Speech Communication, 54(2):175-188, 2012.
[ bib |
DOI |
http |
Abstract ]
|
[79]
|
S. Andersson, J. Yamagishi, and R.A.J. Clark.
Synthesis and evaluation of conversational characteristics in
HMM-based speech synthesis.
Speech Communication, 54(2):175-188, 2012.
[ bib |
DOI |
Abstract ]
|
[80]
|
Keiichiro Oura, Junichi Yamagishi, Mirjam Wester, Simon King, and Keiichi
Tokuda.
Analysis of unsupervised cross-lingual speaker adaptation for
HMM-based speech synthesis using KLD-based transform mapping.
Speech Communication, 54(6):703-714, 2012.
[ bib |
DOI |
http |
Abstract ]
|
[81]
|
Kei Hashimoto, Junichi Yamagishi, William Byrne, Simon King, and Keiichi
Tokuda.
Impacts of machine translation and speech synthesis on
speech-to-speech translation.
Speech Communication, 54(7):857-866, 2012.
[ bib |
DOI |
http |
Abstract ]
|
[82]
|
Junichi Yamagishi, Christophe Veaux, Simon King, and Steve Renals.
Speech synthesis technologies for individuals with vocal
disabilities: Voice banking and reconstruction.
Acoustical Science and Technology, 33(1):1-5, 2012.
[ bib |
DOI |
http |
.pdf |
Abstract ]
|
[83]
|
Sarah Creer, Stuart Cunningham, Phil Green, and Junichi Yamagishi.
Building personalised synthetic voices for individuals with severe
speech impairment.
Computer Speech and Language, 27(6):1178-1193, 2012.
[ bib |
DOI |
http |
Abstract ]
|
[84]
|
Oliver Watts, Junichi Yamagishi, and Simon King.
Unsupervised continuous-valued word features for phrase-break
prediction without a part-of-speech tagger.
In Proc. Interspeech, pages 2157-2160, Florence, Italy, August
2011.
[ bib |
.pdf |
Abstract ]
|
[85]
|
Cassia Valentini-Botinhao, Junichi Yamagishi, and Simon King.
Can objective measures predict the intelligibility of modified
HMM-based synthetic speech in noise?
In Proc. Interspeech, August 2011.
[ bib |
.pdf |
Abstract ]
|
[86]
|
Ming Lei, Junichi Yamagishi, Korin Richmond, Zhen-Hua Ling, Simon King, and
Li-Rong Dai.
Formant-controlled HMM-based speech synthesis.
In Proc. Interspeech, pages 2777-2780, Florence, Italy, August
2011.
[ bib |
.pdf |
Abstract ]
|
[87]
|
Zhen-Hua Ling, Korin Richmond, and Junichi Yamagishi.
Feature-space transform tying in unified acoustic-articulatory
modelling of articulatory control of HMM-based speech synthesis.
In Proc. Interspeech, pages 117-120, Florence, Italy, August
2011.
[ bib |
.pdf |
Abstract ]
|
[88]
|
S. Andraszewicz, J. Yamagishi, and S. King.
Vocal attractiveness of statistical speech synthesisers.
In Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE
International Conference on, pages 5368-5371, May 2011.
[ bib |
DOI |
Abstract ]
|
[89]
|
P.L. De Leon, I. Hernaez, I. Saratxaga, M. Pucher, and J. Yamagishi.
Detection of synthetic speech for the problem of imposture.
In Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE
International Conference on, pages 4844-4847, May 2011.
[ bib |
DOI |
Abstract ]
|
[90]
|
Cassia Valentini-Botinhao, Junichi Yamagishi, and Simon King.
Evaluation of objective measures for intelligibility prediction of
HMM-based synthetic speech in noise.
In Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE
International Conference on, pages 5112-5115, May 2011.
[ bib |
DOI |
.pdf |
Abstract ]
|
[91]
|
J.P. Cabral, S. Renals, J. Yamagishi, and K. Richmond.
HMM-based speech synthesiser using the LF-model of the glottal
source.
In Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE
International Conference on, pages 4704-4707, May 2011.
[ bib |
DOI |
.pdf |
Abstract ]
|
[92]
|
K. Hashimoto, J. Yamagishi, W. Byrne, S. King, and K. Tokuda.
An analysis of machine translation and speech synthesis in
speech-to-speech translation system.
In Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE
International Conference on, pages 5108-5111, May 2011.
[ bib |
DOI |
Abstract ]
|
[93]
|
T. Raitio, A. Suni, J. Yamagishi, H. Pulakka, J. Nurminen, M. Vainio, and
P. Alku.
HMM-based speech synthesis utilizing glottal inverse filtering.
IEEE Transactions on Audio, Speech and Language Processing,
19(1):153-165, January 2011.
[ bib |
DOI |
Abstract ]
|
[94]
|
J. Dines, J. Yamagishi, and S. King.
Measuring the gap between HMM-based ASR and TTS.
IEEE Selected Topics in Signal Processing, 2011.
(in press).
[ bib |
DOI |
Abstract ]
|
[95]
|
Adriana Stan, Junichi Yamagishi, Simon King, and Matthew Aylett.
The Romanian speech synthesis (RSS) corpus: Building a high
quality HMM-based speech synthesis system using a high sampling rate.
Speech Communication, 53(3):442-450, 2011.
[ bib |
DOI |
http |
Abstract ]
|
[96]
|
Zhen-Hua Ling, Korin Richmond, and Junichi Yamagishi.
An analysis of HMM-based prediction of articulatory movements.
Speech Communication, 52(10):834-846, October 2010.
[ bib |
DOI |
Abstract ]
|
[97]
|
Zhen-Hua Ling, Korin Richmond, and Junichi Yamagishi.
HMM-based text-to-articulatory-movement prediction and analysis of
critical articulators.
In Proc. Interspeech, pages 2194-2197, Makuhari, Japan,
September 2010.
[ bib |
.pdf |
Abstract ]
|
[98]
|
Yong Guan, Jilei Tian, Yi-Jian Wu, Junichi Yamagishi, and Jani Nurminen.
A unified and automatic approach of Mandarin HTS system.
In Proc. SSW7, Kyoto, Japan, September 2010.
[ bib |
.pdf ]
|
[99]
|
João Cabral, Steve Renals, Korin Richmond, and Junichi Yamagishi.
Transforming voice source parameters in a HMM-based speech
synthesiser with glottal post-filtering.
In Proc. 7th ISCA Speech Synthesis Workshop (SSW7), pages
365-370, NICT/ATR, Kyoto, Japan, September 2010.
[ bib |
.pdf |
Abstract ]
|
[100]
|
Oliver Watts, Junichi Yamagishi, and Simon King.
The role of higher-level linguistic features in HMM-based speech
synthesis.
In Proc. Interspeech, pages 841-844, Makuhari, Japan,
September 2010.
[ bib |
.pdf |
Abstract ]
|
[101]
|
Junichi Yamagishi, Oliver Watts, Simon King, and Bela Usabaev.
Roles of the average voice in speaker-adaptive HMM-based speech
synthesis.
In Proc. Interspeech, pages 418-421, Makuhari, Japan,
September 2010.
[ bib |
.pdf |
Abstract ]
|
[102]
|
Mirjam Wester, John Dines, Matthew Gibson, Hui Liang, Yi-Jian Wu, Lakshmi
Saheer, Simon King, Keiichiro Oura, Philip N. Garner, William Byrne, Yong
Guan, Teemu Hirsimäki, Reima Karhila, Mikko Kurimo, Matt Shannon, Sayaka
Shiota, Jilei Tian, Keiichi Tokuda, and Junichi Yamagishi.
Speaker adaptation and the evaluation of speaker similarity in the
EMIME speech-to-speech translation project.
In Proc. 7th ISCA Speech Synthesis Workshop, Kyoto, Japan,
September 2010.
[ bib |
.pdf |
Abstract ]
|
[103]
|
Michael Pucher, Dietmar Schabus, and Junichi Yamagishi.
Synthesis of fast speech with interpolation of adapted HSMMs and
its evaluation by blind and sighted listeners.
In Proc. Interspeech, pages 2186-2189, Makuhari, Japan,
September 2010.
[ bib |
.pdf |
Abstract ]
|
[104]
|
Sebastian Andersson, Junichi Yamagishi, and Robert Clark.
Utilising spontaneous conversational speech in HMM-based speech
synthesis.
In The 7th ISCA Tutorial and Research Workshop on Speech
Synthesis, September 2010.
[ bib |
.pdf |
Abstract ]
|
[105]
|
Oliver Watts, Junichi Yamagishi, and Simon King.
Letter-based speech synthesis.
In Proc. Speech Synthesis Workshop 2010, pages 317-322, Nara,
Japan, September 2010.
[ bib |
.pdf |
Abstract ]
|
[106]
|
O. Watts, J. Yamagishi, S. King, and K. Berkling.
Synthesis of child speech with HMM adaptation and voice conversion.
Audio, Speech, and Language Processing, IEEE Transactions on,
18(5):1005-1016, July 2010.
[ bib |
DOI |
.pdf |
Abstract ]
|
[107]
|
Mikko Kurimo, William Byrne, John Dines, Philip N. Garner, Matthew Gibson, Yong
Guan, Teemu Hirsimäki, Reima Karhila, Simon King, Hui Liang, Keiichiro
Oura, Lakshmi Saheer, Matt Shannon, Sayaka Shiota, Jilei Tian, Keiichi
Tokuda, Mirjam Wester, Yi-Jian Wu, and Junichi Yamagishi.
Personalising speech-to-speech translation in the EMIME project.
In Proc. ACL 2010 System Demonstrations, Uppsala, Sweden, July
2010.
[ bib |
.pdf |
Abstract ]
|
[108]
|
J. Yamagishi, B. Usabaev, S. King, O. Watts, J. Dines, J. Tian, R. Hu, Y. Guan,
K. Oura, K. Tokuda, R. Karhila, and M. Kurimo.
Thousands of voices for HMM-based speech synthesis - analysis and
application of TTS systems built on various ASR corpora.
IEEE Transactions on Audio, Speech and Language Processing,
18(5):984-1004, July 2010.
[ bib |
DOI |
Abstract ]
|
[109]
|
R. Barra-Chicote, J. Yamagishi, S. King, J. Manuel Monero, and
J. Macias-Guarasa.
Analysis of statistical parametric and unit-selection speech
synthesis systems applied to emotional speech.
Speech Communication, 52(5):394-404, May 2010.
[ bib |
DOI |
Abstract ]
|
[110]
|
P. L. De Leon, V. R. Apsingekar, M. Pucher, and J. Yamagishi.
Revisiting the security of speaker verification systems against
imposture using synthetic speech.
In Proc. ICASSP 2010, Dallas, Texas, USA, 2010.
[ bib |
.pdf ]
|
[111]
|
P.L. De Leon, M. Pucher, and J. Yamagishi.
Evaluation of the vulnerability of speaker verification to synthetic
speech.
In Proc. Odyssey (The speaker and language recognition
workshop) 2010, Brno, Czech Republic, 2010.
[ bib |
.pdf ]
|
[112]
|
Keiichiro Oura, Keiichi Tokuda, Junichi Yamagishi, Mirjam Wester, and Simon
King.
Unsupervised cross-lingual speaker adaptation for HMM-based speech
synthesis.
In Proc. ICASSP, volume I, pages 4954-4957, 2010.
[ bib |
.pdf |
Abstract ]
|
[113]
|
Michael Pucher, Dietmar Schabus, Junichi Yamagishi, Friedrich Neubarth, and
Volker Strom.
Modeling and interpolation of Austrian German and Viennese dialect
in HMM-based speech synthesis.
Speech Communication, 52(2):164-179, 2010.
[ bib |
DOI |
Abstract ]
|
[114]
|
J. Yamagishi and S. King.
Simple methods for improving speaker-similarity of HMM-based speech
synthesis.
In Proc. ICASSP 2010, Dallas, Texas, USA, 2010.
[ bib |
.pdf ]
|
[115]
|
Heiga Zen, Keiichiro Oura, Takashi Nose, Junichi Yamagishi, Shinji Sako, Tomoki
Toda, Takashi Masuko, Alan W. Black, and Keiichi Tokuda.
Recent development of the HMM-based speech synthesis system
(HTS).
In Proc. 2009 Asia-Pacific Signal and Information Processing
Association (APSIPA), Sapporo, Japan, October 2009.
[ bib |
.pdf |
Abstract ]
|
[116]
|
J. Sebastian Andersson, Joao P. Cabral, Leonardo Badino, Junichi Yamagishi, and
Robert A.J. Clark.
Glottal source and prosodic prominence modelling in HMM-based
speech synthesis for the Blizzard Challenge 2009.
In The Blizzard Challenge 2009, Edinburgh, U.K., September
2009.
[ bib |
.pdf |
Abstract ]
|
[117]
|
Oliver Watts, Junichi Yamagishi, Simon King, and Kay Berkling.
HMM adaptation and voice conversion for the synthesis of child
speech: A comparison.
In Proc. Interspeech 2009, pages 2627-2630, Brighton, U.K.,
September 2009.
[ bib |
.pdf |
Abstract ]
|
[118]
|
Leonardo Badino, J. Sebastian Andersson, Junichi Yamagishi, and Robert A.J.
Clark.
Identification of contrast and its emphatic realization in
HMM-based speech synthesis.
In Proc. Interspeech 2009, Brighton, U.K., September 2009.
[ bib |
.PDF |
Abstract ]
|
[119]
|
Junichi Yamagishi, Mike Lincoln, Simon King, John Dines, Matthew Gibson, Jilei
Tian, and Yong Guan.
Analysis of unsupervised and noise-robust speaker-adaptive
HMM-based speech synthesis systems toward a unified ASR and TTS
framework.
In Proc. Interspeech 2009, Edinburgh, U.K., September 2009.
[ bib |
Abstract ]
|
[120]
|
J. Dines, J. Yamagishi, and S. King.
Measuring the gap between HMM-based ASR and TTS.
In Proc. Interspeech, pages 1391-1394, Brighton, U.K.,
September 2009.
[ bib |
Abstract ]
|
[121]
|
J. Yamagishi, Bela Usabaev, Simon King, Oliver Watts, John Dines, Jilei Tian,
Rile Hu, Yong Guan, Keiichiro Oura, Keiichi Tokuda, Reima Karhila, and Mikko
Kurimo.
Thousands of voices for HMM-based speech synthesis.
In Proc. Interspeech, pages 420-423, Brighton, U.K., September
2009.
[ bib |
http |
Abstract ]
|
[122]
|
Z. Ling, K. Richmond, J. Yamagishi, and R. Wang.
Integrating articulatory features into HMM-based parametric speech
synthesis.
IEEE Transactions on Audio, Speech and Language Processing,
17(6):1171-1185, August 2009.
IEEE SPS 2010 Young Author Best Paper Award.
[ bib |
DOI |
Abstract ]
|
[123]
|
J. Cabral, S. Renals, K. Richmond, and J. Yamagishi.
HMM-based speech synthesis with an acoustic glottal source model.
In Proc. The First Young Researchers Workshop in Speech
Technology, April 2009.
[ bib |
.pdf |
Abstract ]
|
[124]
|
Sarah Creer, Phil Green, Stuart Cunningham, and Junichi Yamagishi.
Building personalised synthesised voices for individuals with
dysarthria using the HTS toolkit.
In John W. Mullennix and Steven E. Stern, editors, Computer
Synthesized Speech Technologies: Tools for Aiding Impairment. IGI Global,
1st edition, 2009.
in press.
[ bib |
Abstract ]
|
[125]
|
Matthew P. Aylett, Simon King, and Junichi Yamagishi.
Speech synthesis without a phone inventory.
In Interspeech, pages 2087-2090, 2009.
[ bib |
.pdf |
Abstract ]
|
[126]
|
Junichi Yamagishi, Takashi Nose, Heiga Zen, Zhenhua Ling, Tomoki Toda, Keiichi
Tokuda, Simon King, and Steve Renals.
Robust speaker-adaptive HMM-based text-to-speech synthesis.
IEEE Transactions on Audio, Speech and Language Processing,
17(6):1208-1230, 2009.
[ bib |
http |
www: |
Abstract ]
|
[127]
|
R. Barra-Chicote, J. Yamagishi, J.M. Montero, S. King, S. Lutfi, and
J. Macias-Guarasa.
Generacion de una voz sintetica en Castellano basada en HSMM para
la Evaluacion Albayzin 2008: conversion texto a voz.
In V Jornadas en Tecnologia del Habla, pages 115-118, November
2008.
(in Spanish).
[ bib |
.pdf ]
|
[128]
|
Oliver Watts, Junichi Yamagishi, Kay Berkling, and Simon King.
HMM-based synthesis of child speech.
In Proc. 1st Workshop on Child, Computer and Interaction
(ICMI'08 post-conference workshop), Crete, Greece, October 2008.
[ bib |
.pdf |
Abstract ]
|
[129]
|
Zhen-Hua Ling, Korin Richmond, Junichi Yamagishi, and Ren-Hua Wang.
Articulatory control of HMM-based parametric speech synthesis
driven by phonetic knowledge.
In Proc. Interspeech, pages 573-576, Brisbane, Australia,
September 2008.
[ bib |
.PDF |
Abstract ]
|
[130]
|
Matthew P. Aylett and Junichi Yamagishi.
Combining statistical parameteric speech synthesis and unit-selection
for automatic voice cloning.
In Proc. LangTech 2008, Brisbane, Australia, September 2008.
[ bib |
.pdf |
Abstract ]
|
[131]
|
Junichi Yamagishi, Zhenhua Ling, and Simon King.
Robustness of hmm-based speech synthesis.
In Proc. Interspeech 2008, pages 581-584, Brisbane, Australia,
September 2008.
[ bib |
.pdf |
Abstract ]
|
[132]
|
J. Cabral, S. Renals, K. Richmond, and J. Yamagishi.
Glottal spectral separation for parametric speech synthesis.
In Proc. Interspeech, pages 1829-1832, Brisbane, Australia,
September 2008.
[ bib |
.PDF |
Abstract ]
|
[133]
|
Gregor Hofer, Junichi Yamagishi, and Hiroshi Shimodaira.
Speech-driven lip motion generation with a trajectory HMM.
In Proc. Interspeech 2008, pages 2314-2317, Brisbane,
Australia, September 2008.
[ bib |
.pdf |
Abstract ]
|
[134]
|
Simon King, Keiichi Tokuda, Heiga Zen, and Junichi Yamagishi.
Unsupervised adaptation for hmm-based speech synthesis.
In Proc. Interspeech, pages 1869-1872, Brisbane, Australia,
September 2008.
[ bib |
.PDF |
Abstract ]
|
[135]
|
Junichi Yamagishi, Heiga Zen, Yi-Jian Wu, Tomoki Toda, and Keiichi Tokuda.
The HTS-2008 system: Yet another evaluation of the speaker-adaptive
HMM-based speech synthesis system in the 2008 Blizzard Challenge.
In Proc. Blizzard Challenge 2008, Brisbane, Australia,
September 2008.
[ bib |
.pdf |
Abstract ]
|
[136]
|
Junichi Yamagishi, Hisashi Kawai, and Takao Kobayashi.
Phone duration modeling using gradient tree boosting.
Speech Communication, 50(5):405-415, May 2008.
[ bib |
DOI |
Abstract ]
|
[137]
|
Junichi Yamagishi, Takashi Nose, Heiga Zen, Tomoki Toda, and Keiichi Tokuda.
Performance evaluation of the speaker-independent HMM-based speech
synthesis system "HTS-2007" for the Blizzard Challenge 2007.
In Proc. ICASSP 2008, pages 3957-3960, Las Vegas, U.S.A, April
2008.
[ bib |
DOI |
Abstract ]
|
[138]
|
Junichi Yamagishi, Takao Kobayashi, Yuji Nakano, Katsumi Ogata, and Juri
Isogai.
Analysis of speaker adaptation algorihms for HMM-based speech
synthesis and a constrained SMAPLR adaptation algorithm.
IEEE Transactions on Audio, Speech and Language Processing,
2008.
In print.
[ bib |
Abstract ]
|
[139]
|
Takashi Nose, Junichi Yamagishi, and Takao Kobayashi.
A style control technique for HMM-based expressive speech
synthesis.
IEICE Trans. Information and Systems, E90-D(9):1406-1413,
September 2007.
[ bib |
http |
Abstract ]
|
[140]
|
Junichi Yamagishi, Takao Kobayashi, Steve Renals, Simon King, Heiga Zen, Tomoki
Toda, and Keiichi Tokuda.
Improved average-voice-based speech synthesis using gender-mixed
modeling and a parameter generation algorithm considering GV.
In Proc. 6th ISCA Workshop on Speech Synthesis (SSW-6), August
2007.
[ bib |
.pdf |
Abstract ]
|
[141]
|
Toshio Hirai, Junichi Yamagishi, and Seiichi Tenpaku.
Utilization of an HMM-based feature generation module in 5 ms
segment concatenative speech synthesis.
In Proc. 6th ISCA Workshop on Speech Synthesis (SSW-6), August
2007.
[ bib |
Abstract ]
|
[142]
|
Heiga Zen, Takashi Nose, Junichi Yamagishi, Shinji Sako, Takashi Masuko, Alan
Black, and Keiichi Tokuda.
The HMM-based speech synthesis system (HTS) version 2.0.
In Proc. 6th ISCA Workshop on Speech Synthesis (SSW-6), August
2007.
[ bib |
Abstract ]
|
[143]
|
Makoto Tachibana, Keigo Kawashima, Junichi Yamagishi, and Takao Kobayashi.
Performance evaluation of HMM-based style classification with a
small amount of training data.
In Proc. Interspeech 2007, August 2007.
[ bib |
Abstract ]
|
[144]
|
Junichi Yamagishi, Heiga Zen, Tomoki Toda, and Keiichi Tokuda.
Speaker-independent HMM-based speech synthesis system - HTS-2007
system for the Blizzard Challenge 2007.
In Proc. Blizzard Challenge 2007, August 2007.
[ bib |
.pdf |
Abstract ]
|
[145]
|
K. Richmond, V. Strom, R. Clark, J. Yamagishi, and S. Fitt.
Festival multisyn voices for the 2007 blizzard challenge.
In Proc. Blizzard Challenge Workshop (in Proc. SSW6), Bonn,
Germany, August 2007.
[ bib |
.pdf |
Abstract ]
|
[146]
|
Junichi Yamagishi and Takao Kobayashi.
Average-voice-based speech synthesis using hsmm-based speaker
adaptation and adaptive training.
IEICE Trans. Information and Systems, E90-D(2):533-543,
February 2007.
[ bib |
Abstract ]
|
[147]
|
J. Yamagishi, T. Kobayashi, M. Tachibana, K. Ogata, and Y. Nakano.
Model adaptation approach to speech synthesis with diverse voices and
styles.
In Proc. ICASSP, pages 1233-1236, 2007.
[ bib |
Abstract ]
|
[148]
|
Gregor Hofer, Hiroshi Shimodaira, and Junichi Yamagishi.
Speech-driven head motion synthesis based on a trajectory model.
Poster at Siggraph 2007, 2007.
[ bib |
.pdf ]
|
[149]
|
Gregor Hofer, Hiroshi Shimodaira, and Junichi Yamagishi.
Lip motion synthesis using a context dependent trajectory hidden
Markov model.
Poster at SCA 2007, 2007.
[ bib |
.pdf ]
|
[150]
|
J. Cabral, S. Renals, K. Richmond, and J. Yamagishi.
Towards an improved modeling of the glottal source in statistical
parametric speech synthesis.
In Proc.of the 6th ISCA Workshop on Speech Synthesis, Bonn,
Germany, 2007.
[ bib |
.pdf |
Abstract ]
|
[151]
|
Hisashi Kawai, Tomoki Toda, Junichi Yamagishi, Toshio Hirai, Jinfu Ni, Nobuyuki
Nishizawa, Minoru Tsuzaki, and Keiichi Tokuda.
Ximera: a concatenative speech synthesis system with large scale
corpora.
IEICE Trans. Information and Systems, J89-D-II(12):2688-2698,
December 2006.
[ bib ]
|