You are here

Publications

Export 5 results:
Author Title Type [ Year(Asc)]
2015
T. Merritt, Latorre, J., and King, S., Attributing modelling errors in HMM synthesis by stepping gradually from natural to modelled speech, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Brisbane, 2015.
Z. Wu, Valentini-Botinhao, C., Watts, O., and King, S., Deep neural networks employing multi-task learning and stacked bottleneck features for speech synthesis, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015.
P. Swietojanski and Renals, S., Differentiable Pooling for Unsupervised Speaker Adaptation, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015.
N. Obin, Veaux, C., and Lanchantin, P., Exploiting Alternatives for Text-To-Speech Synthesis: From Machine to Human. Springer Verlag, 2015, pp. 189-202.
X. Chen, Liu, X., Gales, M., and Woodland, P., Improving the training and evaluation efficiency of recurrent neural network language models, in Proc. ICASSP, Brisbane, Australia, 2015.
Y. Liu, Karanasou, P., and Hain, T., An Investigation Into Speaker Informed DNN Front-end for LVCSR, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015.
Q. Hu, Stylianou, Y., Maia, R., Richmond, K., and Yamagishi, J., Methods for applying dynamic sinusoidal models to statistical parametric speech synthesis, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015.
B. Uria, Murray, I., Renals, S., and Valentini-Botinhao, C., Modelling acoustic feature dependencies with artificial neural networks: Trajectory-RNADE, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015.
C. Wu and Gales, M., MULTI-BASIS ADAPTIVE NEURAL NETWORK FOR RAPID ADAPTATION IN SPEECH RECOGNITION, in Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on, 2015.
X. Liu, Chen, X., Gales, M., and Woodland, P., Paraphrastic Recurrent Neural Network Language Models, in Proc. ICASSP, Brisbane, Australia, 2015.
X. Chen, Liu, X., Gales, M., and Woodland, P., Recurrent neural network language model training with noise contrastive estimation for speech recognition, in Proc. ICASSP, Brisbane, Australia, 2015.
N. Obin and Lanchantin, P., Symbolic Modelling of Speech Prosody: From Linguistics to Statistics, IEEE Transactions on Audio, Speech, and Language Processing, vol. 23, 2015.
Z. Wu, Khodabakhsh, A., Demiroglu, C., Yamagishi, J., Saito, D., Toda, T., and King, S., SAS: A speaker verification spoofing database containing diverse attacks, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015.
P. Swietojanski, Bell, P., and Renals, S., Structured Output Layer with Auxiliary Targets for Context-Dependent Acoustic Modelling, in Proc. Interspeech, Dresden, Germany, 2015.
2014
P. Karanasou, Wang, Y., Gales, M., and Woodland, P., Adaptation of Deep Neural Network Acoustic Models Using Factorised I-vectors, in Proceedings of Interspeech’14, 2014.
I. Casanueva, Christensen, H., Hain, T., and Green, P., "Adaptive speech recognition and dialogue management for users with speech disorders, in Proceedings of Interspeech'14, 2014.
H. Christensen, Casanueva, I., Cunningham, S., Green, P., and Hain, T., Automatic Selection of Speakers for Improved Acoustic Modelling : Recognition of Disordered Speech with Sparse Data, in Spoken Language Technology Workshop, SLT'14, Lake Tahoe, 2014.
O. Saz, Doulaty, M., and Hain, T., Background-tracking acoustic features for genre identification of broadcast shows, in Proceedings of the 2014 IEEE Spoken Language Technology Workshop (SLT), South Lake Tahoe, NV, 2014, pp. 118–123.
P. Swietojanski, Ghoshal, A., and Renals, S., Convolutional Neural Networks for Distant Speech Recognition, Signal Processing Letters, IEEE, vol. 21, pp. 1120-1124, 2014.
L. Lu, Ghoshal, A., and Renals, S., Cross-lingual subspace Gaussian mixture model for low-resource speech recognition, IEEE Transactions on Audio, Speech and Language Processing, 2014.
R. Dall, Wester, M., and Corley, M., The Effect of Filled Pauses and Speaking Rate on Speech Comprehension in Natural, Vocoded and Synthetic Speech, in Proceedings of Interspeech, 2014.
X. Chen, Wang, Y., Liu, X., Gales, M., and Woodland, P., Efficient GPU-based training of recurrent neural network language models using spliced sentence bunch, in Proc. Interspeech, Singapore, 2014.
X. Liu, Wang, Y., Chen, X., Gales, M., and Woodland, P., EFFICIENT LATTICE RESCORING USING RECURRENT NEURAL NETWORK LANGUAGE MODELS, in IEEE ICASSP2014, Florence, Italy, 2014.
M. P. Aylett, Dall, R., Ghoshal, A., Henter, G. Eje, and Merritt, T., A Flexible Front-End for HTS, in Proc. Interspeech, Singapore, 2014.
R. Dall, Tomalin, M., Wester, M., Byrne, W., and King, S., Investigating Automatic & Human Filled Pause Insertion for Speech Synthesis, in Proceedings of Interspeech, 2014.

Pages