CORE
🇺🇦
make metadata, not war
Services
Research
Services overview
Explore all CORE services
Access to raw data
API
Dataset
FastSync
Content discovery
Recommender
Discovery
OAI identifiers
OAI Resolver
Managing content
Dashboard
Bespoke contracts
Consultancy services
Support us
Support us
Membership
Sponsorship
Community governance
Advisory Board
Board of supporters
Research network
About
About us
Our mission
Team
Blog
FAQs
Contact us
Personalising speech-to-speech translation in the EMIME project
Authors
,
W Byrne
+18 more
J Dines
PN Garner
M Gibson
Y Guan
T Hirsimäki
R Karhila
S King
M Kurimo
H Liang
K Oura
L Saheer
M Shannon
S Shiota
J Tian
K Tokuda
M Wester
YJ Wu
J Yamagishi
Publication date
1 December 2010
Publisher
Abstract
In the EMIME project we have studied unsupervised cross-lingual speaker adaptation. We have employed an HMM statistical framework for both speech recognition and synthesis which provides transformation mechanisms to adapt the synthesized voice in TTS (text-to-speech) using the recognized voice in ASR (automatic speech recognition). An important application for this research is personalised speech-to-speech translation that will use the voice of the speaker in the input language to utter the translated sentences in the output language. In mobile environments this enhances the users' interaction across language barriers by making the output speech sound more like the original speaker's way of speaking, even if she or he could not speak the output language. © 2010 Association for Computational Linguistics
Similar works
Full text
Available Versions
CUED - Cambridge University Engineering Department
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:generic.eprints.org:533116...
Last time updated on 15/07/2020