13.07.2015 Views

Proceedings Fonetik 2009 - Institutionen för lingvistik

Proceedings Fonetik 2009 - Institutionen för lingvistik

Proceedings Fonetik 2009 - Institutionen för lingvistik

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

<strong>Proceedings</strong>, FONETIK <strong>2009</strong>, Dept. of Linguistics, Stockholm UniversityStudies on using the SynFace talking head for the hearingimpairedSamer Al Moubayed 1 , Jonas Beskow 1 , Ann-Marie Öster 1 , Giampiero Salvi 1 , Björn Granström 1 , Nicvan Son 2 , Ellen Ormel 2 ,Tobias Herzke 31KTH Centre for Speech Technology, Stockholm, Sweden.2 Viataal, Nijmegen, The Netherlands.3 HörTech gGmbH, Germany.sameram@kth.se, {beskow, annemarie, giampi, bjorn}@speech.kth.se, n.vson@viataal.nl, elleno@socsci.ru.nl,t.herzke@hoertech.deAbstractSynFace is a lip-synchronized talking agentwhich is optimized as a visual reading supportfor the hearing impaired. In this paper wepresent the large scale hearing impaired userstudies carried out for three languages in theHearing at Home project. The user tests focuson measuring the gain in Speech ReceptionThreshold in Noise and the effort scaling whenusing SynFace by hearing impaired people,where groups of hearing impaired subjects withdifferent impairment levels from mild to severeand cochlear implants are tested. Preliminaryanalysis of the results does not show significantgain in SRT or in effort scaling. But looking atlarge cross-subject variability in both tests, it isclear that many subjects benefit from SynFaceespecially with speech with stereo babble.IntroductionThere is a growing number of hearing impairedpersons in the society today. In the ongoingEU-project Hearing at Home (HaH)(Beskow et al., 2008), the goal is to develop thenext generation of assistive devices that willallow this group - which predominantly includesthe elderly - equal participation in communicationand empower them to play a fullrole in society. The project focuses on theneeds of hearing impaired persons in home environments.For a hearing impaired person, it is oftennecessary to be able to lip-read as well as hearthe person they are talking with in order tocommunicate successfully. Often, only the audiosignal is available, e.g. during telephoneconversations or certain TV broadcasts. One ofthe goals of the HaH project is to study the useof visual lip-reading support by hard of hearingpeople for home information, home entertainment,automation, and care applications.The SynFace Lip-SynchronizedTalking AgentSynFace (Beskow et al, 2008) is a supportivetechnology for hearing impaired persons,which aims to re-create the visible articulationof a speaker, in the form of an animated talkinghead. SynFace employs a specially developedreal-time phoneme recognition system, basedon a hybrid of recurrent artificial neural networks(ANNs) and Hidden Markov Models(HMMs) that delivers information regardingthe speech articulation to a speech animationmodule that renders the talking face to thecomputer screen using 3D graphics.SynFace previously has been trained on fourlanguages: English, Flemish, German and Swedish.The training used the multilingualSpeechDat corpora. To align the corpora, theHTK (Hidden markov models ToolKit) basedRefRec recogniser (Lindberg et al, 2000) wastrained to derive the phonetic transcription ofthe corpus. Table 1 presents the % correctframe of the recognizers of the four languagesSynFace contains.Table 1. Complexity and % correct frame of the recognizersof different languages in SynFace.Language Connections % correct frameSwedish 541,250 54.2English 184,848 53.0German 541,430 61.0Flemish 186,853 51.0User StudiesThe SynFace has been previously evaluatedby subjects in many ways in Agelfors et al(1998), Agelfors et al (2006) and Siciliano et al(2003). In the present study, a large scale test ofthe use of SynFace as an audio-visual support140

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!