11
Hindawi Publishing Corporation EURASIP Journal on Audio, Speech, and Music Processing Volume 2009, Article ID 191940, 10 pages doi:10.1155/2009/191940 Research Article SynFace—Speech-Driven Facial Animation for Virtual Speech-Reading Support Giampiero Salvi, Jonas Beskow, Samer Al Moubayed, and Bj¨ orn Granstr¨ om KTH, School of Computer Science and Communication, Deptarment for Speech, Music, and Hearing, Lindstedtsv¨ agen 24, SE-100 44 Stockholm, Sweden Correspondence should be addressed to Giampiero Salvi, [email protected] Received 13 March 2009; Revised 23 July 2009; Accepted 23 September 2009 Recommended by G´ erard Bailly This paper describes SynFace, a supportive technology that aims at enhancing audio-based spoken communication in adverse acoustic conditions by providing the missing visual information in the form of an animated talking head. Firstly, we describe the system architecture, consisting of a 3D animated face model controlled from the speech input by a specifically optimised phonetic recogniser. Secondly, we report on speech intelligibility experiments with focus on multilinguality and robustness to audio quality. The system, already available for Swedish, English, and Flemish, was optimised for German and for Swedish wide-band speech quality available in TV, radio, and Internet communication. Lastly, the paper covers experiments with nonverbal motions driven from the speech signal. It is shown that turn-taking gestures can be used to aect the flow of human-human dialogues. We have focused specifically on two categories of cues that may be extracted from the acoustic signal: prominence/emphasis and interactional cues (turn-taking/back-channelling). Copyright © 2009 Giampiero Salvi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. 1. Introduction For a hearing impaired person, and for a normal hearing person in adverse acoustic conditions, it is often necessary to be able to lip-read as well as hear the person they are talking with in order to communicate successfully. Apart from the lip movements, nonverbal visual information is also essential to keep a normal flow of conversation. Often, only the audio signal is available, for example, during telephone conversations or certain TV broadcasts. The idea behind SynFace is to try to recreate the visible articulation of the speaker, in the form of an animated talking head. The visual signal is presented in synchrony with the acoustic speech signal, which means that the user can benefit from the com- bined synchronised audiovisual perception of the original speech acoustics and the resynthesised visible articulation. When compared to video telephony solutions, SynFace has the distinct advantage that only the user on the receiving end needs special equipment—the speaker at the other end can use any telephone terminal and technology: fixed, mobilem, or IP-telephony. Several methods have been proposed to drive the lip movements of an avatar from the acoustic speech signal with varying synthesis models and acoustic-to-visual maps. Tamura et al. [1] used hidden Markov models (HMMs) that are trained on parameters that represent both auditory and visual speech features. Similarly, Nakamura and Yamamoto [2] propose to estimate the audio-visual joint probability using HMMs. Wen et al. [3] extract the visual information from the output of a formant analyser. Al Moubayed et al. [4] map from the lattice output of a phonetic recogniser to texture parameters using neural networks. Hofer et al. [5] used trajectory hidden Markov models to predict visual speech parameters from an observed sequence. Most existing approaches to acoustic-to-visual speech mapping can be categorised as either regression based or classification based. Regression-based systems try to map features of the incoming sounds into continuously varying articulatory (or visual) parameters. Classification-based sys- tems, such as SynFace, consider an intermediate phonetic level, thus solving a classification problem, and generating the final face parameters with a rule-based system. This

SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

  • Upload
    lythuy

  • View
    217

  • Download
    0

Embed Size (px)

Citation preview

Page 1: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

Hindawi Publishing CorporationEURASIP Journal on Audio, Speech, and Music ProcessingVolume 2009, Article ID 191940, 10 pagesdoi:10.1155/2009/191940

Research Article

SynFace—Speech-Driven Facial Animation forVirtual Speech-Reading Support

Giampiero Salvi, Jonas Beskow, Samer Al Moubayed, and Bjorn Granstrom

KTH, School of Computer Science and Communication, Deptarment for Speech, Music, and Hearing,Lindstedtsvagen 24, SE-100 44 Stockholm, Sweden

Correspondence should be addressed to Giampiero Salvi, [email protected]

Received 13 March 2009; Revised 23 July 2009; Accepted 23 September 2009

Recommended by Gerard Bailly

This paper describes SynFace, a supportive technology that aims at enhancing audio-based spoken communication in adverseacoustic conditions by providing the missing visual information in the form of an animated talking head. Firstly, we describe thesystem architecture, consisting of a 3D animated face model controlled from the speech input by a specifically optimised phoneticrecogniser. Secondly, we report on speech intelligibility experiments with focus on multilinguality and robustness to audio quality.The system, already available for Swedish, English, and Flemish, was optimised for German and for Swedish wide-band speechquality available in TV, radio, and Internet communication. Lastly, the paper covers experiments with nonverbal motions drivenfrom the speech signal. It is shown that turn-taking gestures can be used to affect the flow of human-human dialogues. Wehave focused specifically on two categories of cues that may be extracted from the acoustic signal: prominence/emphasis andinteractional cues (turn-taking/back-channelling).

Copyright © 2009 Giampiero Salvi et al. This is an open access article distributed under the Creative Commons AttributionLicense, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properlycited.

1. Introduction

For a hearing impaired person, and for a normal hearingperson in adverse acoustic conditions, it is often necessaryto be able to lip-read as well as hear the person they aretalking with in order to communicate successfully. Apartfrom the lip movements, nonverbal visual information is alsoessential to keep a normal flow of conversation. Often, onlythe audio signal is available, for example, during telephoneconversations or certain TV broadcasts. The idea behindSynFace is to try to recreate the visible articulation of thespeaker, in the form of an animated talking head. The visualsignal is presented in synchrony with the acoustic speechsignal, which means that the user can benefit from the com-bined synchronised audiovisual perception of the originalspeech acoustics and the resynthesised visible articulation.When compared to video telephony solutions, SynFace hasthe distinct advantage that only the user on the receiving endneeds special equipment—the speaker at the other end canuse any telephone terminal and technology: fixed, mobilem,or IP-telephony.

Several methods have been proposed to drive the lipmovements of an avatar from the acoustic speech signalwith varying synthesis models and acoustic-to-visual maps.Tamura et al. [1] used hidden Markov models (HMMs) thatare trained on parameters that represent both auditory andvisual speech features. Similarly, Nakamura and Yamamoto[2] propose to estimate the audio-visual joint probabilityusing HMMs. Wen et al. [3] extract the visual informationfrom the output of a formant analyser. Al Moubayed et al.[4] map from the lattice output of a phonetic recogniserto texture parameters using neural networks. Hofer et al.[5] used trajectory hidden Markov models to predict visualspeech parameters from an observed sequence.

Most existing approaches to acoustic-to-visual speechmapping can be categorised as either regression based orclassification based. Regression-based systems try to mapfeatures of the incoming sounds into continuously varyingarticulatory (or visual) parameters. Classification-based sys-tems, such as SynFace, consider an intermediate phoneticlevel, thus solving a classification problem, and generatingthe final face parameters with a rule-based system. This

Page 2: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

2 EURASIP Journal on Audio, Speech, and Music Processing

(a) (b)

Figure 1: One of the talking head models used in SynFace, to theright running on a mobile device.

approach has proved to be more appropriate when the focusis on a real-life application, where additional requirementsare to be met, for example, speaker independence, and low-latency. Ohman and Salvi [6] compared two examples ofthe two paradigms. A time-delayed neural network was usedto estimate the face parameter trajectories from spectralfeatures of speech, whereas an HMM phoneme recogniserwas used to extract the phonetic information needed todrive the rule-based visual synthesis system. Although theresults are dependent on our implementation, we observedthat the first method could learn the general trend of theparameter trajectories, but was not accurate enough toprovide useful visual information. The same is also observedin Hofer et al. [5] and Massaro et al. [7]. (Although somespeech-reading support was obtained for isolated wordsfrom a single speaker in Massaro’s paper, this result didnot generalise well to extemporaneous speech from differentspeakers (which is indeed one of the goals of SynFace).) Thesecond method resulted in large errors in the trajectoriesin case of misrecognition, but provided, in general, morereliable results.

As for the actual talking head image synthesis, thiscan be produced using a variety of techniques, typicallybased on manipulation of video images [8, 9] parametricallydeformable models of the human face and/or speech organs[10, 11] or as a combination thereof [12]. In our system weemploy a deformable 3D model (see Section 2) for reasons ofspeed and simplicity.

This paper summarises the research that led to thedevelopment of the SynFace system and discusses a numberof aspects involved in its development, along with novelexperiments in multilinguality, dependency on the quality ofthe speech input, and extraction of nonverbal gestures fromthe acoustic signal.

The SynFace architecture is described for the first timeas a whole in Section 2; Section 3 describes the additionalnonverbal gestures. Experiments in German and with wide-band speech quality are described in Section 4. Finally,Section 5 discusses and concludes the paper.

2. SynFace Architecture

The processing chain in SynFace is illustrated in Figure 2.SynFace employs a specially developed real-time phonemerecognition system, that delivers information regarding thespeech signal-to-a speech animation module that renders thetalking face on the computer screen using 3D graphics. Thetotal delay from speech input to animation is only about200 milliseconds, which is low enough not to disturb the flowof conversation, (e.g., [13]). However, in order for face andvoice to be perceived coherently, the acoustic signal also hasto be delayed by the same amount [14].

2.1. Synthesis. The talking head model depicted in Figures1 and 2 includes face, tongue, and teeth, and is based onstatic 3D-wireframe meshes that are deformed using directparametrisation by applying weighted transformations totheir vertices according to principles first introduced byParke [15]. These transformations are in turn described byhigh-level articulatory parameters [16], such as jaw opening,lip rounding and bilabial occlusion. The talking head modelis lightweight enough to allow it to run at interactive rateson a mobile device [17]. A real-time articulatory controlmodel is responsible for driving the talking head’s lip, jawand tongue movements based on the phonetic input derivedby the speech recogniser (see below) as well as other facialmotion (nodding, eyebrow movements, gaze, etc.) furtherdescribed in Section 3.

The control model is based on the rule-based look-aheadmodel proposed by Beskow [16], but modified for low-latency operation. In this model, each phoneme is assigneda target vector of articulatory control parameters. To allowthe targets to be influenced by coarticulation, the targetvector may be under-specified, that is, some parametervalues can be left undefined. If a target is left undefined,the value is inferred from context using interpolation,followed by smoothing of the resulting trajectory. As anexample, consider the lip rounding parameter in aV1CCCV2

utterance where V1 is an unrounded vowel, CCC representsa consonant cluster and V2 is a rounded vowel. Accordingto the rules set, lip rounding would be unspecified for theconsonants, leaving these targets to be determined from thevowel context by linear interpolation from the unroundedV1, across the consonant cluster, to the rounded V2.

To allow for low-latency operation, the look-ahead modelhas been modified by limiting the look-ahead time window(presently a value of 100 milliseconds is used) which meansthat no anticipatory coarticulation beyond this window willoccur.

For comparison, the control model has also been eval-uated against several data-driven schemes [18]. In theseexperiments, different models are implemented and trainedto reproduce the articulatory patterns of a real speaker,based on a corpus of optical measurements. Two of themodels, (Cohen-Massaro and Ohman) are based on coar-ticulation models from speech production theory and oneuses artificial neural networks (ANNs). The different modelswere evaluated through a perceptual intelligibility experi-ment, where the data-driven models were compared against

Page 3: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

EURASIP Journal on Audio, Speech, and Music Processing 3

Acousticsource Phonetic

recogniserTrajectorygenerator

Renderer

Acousticdelay

Figure 2: Illustration of the signal flow in the SynFace system.

Table 1: Summary of intelligibility test of visual speech synthesiscontrol models, from Beskow [18].

Control model % keywords correct

Audio only 62.7

Cohen-Massaro 74.8

Ohman 75.3

ANN 72.8

Rule-based 81.1

the rule-based model as well as an audio-alone condition.In order to only evaluate the control models, and not therecognition, the phonetic input to all models was generatedusing forced alignment Sjolander [19]. Also, since the intentwas a general comparison of the relative merits of the controlmodels, that is, not only for real time applications, no low-latency constraints were applied in this evaluation. Thismeans that all models had access to all segments in eachutterance, but in practise the models differ in their use oflook-ahead information. The “Cohen-Massaro” model bydesign always uses all segments; the “Ohman” model looksahead until the next upcoming vowel; while the ANN model,which was specially conceived for low-latency operation,used a constant look-ahead of 50 milliseconds.

Table 1 summarises the results; all models give sig-nificantly increased speech intelligibility over the audio-alone case, with the rule-based model yielding the highestintelligibility score. While the data-driven models seem toprovide movements that are in some sense more naturalistic,the intelligibility is the single most important aspect of theanimation in SynFace, which is why the rule-based model isused in the system.

2.2. Phoneme Recognition. The constraints imposed on thephoneme recogniser (PR) for this application are speakerindependence, task independence and low latency. However,the demands on the PR performance are limited by the factthat some phonemes map to the same visemes (targets) forsynthesis.

The phoneme recogniser used in SynFace, is based ona hybrid of recurrent neural networks (RNNs) and hiddenMarkov models (HMMs) [20]. Mel frequency cepstralcoefficients (MFCCs) are extracted on 10 milliseconds spacedframes of speech samples. The neural networks are used

to estimate the posterior probabilities of each phoneticclass given a number of feature vectors in time [21]. Thenetworks are trained using Back Propagation through time[22] with a cross-entropy error measure [23]. This ensures anapproximately linear relation between the output activitiesof the RNN and the posterior probabilities of each phoneticclass, given the input observation. As in Strom [24], amixture of time delayed and recurrent connections is used.All the delays are positive, ensuring that no future context isused and thus reducing the total latency of the system at thecost of slightly lower recognition accuracy.

The posterior probabilities estimated by the RNN arefed into an HMM with the main purpose of smoothingthe results. The model defines a simple loop of phonemes,where each phoneme is a left-to-right three-state HMM. Aslightly modified Viterbi decoder is used to allow low-latencydecoding. Differently from the RNN model, the decodermakes use of some future context (look-ahead). The amountof look-ahead is one of the parameters that can be controlledin the algorithm.

During the Synface project (IST-2001-33327), the recog-niser was trained and evaluated on the SpeechDat recordings[25] for three languages: Swedish, English and Flemish.In Salvi [20, 26], the effect of limiting the look-ahead inthe Viterbi decoder was studied. No improvements in theresults were observed for look-ahead lengths greater than100 milliseconds. In the SynFace system, the look-aheadlength was further limited to 30 milliseconds, resulting in arelative 4% drop in performance in terms of correct frames.

3. Nonverbal Gestures

While enhancing speech perception through visible artic-ulation has been the main focus of SynFace, recent workhas been aimed at improving the overall communicativeexperience through nonarticulatory facial movements. Itis well known that a large part of information transferin face-to-face interaction is nonverbal, and it has beenshown that speech intelligibility is also affected by nonverbalactions such as head movements [27]. However, while thereis a clear correlation between the speech signal and thearticulatory movements of the speaker that can be exploitedfor driving the face articulation, it is less clear how toprovide meaningful nonarticulatory movements based solelyon the acoustics. We have chosen to focus on two classes

Page 4: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

4 EURASIP Journal on Audio, Speech, and Music Processing

of nonverbal movements that have found to play importantroles in communication and that also may be driven byacoustic features that can be reliably estimated from speech.The first category is speech-related movements linked toemphasis or prominence, the second category is gesturesrelated to interaction control in a dialogue situation. For thetime being, we have not focused on expressiveness of thevisual synthesis in terms of emotional content as in Cao et al.[28].

Hadar et al. [29] found that increased head movementactivity co-occurs with speech, and Beskow et al. [30] found,by analysing facial motion for words in focal and nonfocalposition, that prominence is manifested visually in all partsof the face, and that the particular realisation chosen isdependent on the context. In particular these results suggestthat there is not one way of signalling prominence visuallybut it is likely that several cues are used interchangeably orin combination. One issue that we are currently working onis how to reliably extract prominence based on the audiosignal alone, with the goal of driving movements in thetalking head. In a recent experiment Al Moubayed et al.[4] it was shown that adding small eyebrow movementson syllables with large pitch movements, resulted in asignificant intelligibility improvement over the articulation-only condition, but less so than a condition where manuallylabelled prominence was used to drive the gestures.

When people are engaged in face-to-face conversation,they take a great number of things into consideration in orderto manage the flow of the interaction. We call this interactioncontrol—the term is wider than turn-taking and does notpresuppose the existence of “turns.” Examples of featuresthat play a part in interaction control include auditorycues such as pitch, intensity, pause and disfluency, hyper-articulation; visual cues such as gaze, facial expressions,gestures, and mouth movements (constituting the regulatorscategory above) and cues like pragmatic, semantic, andsyntactic completeness.

In order to investigate the effect of visual interactioncontrol cues in a speech driven virtual talking head, weconducted an experiment with human-human interactionover a voice connection supplemented by the SynFace talkinghead at each end, where visual interaction control gestureswere automatically controlled from the audio stream. Thegoal of the experiment was to find out to what degree subjectswere affected by the interaction control cues. In what followsis a summary, for full details see Edlund and Beskow [31].

In the experiment, a bare minimum of gestures wasimplemented that can be said to represent a stylised versionof the gaze behaviours observed by Kendon [32] and recentgaze-tracking experiments [33].

(i) A turn-taking/keeping gesture, where the avatarmakes a slight turn of the head to the side incombination with shifting the gaze away a little,signalling a wish to take or keep the floor.

(ii) A turn-yielding/listening gesture, where the avatarlooks straight forward, at the subject, with slightlyraised eyebrows, signalling attention and willingnessto listen.

(iii) A feedback/agreement gesture, consisting of a smallnod. In the experiment described here, this gestureis never used alone, but is added at the end of thelistening gesture to add to its responsiveness. In thefollowing, simply assume it is present in the turnyielding/listening gesture.

The audio-signal from each participant was processed bya voice activity detector (VAD). The VAD reports a changeto the SPEECH state each time it detected a certain numberof consecutive speech frames whilst in the SILENCE state,and vice-versa. Based on these state transitions, gestures weretriggered in the respective SynFace avatar.

To be able to assess the degree to which subjects wereinfluenced by the gestures, the avatar on each side couldwork in one of two modes: ACTIVE or PASSIVE. In theACTIVE mode, gestures were chosen as to encourage oneparty to take and keep turns, while PASSIVE mode impliedthe opposite—to discourage the user to speak. In order tocollect balanced data of the two participants behaviour, themodes were shifted regularly (every 10 turns), but they werealways complementary—ACTIVE on one side and PASSIVEon the other. The number 10 was chosen to be small enoughto make sure that both parties got exposed to both modesseveral times during the test (10 minutes), but large enoughto allow subjects to accommodate to the situation.

The subjects were placed in separate rooms and equippedwith head-sets connected to a Voice-over-IP call. On eachside, the call is enhanced by the SynFace animated talkinghead representing the other participant, providing real-time lip-synchronised visual speech animation. The task wasto speak about any topic freely for around ten minutes.There were 12 participants making up 6 pairs. None of theparticipants had any previous knowledge of the experimentsetup.

The results were analysed by counting the percentageof times that the turn changed when a speaker paused.The percentage of all utterances followed by a turn changeis larger under the PASSIVE condition than under theACTIVE condition for each participant without exception.The difference is significant (P < .01), which shows thatsubjects were consistently affected by the interaction controlcues in the talking head. As postinterviews revealed that mostsubjects never even noticed the gestures consciously, and nosubject connected them directly to interaction control, thisresult shows that it is possible to unobtrusively influencethe interaction behaviour of two interlocutors in a givendirection—that is to make a person take the floor more orless often—by way of facial gestures in an animated talkinghead in the role of an avatar.

4. Evaluation Experiments

In the SynFace application, speech intelligibility enhance-ment is the main function. Speech reading and audio-visualspeech intelligibility have been extensively studied by manyresearchers, for natural speech as well as for visual speechsynthesis systems driven by text or phonetically transcribedinput. Massaro et al. [7], for example, evaluated visual-only intelligibility of a speaker dependent speech driven

Page 5: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

EURASIP Journal on Audio, Speech, and Music Processing 5

system on isolated words. To date, however, we have notseen any published results on speaker independent speechdriven facial animation systems, where the intelligibilityenhancement (i.e., audiovisual compared to audio-onlycondition) has been investigated. Below, we report on twoexperiments were audiovisual intelligibility of SynFace hasbeen evaluated for different configurations and languages.

The framework adopted in SynFace allows for evaluationof the system at different points in the signal chain shown inFigure 2. We can measure accuracy

(i) at the phonetic level, by measuring the phoneme(viseme) accuracy of the speech recogniser,

(ii) at the face parameter level, by computing the distancebetween the face parameters generated by the systemand the optimal trajectories, for example, trajectoriesobtained from phonetically annotated speech,

(iii) at the intelligibility level, by performing listeningtests with hearing impaired subjects, or with normalhearing subjects and a degraded acoustic signal.

The advantage of the first two methods is simplicity. Thecomputations can be performed automatically, if we assumethat a good reference is available (phonetically annotatedspeech). The third method, however, is the most reliablebecause it tests the effects of the system as a whole.

Evaluating the Phoneme Recogniser. Measuring the perfor-mance at the phonetic level can be done in at least two ways:By measuring the percentage of frames that are correctlyclassified, or by computing the Levenshtein (edit) distanceLevenshtein [34] between the string of phonemes outputby the recogniser and the reference transcription. The firstmethod does not explicitly consider the stability of theresults in time and, therefore, may overestimate the perfor-mance of a recogniser that produces many short insertions.These insertions, however, do not necessarily result in adegradation of the face parameter trajectories, because thearticulatory model, the face parameter generation is basedon, often acts as a low-pass filter. On the other hand, theLevenshtein distance does not consider the time alignment ofthe two sequences, and may result in misleading evaluationin the case that two phonetic subsequences that are not co-occurring in time are aligned by mistake. To make the lattermeasure homogeneous with the correct frames %, we expressit in terms of accuracy, defined as (1 − l/n) × 100, where l isLevenshtein (Edit) distance and n the length of the referencetranscription.

Intelligibility Tests. Evaluating the intelligibility is performedby listening tests with a number of hearing impaired ornormal hearing subjects. Using normal hearing subject anddistorting the audio signal has been shown to be a viablesimulation of perception by hearing impaired [35, 36]. Thespeech material is presented to the subjects in differentconditions. These may include audio alone, audio and naturalface, audio and synthetic face. In the last case, the syntheticface may be driven by different methods (e.g., differentversions of the PR that we want to compare). It may also

−2

−1

0

1

Del

taSR

T(d

B)

50 55 60 65 70

Correct frames (%)

SRT versus % correct frames (5 subjects)

Figure 3: Delta SRT versus correct frames % for three differentrecognisers (correlation r = −0.89) on a 5-subject listening test.

be driven by carefully obtained annotations of the speechmaterial, if the aim is to test the effects of the visual synthesismodels alone.

Two listening test methods have been used in the currentexperiments. The first method is based on a set of carefullydesigned short sentences containing a number of key-words.The subject’s task is to repeat the sentences, and intelligibilityis measured in terms of correctly recognised key-words. Incase of normal hearing subjects, the acoustic signal may bedegraded by noise in order to simulate hearing impairment.In the following, we will refer to this methodology as “key-word” test.

The second methodology Hagerman and Kinnefors [37]relies on the adaptive use of noise to assess the level ofintelligibility. Lists of 5 words are presented to the subjectsin varying noise conditions. The signal-to-noise ratio (SNRdB) is adjusted during the test until the subject is able tocorrectly report about 50% of the words. This level of noise isreferred to as the Speech Reception Threshold (SRT dB) andindicates the amount of noise the subject is able to toleratebefore the intelligibility drops below 50%. Lower values ofSRT correspond to better performance (the intelligibility ismore robust to noise). We will refer to this methodology as“SRT” test.

SRT Versus Correct Frames %. Figure 3 relates the changeof SRT level between audio-alone and SynFace conditions(Delta SRT) to the correct frames % of the correspondingphoneme recogniser. Although the data is based on a smalllistening experiment (5 subjects), the high correlation shownin the figure motivates the use of the correct frames % mea-sure for developmental purposes. We believe, however, thatreliable evaluations should always include listening tests. Inthe following, we report results on the recent developmentsof SynFace using both listening tests and PR evaluation.

Page 6: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

6 EURASIP Journal on Audio, Speech, and Music Processing

Table 2: Number of connections in the RNN and correct frames %of the SynFace RNN phonetic classifiers.

Language Connections Correct frames %

English 184,848 46.1

Flemish 186,853 51.0

German 541,430 61.0

Swedish 541,250 54.2

These include the newly developed German recogniser, wide-versus narrow-band speech recognition experiments andcross-language tests. All experiments are performed with thereal-time, low-latency implementation of the system, thatis, the phoneme recogniser uses 30 milliseconds look-aheadlength, and the total delay of the system in the intelligibilitytests is 200 milliseconds.

4.1. SynFace in German. To extend SynFace to German,a new recogniser was trained on the SpeechDat Germanrecordings. These consist of around 200 hours of telephonespeech spoken by 4000 speakers. As for the previouslanguages, the HTK-based RefRec recogniser Lindberg et al.[38] was trained and used to derive phonetic transcriptionsof the corpus. Whereas the recogniser for Swedish, Englishand Flemish, was trained exclusively on the phoneticallyrich sentences, the full training set, also containing isolatedwords, digits, and spellings, was used to train the Germanmodels. Table 2 shows the results in terms of correct frames% for the different languages. Note however that these resultsare not directly comparable because they are obtained ondifferent test sets.

The same synthesis rules used for Swedish are appliedto the German system, simply by mapping the phoneme(viseme) inventory of the two languages.

To evaluate the German version of the SynFace system,a small “key-word” intelligibility test was performed. A setof twenty short (4–6 words) sentences from the Gottingersatsztest set [39], spoken by a male native German speaker,were presented to a group of six normal hearing Germanlisteners. The audio presented to the subjects was degraded inorder to avoid ceiling effects, using a 3-channel noise excitedvocoder shannon et al. [40]. This type of signal degradationhas been used in the previous audio-visual intelligibilityexperiments Siciliano et al. [41] and can be viewed as away of simulating the information reduction experienced bycochlear implant patients. Clean speech was used to driveSynFace. 10 sentences were presented with audio-only and10 sentences were presented with SynFace support. Subjectswere presented with four training sentences before the teststarted. The listeners were instructed to watch the screen andwrite down what they perceived.

Figure 4 summarises the results for each subject. Themean score (% correctly recognised key-words) for the audioonly condition was extremely low (2.5%). With SynFacesupport, a mean score of 16.7% was obtained. While therewas a large intersubject variability, subjects consistentlyshowed to take advantage of the use of SynFace. An ANOVA

0

10

20

30

40

Cor

rect

wor

ds(%

)

Audio-alone Synface

Condition

Listening experiments (6 subjects)

2

3

145

61

2364

5

Figure 4: Subjective evaluation results for German version ofSynFace (% correct word recognition).

analysis shows significant differences (P < .01) between theaudio-alone and SynFace conditions.

4.2. Narrow- Versus Wide-Band PR. In the Hearing at Homeproject, SynFace is employed in a range of applications thatinclude speech signals that are streamed through differentmedia (Telephone, Internet, TV). The signal is often of ahigher quality compared to the land-line telephone settings.This opens the possibility for improvements in the signalprocessing part of the system.

In order to take advantage of the available audio bandin these applications, the SynFace recogniser was trainedon wide-band speech data from the SpeeCon corpus [42].SpeeCon contains recordings in several languages and con-ditions. Only recordings in office settings of Swedish werechosen. The corpus contains word level transcriptions, andannotations for speaker noise, background noise, and filledpauses. As in the SpeechDat training, the silence at theboundaries of every utterance was reduced, in order toimprove balance between the number of frames for thesilence class and for any other phonetic class. Differentlyfrom the SpeechDat training, NALIGN Sjolander [19] wasused in order to create time aligned phonetic transcriptionsof the corpus based on the orthographic transcriptions.

The bank of filters, used to compute the MFCCs that areinput to the recogniser, was defined in a way that the filtersbetween 0 and 4 kHz coincide with the narrow-band filter-bank definition. Additional filters are added for the upper4 kHz frequencies offered by the wide-band signal.

Table 3 shows the results for the network trained on theSpeeCon database. The results obtained on the SpeechDatmaterial are also given for comparison. Note, however, thatthese results cannot be compared directly because the testswere performed on different test sets.

Page 7: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

EURASIP Journal on Audio, Speech, and Music Processing 7

Table 3: Comparison between the SpeechDat telephone quality(TF), SpeeCon narrow-band (NB) and SpeeCon wide-band (WB)recognisers. Results are given in terms of correct frames % forphonemes (ph) and visemes (vi), and accuracy.

Database SpeechDat SpeeCon

Data size (ca. hours) 200 40

Speakers (#) 5000 550

Speech quality TF NB WB

Sampling (kHz) 8 8 16

Correct frames (% ph) 54.2 65.2 68.7

Correct frames (% vi) 59.3 69.0 74.5

Accuracy (% ph) 56.5 62.2 63.2

−11

−10

−9

−8

−7

SRT

(dB

)

Audio-alone Narrow-band Wide-band

Condition

Listening experiments (5 subjects)

22

25

5

51

1

4

3

3

3

1

4

4

Figure 5: Speech Reception Threshold (SRT dB) for differentconditions and subjects.

In order to have a more controlled comparison betweenthe narrow- and the wide-band networks for Swedish, anetwork was trained on a downsampled (8 kHz) version ofthe same SpeeCon database. The middle column of Table 3shows the results for the networks trained and tested onthe narrow-band (downsampled) version of the SpeeCondatabase. Results are shown in terms of % of correct framesfor phonemes, visemes, and phoneme accuracy.

Finally, a small-scale “SRT” intelligibility experiment,was performed in order to confirm the improvement inperformance that we see in the wide-band case. The con-ditions include audio alone and SynFace driven by differentversions of the recogniser. The tests were carried out usingfive normal hearing subjects. The stimuli consisted of listsof 5 words randomly selected from a set of 50 words. Atraining session was performed before the real test to controlthe learning effect. Figure 5 shows the SRT levels obtainedin the different conditions, where each line correspondsto a subject. An ANOVA analysis and successive multiplecomparison analysis confirm that there is a significant

b

a

c

d

e

b

c

d

f

g

(a)

b

a

c

d

e

b

c

d

e

p

t

(b)

p(b)

p(c)

p(d)

p(e)

p(b)

p(c)

p(d)

p( f )

p(g)

(c)

Figure 6: Structure of the language models mappers. (a) Phoneticmapper: only identical phonemes are matched between the twolanguages, (b) best match mapper: the phoneme are blindlymatched in a way that recognition results are maximised, (c) linearregression mapper: posterior probabilities for the target languageare estimated from the outputs of the phonetic recogniser.

Table 4: correct frames % for different languages (columns)recognised by different models (rows). The languages are: German(de), English (en), Flemish (fl) and Swedish (sv). Numbers inparentheses are the % of correct frames for perfect recognition,given the mismatch in phonetic inventory across languages.

de sv fl en

de 61.0 (100) 30.3 (82.6) 27.1 (73.5) 26.2 (71.6)

sv 31.5 (86.2) 54.2 (100) 26.3 (72.1) 23.5 (74.2)

fl 34.2 (85.7) 31.6 (77.9) 51.0 (100) 26.9 (69.8)

en 24.5 (74.6) 23.7 (72.3) 21.5 (66.8) 46.1 (100)

decrease (improvement) of SRT (P < .001) for the wide-band recogniser over the narrow-band trained network andthe audio-alone condition.

4.3. Multilinguality. SynFace is currently available inSwedish, German, Flemish and English. In order toinvestigate the possibility of using the current recognitionmodels on new languages, we performed cross-languageevaluation tests.

Each language has its unique phonetic inventory. Inorder to map between the inventory of the recognitionmodel and that of the target language, we considered threedifferent paradigms illustrated in Figure 6. In the first case(Figure 6(a)) we relay on perfect matching of the phonemes.This is a very strict evaluation criterion because it does nottake into account the acoustic and visual similarities betweenphonemes in different languages that do not share the samephonetic code.

Table 4 presents the correct frames % when only match-ing phonetic symbols are considered. The numbers inparentheses show the highest possible performance, giventhe fact that some of the phonemes in the test set do notexist in the recognition model. As expected, the accuracyof recognition drops drastically when we use cross-languagemodels. This could be considered as a lower bound toperformance.

Page 8: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

8 EURASIP Journal on Audio, Speech, and Music Processing

0

10

20

30

40

50

60

Cor

rect

fram

era

te(%

)

German PR Flemish PR English PR

Phoneme mapperBest match mapper

Linear regression (NN)Swedish PR

Figure 7: correct frames % for the three mappers when the targetlanguage is Swedish. The dashed line represents the results obtainedwith the Swedish phoneme recogniser (language matching case).

The second mapping criterion depicted in Figure 6(b),considers as correct the association between model andtarget phonemes that was most frequently adopted by therecognition models on that particular target language. Ifwe consider all possible maps between the set of modelphonemes and the set of target phonemes, this correspondsto an upper bound of the results. Compared to the resultsin Table 4, this evaluation method gives about 10% increasedcorrect frames in average. In this case, there is no guaranteethat the chosen mapping bares phonetic significance.

The previous mappings were introduced to allow forsimple cross-language evaluation of the phonetic recognisers.Figure 6(c) shows a mapping method that is more realisticin terms of system performance. In this case we do notmap between two discrete sets of phonemes, but, rather,between the posterior probabilities of the first set and thesecond set. This way, we can, in principle, obtain betterresults than the above upper bound, and possibly even betterresults than for the original language. In the experiments theprobability mapping was performed by a one-layer neuralnetwork that implements linear regression. Figure 7 showsthe results when the target language is Swedish, and thephoneme recognition (PR) models were trained on German,English and Flemish. Only 30 minutes of an independenttraining set were used to train the linear regression mapper.The performance of this mapper is above the best matchresults, and comes close to the Swedish PR results (dashedline in the figure) for the German PR models.

5. Conclusions

The purpose of SynFace is to enhance spoken communi-cation for the hearing impaired, rather than solving theacoustic-to-visual speech mapping per se. The methods

employed here are, therefore, tailored to achieving this goalin the most effective way. Beskow [18] showed that, whereasdata-driven visual synthesis resulted in more realistic lipmovements, the rule-based system enhanced the intelligi-bility. Similarly, mapping from the acoustic speech directlyinto visual parameters is an appealing research problem.However, when the ambition is to develop a tool that can beapplied in real-life conditions, it is necessary to constrain theproblem. The system discussed in this paper

(i) works in real time and with low latency, allowing real-istic conditions for a natural spoken communication,

(ii) is light-weight and can be run on standard commer-cially available hardware,

(iii) is speaker independent, allowing the user to commu-nicate with any person,

(iv) is being developed for different languages (currently,Swedish, English, Flemish, and German are avail-able),

(v) is optimised for different acoustic conditions, rangingfrom telephone speech quality to wide-band speechavailable in, for example, Internet communicationsand radio/TV broadcasting,

(vi) is being extensively evaluated in realistic settings,with hearing impaired subjects or by simulatinghearing impairment.

Even though speech intelligibility is the focus of theSynFace system, extra-linguistic aspects of speech commu-nication have also been described in the paper. Modellingnonverbal gestures proved to be a viable way of enhancingthe turn-taking mechanism in telephone communication.

Future work will be aimed at increasing the generalityof the methods, for example, by studying ways to achievelanguage independence or by simplifying the process ofoptimising the system to a new language, based on thepreliminary results shown in this paper. Reliably extractingextra-linguistic information, as well as synthesis and evalu-ation of nonverbal gestures will also be the focus of futurework.

Acknowledgments

The work presented here was funded in part by EuropeanCommission Project IST-045089 (Hearing at Home) andSwedish Research Council Project 621-2005-3488 (Mod-elling multimodal communicative signal and expressivespeech for embodied conversational agents).

References

[1] M. Tamura, T. Masuko, T. Kobayashi, and K. Tokuda, “Visualspeech synthesis based on parameter generation from hmm:speech-driven and text-and-speechdriven approaches,” in Pro-ceedings of the Auditory-Visual Speech Processing (AVSP ’98),pp. 221–226, 1998.

[2] S. Nakamura and E. Yamamoto, “Speech-to-lip movementsynthesis by maximizing audio-visual joint probability basedon the EM algorithm,” Journal of VLSI Signal Processing

Page 9: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

EURASIP Journal on Audio, Speech, and Music Processing 9

Systems for Signal, Image, and Video Technology, vol. 27, no.1-2, pp. 119–126, 2001.

[3] Z. Wen, P. Hong, and T. Huang, “Real time speech driven facialanimation using formant analysis,” in Proceedings of the IEEEInternational Conference on Multimedia and Expo (ICME ’01),pp. 817–820, 2001.

[4] S. Al Moubayed, M. De Smet, and H. Van Hamme, “Lipsynchronization:from phone lattice to PCA eigenprojectionsusing neural networks,” in Proceedings of the Biennial Con-ference of the International Speech Communication Association(Interspeech ’08), Brisbane, Australia, 2008.

[5] G. Hofer, J. Yamagishi, and H. Shimodaira, “Speech-driven lipmotion generation with a trajectory hmm,” in Proceedings ofthe Biennial Conference of the International Speech Communi-cation Association (Interspeech ’08), Brisbane, Australia, 2008.

[6] T. Ohman and G. Salvi, “Using HMMs and ANNs for mappingacoustic to visual speech,” TMH-QPSR, vol. 40, no. 1-2, pp.45–50, 1999.

[7] D. Massaro, J. Beskow, M. Cohen, C. Fry, and T. Rodgriguez,“Picture my voice: audio to visual speech synthesis usingartificial neural networks,” in Proceedings of the InternationalConference on Auditory-Visual Speech Processing (ISCA ’99),1999.

[8] T. Ezzat, G. Geiger, and T. Poggio, “Trainable videorealisticspeech animation,” in Proceedings of the 29th Annual Con-ference on Computer Graphics and Interactive Techniques, pp.388–398, ACM, New York, NY, USA, 2002.

[9] K. Liu and J. Ostermann, “Realistic facial animation system forinteractive services,” in Proceedings of the Biennial Conferenceof the International Speech Communication Association (Inter-speech ’08), Brisbane, Australia, 2008.

[10] M. Cohen and D. Massaro, “Models and techniques incomputer animation,” in Modeling Coarticulation in SyntheticVisual Speech, vol. 92, Springer, Tokyo, Japan, 1993.

[11] M. Zelezny, Z. Krnoul, P. Cısar, and J. Matousek, “Design,implementation and evaluation of the Czech realistic audio-visual speech synthesis,” Signal Processing, vol. 86, no. 12, pp.3657–3673, 2006.

[12] B. J. Theobald, J. A. Bangham, I. A. Matthews, and G. G.Cawley, “Near-videorealistic synthetic talking faces: imple-mentation and evaluation,” Speech Communication, vol. 44,no. 1–4, pp. 127–140, 2004.

[13] N. Kitawaki and K. Itoh, “Pure delay effects on speech qualityin telecommunications,” IEEE Journal on Selected Areas inCommunications, vol. 9, no. 4, pp. 586–593, 1991.

[14] M. McGrath and Q. Summerfield, “Intermodal timing rela-tions and audio-visual speech recognition by normal-hearingadults,” Journal of the Acoustical Society of America, vol. 77, no.2, pp. 678–685, 1985.

[15] F. I. Parke, “Parameterized models for facial animation,” IEEEComputer Graphics and Applications, vol. 2, no. 9, pp. 61–68,1982.

[16] J. Beskow, “Rule-based visual speech synthesis,” in Proceedingsof the European Conference on Speech Communication andTechnology (Eurospeech ’95), pp. 299–302, Madrid, Spain,1995.

[17] T. Gjermani, Integration of an animated talking face model ina portable device for multimodal speech synthesis, M.S. thesis,Department for Speech, Music and Hearing, KTH, School ofComputer Science and Communication, Stockholm, Sweden,2008.

[18] J. Beskow, “Trainable articulatory control models for visualspeech synthesis,” International Journal of Speech Technology,vol. 7, no. 4, pp. 335–349, 2004.

[19] K. Sjolander, “An HMM-based system for automatic segmen-tation and alignment of speech,” in Proceedings of Fonetik, pp.93–96, Umea, Sweden, 2003.

[20] G. Salvi, “Truncation error and dynamics in very low latencyphonetic recognition,” in Proceedings of the Non Linear SpeechProcessing (NOLISP ’03), Le Croisic, France, 2003.

[21] A. J. Robinson, “Application of recurrent nets to phoneprobability estimation,” IEEE Transactions on Neural Networks,vol. 5, no. 2, pp. 298–305, 1994.

[22] P. J. Werbos, “Backpropagation through time: what it does andhow to do it,” Proceedings of the IEEE, vol. 78, no. 10, pp. 1550–1560, 1990.

[23] H. Bourlard and N. Morgan, “Continuous speech recognitionby connectionist statistical methods,” IEEE Transactions onNeural Networks, vol. 4, no. 6, pp. 893–909, 1993.

[24] N. Strom, “Development of a recurrent time-delay neural netspeech recognition system,” TMH-QPSR, vol. 26, no. 4, pp. 1–15, 1992.

[25] K. Elenius, “Experiences from collecting two Swedish tele-phone speech databases,” International Journal of SpeechTechnology, vol. 3, no. 2, pp. 119–127, 2000.

[26] G. Salvi, “Dynamic behaviour of connectionist speech recog-nition with strong latency constraints,” Speech Communica-tion, vol. 48, no. 7, pp. 802–818, 2006.

[27] K. G. Munhall, J. A. Jones, D. E. Callan, T. Kuratate, and E.Vatikiotis-Bateson, “Visual prosody and speech intelligibility:head movement improves auditory speech perception,” Psy-chological Science, vol. 15, no. 2, pp. 133–137, 2004.

[28] Y. Cao, W. C. Tien, P. Faloutsos, and F. Pighin, “Expressivespeech-driven facial animation,” ACM Transactions on Graph-ics, vol. 24, no. 4, pp. 1283–1302, 2005.

[29] U. Hadar, T. J. Steiner, E. C. Grant, and F. C. Rose, “Kinematicsof head movements accompanying speech during conversa-tion,” Human Movement Science, vol. 2, no. 1-2, pp. 35–46,1983.

[30] J. Beskow, B. Granstrom, and D. House, “Analysis andsynthesis of multimodal verbal and non-verbal interactionfor animated interface agents,” in Proceedings of the Inter-national Workshop on Verbal and Nonverbal CommunicationBehaviours, vol. 4775 of Lecture Notes in Computer Science, pp.250–263, 2007.

[31] J. Edlund and J. Beskow, “Pushy versus meek-using avatarsto in uence turn-taking behaviour,” in Proceedings of theBiennial Conference of the International Speech CommunicationAssociation (Interspeech ’07), Antwerp, Belgium, 2007.

[32] A. Kendon, “Some functions of gaze-direction in socialinteraction,” Acta Psychologica, vol. 26, no. 1, pp. 22–63, 1967.

[33] V. Hugot, Eye gaze analysis in human-human commuication,M.S. thesis, Department for Speech, Music and Hearing, KTH,School of Computer Science and Communication, Stockholm,Sweden, 2007.

[34] V. I. Levenshtein, “Binary codes capable of correcting dele-tions, insertions and reversals,” Soviet Physics Doklady, vol. 10,p. 707, 1966.

[35] L. E. Humes, B. Espinoza-Varas, and C. S. Watson, “Modelingsensorineural hearing loss. I. Model and retrospective evalua-tion,” Journal of the Acoustical Society of America, vol. 83, no.1, pp. 188–202, 1988.

[36] L. E. Humes and W. Jesteadt, “Models of the effects ofthreshold on loudness growth and summation,” Journal of theAcoustical Society of America, vol. 90, no. 4, pp. 1933–1943,1991.

Page 10: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

10 EURASIP Journal on Audio, Speech, and Music Processing

[37] B. Hagerman and C. Kinnefors, “Efficient adaptive methodsfor measuring speech reception threshold in quiet and innoise,” Scandinavian Audiology, vol. 24, no. 1, pp. 71–77, 1995.

[38] B. Lindberg, F. T. Johansen, N. Warakagoda, et al., “A noiserobust multilingual reference recogniser based on Speech-Dat(II),” in Proceedings of the International Conference onSpoken Language Processing (ICSLP ’00), 2000.

[39] M. Wesselkamp, Messung und modellierung der verstandlichkeitvon sprache, Ph.D. thesis, Universitat Gottingen, 1994.

[40] R. V. Shannon, F.-G. Zeng, V. Kamath, J. Wygonski, and M.Ekelid, “Speech recognition with primarily temporal cues,”Science, vol. 270, no. 5234, pp. 303–304, 1995.

[41] C. Siciliano, G. Williams, J. Beskow, and A. Faulkner, “Evalu-ation of a multilingual synthetic talking face as a communica-tion aid for the hearing impaired,” in Proceedings of the 15thInternational Conference of Phonetic Sciences (ICPhS ’03), pp.131–134, Barcelona, Spain, 2003.

[42] D. Iskra, B. Grosskopf, K. Marasek, H. V. D. Heuvel, F. Diehl,and A. Kiessling, “Speecon—speech databases for consumerdevices: database specification and validation,” in Proceedingsof the International Conference on Language Resources andEvaluation (LREC ’02), pp. 329–333, 2002.

Page 11: SynFace—Speech-DrivenFacialAnimationfor … · EURASIP Journal on Audio, Speech, and Music Processing ... recognition system, ... Summary of intelligibility test of visual speech

Photograph © Turisme de Barcelona / J. Trullàs

Preliminary call for papers

The 2011 European Signal Processing Conference (EUSIPCO 2011) is thenineteenth in a series of conferences promoted by the European Association forSignal Processing (EURASIP, www.eurasip.org). This year edition will take placein Barcelona, capital city of Catalonia (Spain), and will be jointly organized by theCentre Tecnològic de Telecomunicacions de Catalunya (CTTC) and theUniversitat Politècnica de Catalunya (UPC).EUSIPCO 2011 will focus on key aspects of signal processing theory and

li ti li t d b l A t f b i i ill b b d lit

Organizing Committee

Honorary ChairMiguel A. Lagunas (CTTC)

General ChairAna I. Pérez Neira (UPC)

General Vice ChairCarles Antón Haro (CTTC)

Technical Program ChairXavier Mestre (CTTC)

Technical Program Co Chairsapplications as listed below. Acceptance of submissions will be based on quality,relevance and originality. Accepted papers will be published in the EUSIPCOproceedings and presented during the conference. Paper submissions, proposalsfor tutorials and proposals for special sessions are invited in, but not limited to,the following areas of interest.

Areas of Interest

• Audio and electro acoustics.• Design, implementation, and applications of signal processing systems.

l d l d d

Technical Program Co ChairsJavier Hernando (UPC)Montserrat Pardàs (UPC)

Plenary TalksFerran Marqués (UPC)Yonina Eldar (Technion)

Special SessionsIgnacio Santamaría (Unversidadde Cantabria)Mats Bengtsson (KTH)

FinancesMontserrat Nájar (UPC)• Multimedia signal processing and coding.

• Image and multidimensional signal processing.• Signal detection and estimation.• Sensor array and multi channel signal processing.• Sensor fusion in networked systems.• Signal processing for communications.• Medical imaging and image analysis.• Non stationary, non linear and non Gaussian signal processing.

Submissions

Montserrat Nájar (UPC)

TutorialsDaniel P. Palomar(Hong Kong UST)Beatrice Pesquet Popescu (ENST)

PublicityStephan Pfletschinger (CTTC)Mònica Navarro (CTTC)

PublicationsAntonio Pascual (UPC)Carles Fernández (CTTC)

I d i l Li i & E hibiSubmissions

Procedures to submit a paper and proposals for special sessions and tutorials willbe detailed at www.eusipco2011.org. Submitted papers must be camera ready, nomore than 5 pages long, and conforming to the standard specified on theEUSIPCO 2011 web site. First authors who are registered students can participatein the best student paper competition.

Important Deadlines:

P l f i l i 15 D 2010

Industrial Liaison & ExhibitsAngeliki Alexiou(University of Piraeus)Albert Sitjà (CTTC)

International LiaisonJu Liu (Shandong University China)Jinhong Yuan (UNSW Australia)Tamas Sziranyi (SZTAKI Hungary)Rich Stern (CMU USA)Ricardo L. de Queiroz (UNB Brazil)

Webpage: www.eusipco2011.org

Proposals for special sessions 15 Dec 2010Proposals for tutorials 18 Feb 2011Electronic submission of full papers 21 Feb 2011Notification of acceptance 23 May 2011Submission of camera ready papers 6 Jun 2011