photorealistic adaptation and interpolation of …

1
PHOTOREALISTIC ADAPTATION AND INTERPOLATION OF FACIAL EXPRESSIONS USING HMMS AND AAMS FOR AUDIO-VISUAL SPEECH SYNTHESIS BACKGROUND & OBJECTIVES Intelligent agents have a continuous presence in everyday life and speech synthesis (both acoustic & audio-visual) constitutes a vital asset for human – computer interaction Achieving a high degree of naturalness in HCI depends on the ability of the agent to express emotions However, there is a huge data overhead when considering synthesis of expressive speech in a large non-discrete emotional space We tackle the problem this problem by: using HMM adaptation to adapt an existing audio-visual speech synthesis HMM set to a new emotion using a small amount of adaptation data employing HMM interpolation to combine HMM sets to generate speech with intermediate styles ACTIVE APPEARANCE MODELS (AAM) Weights Neutral Anger Happiness Sadness 0.1 − 0.9 8.7 4.35 0 86.96 0.3 − 0.7 18.18 0 0 81.82 0.5 − 0.5 45.83 0 4.17 50 0.7 − 0.3 83.33 0 0 16.67 0.9 − 0.1 91.3 4.35 0 4.35 EXPERIMENTS & RESULTS CONCLUSIONS We can successfully adapt an HMM-based audio-visual speech synthesis system to a target emotion using a small number of adaptation data. Level of expressiveness increases with number of adaptation sentences used. HMM interpolation gives us audio-visual speech with intermediate characteristics between the interpolated emotions. DNN version of the system can be found in [4]. Panagiotis P. Filntisis, Athanasios Katsamanis and Petros Maragos School of ECE, National Technical University of Athens, 15773 Athens, Greece Athena Research and Innovation Center, 15125 Maroussi, Greece HMM-BASED AUDIO-VISUAL SPEECH SYNTHESIS [1] Example of the first eigentexture and the variations it causes to the mean texture Face shape =ത + σ = Face texture () = () + σ = : mean shape : eigenshape coefficients (): mean texture : eigentexture coefficients , : original mean and covariance matrix , : adapted mean and covariance matrix , : transformation bias and matrix Extract acoustic and visual features Train HMMs with EM algorithm Cluster similar phonetic contexts using decision trees TRAINING Analyze input text Generate features from HMMs Reconstruct audio and video SYNTHESIS Adaptation CSMAPLR ([2]) adaptation is employed to adapt a neutral emotion HMM system to another emotion using a small amount of adaptation sentences: =+, = 1.5 1.6 1.7 1.8 1.9 2 2.1 2.2 2.3 2.4 2.5 5 10 20 50 100 EXPRESSIVENESS SCORE NUMBER OF SENTENCES USED FOR ADAPTATION Anger Happiness Sadness Total Database HMM Training Vocoder Waveform AAM reconstruction Image Sequence Fusion Synthesized Audio-Visual Speech Trained HMMS Acoustic features Visual features Acoustic features Visual features Linguistic features Text to synthesize MLPG Text analysis Linguistic features Training Synthesis , : interpolated mean – covariance matrix , : adapted mean – covariance matrix of ith HMM set : interpolation weight for ith HMM set Interpolation Interpolation between observations ([3]) is employed to interpolate statistics of HMMs from different HMM sets: = σ = , = σ = 2 We trained four HMM-based audio-visual speech synthesis systems using the CVSP-EAV([4]) corpus which includes: happiness, sadness, anger, neutral. First Evaluation We adapted the neutral HMM system to the other 3 emotions using a variable number of adaptation sentences. 32 humans evaluated the expressiveness of the agent on a discrete scale of 1 to 3 (increasing). Second Evaluation We Interpolated the 6 emotion combinations for variable weight pairs: (0.9, 0.1), (0.7, 0.3), (0.5, 0.5), (0.3, 0.7), (0.1, 0.9). 28 humans were asked to recognize the emotion in each combination/pair. Emotion classification rate when interpolating the neutral and sadness HMM systems (% scores). The face of the agent is modeled by Active Appearance Models: anger mean texture #1 eigentexture mean + 3 sd eigentexture #1 mean - 3 sd eigentexture #1 This work has been funded by the BabyRobot project, supported by the EU Horizon 2020 Programme under grant 687831. REFERENCES [1 ] H. Zen et al., “The hmm-based speech synthesis system (hts) version 2.0.,” in Proc. ISCA SSW6, pp. 294–299, 2007. [2] J. Yamagishi et al., “Analysis of speaker adaptation algorithms for hmm-based speech synthesis and a constrained smaplr adaptation algorithm,”IEEE Trans. Audio, Speech, Language Processing, vol. 17, pp. 66–83, 2009. [3] T. Yoshimura et al., “Speaker interpolation for hmm-based speech synthesis system,” Acoustical Science and Technology, vol. 21, pp. 199–206, 2001. [4] P.P. Filntisis et al., “Video-realistic expressive audio-visual speech synthesis for the Greek language”, 2017, https://doi.org/10.1016/j.specom.2017.08.011 Interpolating the anger and happiness HMM sets. (respective weights shown under each image). Subjective expressiveness score for a variable number of adaptation sentences. Adapting the neutral emotion to other emotions. neutral 0.1 − 0.9 0.3 − 0.7 0.5 − 0.5 0.7 − 0.3 0.9 − 0.1

Upload: others

Post on 26-Apr-2022

8 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: PHOTOREALISTIC ADAPTATION AND INTERPOLATION OF …

Printing:This poster is 48” wide by 36” high. It’s designed to be printed on a large

Customizing the Content:The placeholders in this Type in the placeholders to add text, or click an icon to add a table, chart, SmartArt graphic, picture or multimedia file.

Tthe Bullets button on the Home tab.

If you need more placeholders for titles, body text, just make a copy of what you need and drag it into place. PowerPoint’s Smart Guides will help you align it with everything else.

Want to use your own pictures instead of ours? No problem! Just Picture. Maintain the proportion of pictures as you resize by dragging a corner.

PHOTOREALISTIC ADAPTATION AND INTERPOLATION OF FACIAL EXPRESSIONS USING HMMS AND AAMS FOR AUDIO-VISUAL SPEECH SYNTHESIS

BACKGROUND & OBJECTIVES

• Intelligent agents have a continuous presence in everyday life and speech synthesis (both acoustic & audio-visual) constitutes a vital asset for human –computer interaction

• Achieving a high degree of naturalness in HCI depends on the ability of the agent to express emotions

• However, there is a huge data overhead when considering synthesis of expressive speech in a large non-discrete emotional space

• We tackle the problem this problem by:➢using HMM adaptation to adapt an existing audio-visual speech synthesis

HMM set to a new emotion using a small amount of adaptation data➢employing HMM interpolation to combine HMM sets to generate speech

with intermediate styles

ACTIVE APPEARANCE MODELS (AAM)

Weights Neutral Anger Happiness Sadness

0.1 − 0.9 8.7 4.35 0 86.96

0.3 − 0.7 18.18 0 0 81.82

0.5 − 0.5 45.83 0 4.17 50

0.7 − 0.3 83.33 0 0 16.67

0.9 − 0.1 91.3 4.35 0 4.35

EXPERIMENTS & RESULTS

CONCLUSIONS

• We can successfully adapt an HMM-based audio-visual speech synthesis system to a target emotion using a small number of adaptation data. Level of expressiveness increases with number of adaptation sentences used.

• HMM interpolation gives us audio-visual speech with intermediate characteristics between the interpolated emotions.

• DNN version of the system can be found in [4].

Panagiotis P. Filntisis, Athanasios Katsamanis and Petros MaragosSchool of ECE, National Technical University of Athens, 15773 Athens, Greece

Athena Research and Innovation Center, 15125 Maroussi, Greece

HMM-BASED AUDIO-VISUAL SPEECH SYNTHESIS [1]

Example of the first eigentexture and the

variations it causes to the mean texture

Face shape 𝒔 = ത𝒔 + σ𝒊=𝟏𝒏 𝑝𝑖𝒔𝒊

Face texture 𝑨(𝒙) = 𝑨(𝒙) + σ𝒊=𝟏𝒍 𝜆𝑖𝑨𝒊

ത𝒔: mean shape𝑝𝑖: eigenshape coefficients𝚨(𝒙): mean texture𝜆𝑖: eigentexture coefficients

𝝁, 𝚺 : original mean and covariance matrixഥ𝝁, ഥ𝚺: adapted mean and covariance matrix𝜺, 𝚭: transformation bias and matrix

• Extract acoustic and visual features

• Train HMMs with EM algorithm• Cluster similar phonetic contexts

using decision trees

TRAINING

• Analyze input text

• Generate features from HMMs

• Reconstruct audio and video

SYNTHESIS

Adaptation

CSMAPLR ([2]) adaptation is employed to adapt a neutral emotion HMM system to another emotion using a small amount of adaptation sentences:

ഥ𝝁 = 𝒁 𝝁 + 𝜺, ഥ𝚺 = 𝜡𝚺𝜡𝚻 1.5

1.6

1.7

1.8

1.9

2

2.1

2.2

2.3

2.4

2.5

5 10 20 50 100

EXP

RES

SIV

ENES

S SC

OR

E

NUMBER OF SENTENCES USED FOR ADAPTATION

Anger Happiness Sadness Total

Database

HMM Training

Vocoder

Waveform

AAM reconstruction

Image Sequence

FusionSynthesized Audio-Visual

Speech

Trained HMMS

Acoustic features Visual

features

Acoustic features

Visualfeatures

Linguisticfeatures

Text to

synthesize

MLPG

Text analysis

Linguisticfeatures

Training

Synthesis

𝝁, 𝚺 : interpolated mean – covariance matrix𝝁𝒊, 𝚺𝐢: adapted mean – covariance matrix of ithHMM set 𝑎𝑖: interpolation weight for ith HMM set

Interpolation

Interpolation between observations ([3]) is employed to interpolate statistics of HMMs from different HMM sets:

𝝁 = σ𝒊=𝟏𝑲 𝛼𝜄𝝁𝒊 , 𝚺 = σ𝒊=𝟏

𝑲 𝛼𝑖2𝚺𝒊

We trained four HMM-based audio-visual speech synthesis systems using the CVSP-EAV([4]) corpus which includes: happiness, sadness, anger, neutral.

First Evaluation

• We adapted the neutral HMM system to the other 3 emotions using a variable number of adaptation sentences.

• 32 humans evaluated the expressiveness of the agent on a discrete scale of 1 to 3 (increasing).

Second Evaluation

• We Interpolated the 6 emotion combinations for variable weight pairs: (0.9, 0.1), (0.7, 0.3), (0.5, 0.5), (0.3, 0.7), (0.1, 0.9).

• 28 humans were asked to recognize the emotion in each combination/pair.

Emotion classification rate when interpolating the neutral and sadness HMM systems (% scores).

The face of the agent is modeled by Active Appearance Models:

anger

mean texture #1 eigentexture mean + 3 sdeigentexture #1

mean - 3 sdeigentexture #1

This work has been funded by the BabyRobot project, supported by the EU Horizon 2020 Programme under grant 687831.

REFERENCES

[1 ] H. Zen et al., “The hmm-based speech synthesis system (hts) version 2.0.,” in Proc. ISCA SSW6, pp. 294–299, 2007.

[2] J. Yamagishi et al., “Analysis of speaker adaptation algorithms for hmm-based speech synthesis and a constrained smaplradaptation algorithm,”IEEE Trans. Audio, Speech, Language Processing, vol. 17, pp. 66–83, 2009.

[3] T. Yoshimura et al., “Speaker interpolation for hmm-based speech synthesis system,” Acoustical Science and Technology, vol. 21, pp. 199–206, 2001.

[4] P.P. Filntisis et al., “Video-realistic expressive audio-visual speech synthesis for the Greek language”, 2017, https://doi.org/10.1016/j.specom.2017.08.011

Interpolating the anger and happiness HMM sets. (respective weights shown under each image).

Subjective expressiveness score for a variable number of adaptation sentences. Adapting the neutral emotion to other emotions.

neutral

0.1 − 0.9 0.3 − 0.7 0.5 − 0.5

0.7 − 0.3 0.9 − 0.1