ExpCLIP: Bridging Text and Facial Expressions via Semantic Alignment

08/28/2023
by   Yicheng Zhong, et al.
0

The objective of stylized speech-driven facial animation is to create animations that encapsulate specific emotional expressions. Existing methods often depend on pre-established emotional labels or facial expression templates, which may limit the necessary flexibility for accurately conveying user intent. In this research, we introduce a technique that enables the control of arbitrary styles by leveraging natural language as emotion prompts. This technique presents benefits in terms of both flexibility and user-friendliness. To realize this objective, we initially construct a Text-Expression Alignment Dataset (TEAD), wherein each facial expression is paired with several prompt-like descriptions.We propose an innovative automatic annotation method, supported by Large Language Models (LLMs), to expedite the dataset construction, thereby eliminating the substantial expense of manual annotation. Following this, we utilize TEAD to train a CLIP-based model, termed ExpCLIP, which encodes text and facial expressions into semantically aligned style embeddings. The embeddings are subsequently integrated into the facial animation generator to yield expressive and controllable facial animations. Given the limited diversity of facial emotions in existing speech-driven facial animation training data, we further introduce an effective Expression Prompt Augmentation (EPA) mechanism to enable the animation generator to support unprecedented richness in style control. Comprehensive experiments illustrate that our method accomplishes expressive facial animation generation and offers enhanced flexibility in effectively conveying the desired style.

READ FULL TEXT

page 1

page 3

page 5

page 7

research
01/05/2023

Expressive Speech-driven Facial Animation with controllable emotions

It is in high demand to generate facial animation with high realism, but...
research
12/30/2022

Imitator: Personalized Speech-driven 3D Facial Animation

Speech-driven 3D facial animation has been widely explored, with applica...
research
12/02/2021

Altering Facial Expression Based on Textual Emotion

Faces and their expressions are one of the potent subjects for digital i...
research
12/12/2019

Unconstrained Facial Expression Transfer using Style-based Generator

Facial expression transfer and reenactment has been an important researc...
research
06/28/2022

Generating near-infrared facial expression datasets with dimensional affect labels

Facial expression analysis has long been an active research area of comp...
research
02/24/2022

AFFDEX 2.0: A Real-Time Facial Expression Analysis Toolkit

In this paper we introduce AFFDEX 2.0 - a toolkit for analyzing facial e...
research
07/23/2021

Label Distribution Amendment with Emotional Semantic Correlations for Facial Expression Recognition

By utilizing label distribution learning, a probability distribution is ...

Please sign up or login with your details

Forgot password? Click here to reset