Generalizable Dynamic Radiance Fields For Talking Head Synthesis With Few-shot

No Thumbnail Available
Date
2023
Journal Title
Journal ISSN
Volume Title
Publisher
The Eurographics Association
Abstract
Audio-driven talking head generation has wide applications in virtual games, hosts, online meetings, etc. Recently, great achievements have been made in synthesizing talking heads based on neural radiance fields. However, the existing few-shot talking head synthesis methods still suffer from inaccurate deformation and lack of visual consistency. Therefore, we propose a Generalizable Dynamic Radiance Field (GDRF), which can rapidly generalize to unseen identities with few-shot. We introduce a warping module with 3D constraints to act in feature volume space, which is identity adaptive and exhibits excellent shape-shifting abilities. Our method can generate more accurately deformed and view consistent target images compared to previous methods. Furthermore, we map the audio signal to 3DMM parameters by applying an LSTM network, which helps get long-term context and generate more continuous and natural video. Extensive experiments demonstrate the superiority of our proposed method.
Description

CCS Concepts: Computing methodologies -> Reconstruction; Animation; Shape representations

        
@inproceedings{
10.2312:pg.20231274
, booktitle = {
Pacific Graphics Short Papers and Posters
}, editor = {
Chaine, Raphaëlle
 and
Deng, Zhigang
 and
Kim, Min H.
}, title = {{
Generalizable Dynamic Radiance Fields For Talking Head Synthesis With Few-shot
}}, author = {
Dang, Rujing
 and
Wang, Shaohui
 and
Wang, Haoqian
}, year = {
2023
}, publisher = {
The Eurographics Association
}, ISBN = {
978-3-03868-234-9
}, DOI = {
10.2312/pg.20231274
} }
Citation