Skip to main content

Real-Time Audio-Guided Multi-Face Reenactment

Jiangning Zhang, Xianfang Zeng, Chao Xu, Yong Liu

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
    Length: 00:10:54
09 May 2022

Audio-guided face reenactment aims to generate authentic target faces that have matched facial expression of the input audio, and many learning-based methods have successfully achieved this. However, most methods can only reenact a particular person once trained or suffer from the low-quality generation of the target images. Also, nearly none of the current reenactment works consider the model size and running speed that are important for practical use. To solve the above challenges, we propose an efficient \emph{A}udio-guided \emph{M}ulti-face reenactment model named \emph{AMNet}, which can reenact target faces among multiple persons with corresponding source faces and drive signals as inputs. Concretely, we design a \emph{Geometric Controller} (GC) module to inject the drive signals so that the model can be optimized in an end-to-end manner and generate more authentic images. Also, we adopt a lightweight network for our face reenactor so that the model can run in real-time on both CPU and GPU devices. Abundant experiments prove our approach's superiority over existing methods, \eg, averagely decreasing FID by 0.12$\downarrow$ and increasing SSIM by 0.031$\uparrow$ than APB2Face, while owning fewer parameters ($\times 4 \downarrow$) and faster CPU speed ($\times 4 \uparrow$).

Tags:

More Like This

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00