We present a novel task, i.e, animating a target 3D object through the motion of a raw driving sequence. In previous works, extra auxiliary correlations between source and target meshes or intermedia factors are inevitable to capture the motions in the driving sequences. Instead, we introduce AniFormer, a novel Transformer-based architecture, that generates animated 3D sequences by directly taking the raw driving sequences and arbitrary same-type target meshes as inputs. Specifically, we customize the Transformer architecture for 3D animation that generates mesh sequences by integrating styles from target meshes and motions from the driving meshes. Besides, instead of the conventional single regression head in the vanilla Transformer, AniFormer generates multiple frames as outputs to preserve the sequential consistency of the generated meshes. To achieve this, we carefully design a pair of regression constraints, i.e., motion and appearance constraints, that can provide strong regularization on the generated mesh sequences. Our AniFormer achieves high-fidelity, realistic, temporally coherent animated results and outperforms compared start-of-the-art methods on benchmarks of diverse categories. Code is available: https://github.com/mikecheninoulu/AniFormer.

AniFormer: Data-driven 3D Animation with Transformer / Chen, Haoyu; Tang, Hao; Sebe, Nicu; Zhao, Guoying. - (2021), pp. 1-13. (Intervento presentato al convegno 32nd British Machine Vision Conference, BMVC 2021 tenutosi a online nel 22nd-25th November 2021).

AniFormer: Data-driven 3D Animation with Transformer

Tang, Hao;Sebe, Nicu;
2021-01-01

Abstract

We present a novel task, i.e, animating a target 3D object through the motion of a raw driving sequence. In previous works, extra auxiliary correlations between source and target meshes or intermedia factors are inevitable to capture the motions in the driving sequences. Instead, we introduce AniFormer, a novel Transformer-based architecture, that generates animated 3D sequences by directly taking the raw driving sequences and arbitrary same-type target meshes as inputs. Specifically, we customize the Transformer architecture for 3D animation that generates mesh sequences by integrating styles from target meshes and motions from the driving meshes. Besides, instead of the conventional single regression head in the vanilla Transformer, AniFormer generates multiple frames as outputs to preserve the sequential consistency of the generated meshes. To achieve this, we carefully design a pair of regression constraints, i.e., motion and appearance constraints, that can provide strong regularization on the generated mesh sequences. Our AniFormer achieves high-fidelity, realistic, temporally coherent animated results and outperforms compared start-of-the-art methods on benchmarks of diverse categories. Code is available: https://github.com/mikecheninoulu/AniFormer.
2021
British Machine Vision Conference (BMVC’21)
Durham, UK
British Machine Vision Association, BMVA
Chen, Haoyu; Tang, Hao; Sebe, Nicu; Zhao, Guoying
AniFormer: Data-driven 3D Animation with Transformer / Chen, Haoyu; Tang, Hao; Sebe, Nicu; Zhao, Guoying. - (2021), pp. 1-13. (Intervento presentato al convegno 32nd British Machine Vision Conference, BMVC 2021 tenutosi a online nel 22nd-25th November 2021).
File in questo prodotto:
File Dimensione Formato  
HaoyuBMVC.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 6.32 MB
Formato Adobe PDF
6.32 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/326200
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 10
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact