This paper introduces a novel deep learning framework for image animation. Given an input image with a target object and a driving video sequence depicting a moving object, our framework generates a video in which the target object is animated according to the driving sequence. This is achieved through a deep architecture that decouples appearance and motion information. Our framework consists of three main modules: (i) a Keypoint Detector unsupervisely trained to extract object keypoints, (ii) a Dense Motion prediction network for generating dense heatmaps from sparse keypoints, in order to better encode motion information and (iii) a Motion Transfer Network, which uses the motion heatmaps and appearance information extracted from the input image to synthesize the output frames. We demonstrate the effectiveness of our method on several benchmark datasets, spanning a wide variety of object appearances, and show that our approach outperforms state-of-the-art image animation and video generation methods.

Animating Arbitrary Objects via Deep Motion Transfer / Siarohin, Aliaksandr; Lathuiliere, Stephane; Tulyakov, Sergey; Ricci, Elisa; Sebe, Nicu. - (2019), pp. 2372-2381. (Intervento presentato al convegno IEEE Comference on Computer Vision and Pattern Recognition (CVPR'19) tenutosi a Long Beach nel June 16-20, 2019) [10.1109/CVPR.2019.00248].

Animating Arbitrary Objects via Deep Motion Transfer

Siarohin, Aliaksandr;Lathuiliere, Stephane;Tulyakov, Sergey;Ricci, Elisa;Sebe, Nicu
2019-01-01

Abstract

This paper introduces a novel deep learning framework for image animation. Given an input image with a target object and a driving video sequence depicting a moving object, our framework generates a video in which the target object is animated according to the driving sequence. This is achieved through a deep architecture that decouples appearance and motion information. Our framework consists of three main modules: (i) a Keypoint Detector unsupervisely trained to extract object keypoints, (ii) a Dense Motion prediction network for generating dense heatmaps from sparse keypoints, in order to better encode motion information and (iii) a Motion Transfer Network, which uses the motion heatmaps and appearance information extracted from the input image to synthesize the output frames. We demonstrate the effectiveness of our method on several benchmark datasets, spanning a wide variety of object appearances, and show that our approach outperforms state-of-the-art image animation and video generation methods.
2019
IEEE Comference on Computer Vision and Pattern Recognition (CVPR'19)
New York
IEEE
978-1-7281-3293-8
Siarohin, Aliaksandr; Lathuiliere, Stephane; Tulyakov, Sergey; Ricci, Elisa; Sebe, Nicu
Animating Arbitrary Objects via Deep Motion Transfer / Siarohin, Aliaksandr; Lathuiliere, Stephane; Tulyakov, Sergey; Ricci, Elisa; Sebe, Nicu. - (2019), pp. 2372-2381. (Intervento presentato al convegno IEEE Comference on Computer Vision and Pattern Recognition (CVPR'19) tenutosi a Long Beach nel June 16-20, 2019) [10.1109/CVPR.2019.00248].
File in questo prodotto:
File Dimensione Formato  
Siarohin_Animating_Arbitrary_Objects_via_Deep_Motion_Transfer_CVPR_2019_paper.pdf

accesso aperto

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Altra licenza (Other type of license)
Dimensione 1.55 MB
Formato Adobe PDF
1.55 MB Adobe PDF Visualizza/Apri
08954170.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 475.06 kB
Formato Adobe PDF
475.06 kB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/250769
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 199
  • ???jsp.display-item.citation.isi??? 140
social impact