Recently, diffusion models have made significant strides in synthesizing realistic 2D human images based on provided text prompts. Building upon this, researchers have extended 2D text-to-image diffusion models into the 3D domain for generating human textures (UV Maps). However, some important problems about UV Map Generative models are still not solved, i.e., how to generate personalized texture maps for any given face image, and how to define and evaluate the quality of these generated texture maps. To solve the above problems, we introduce a novel method, UVMap-ID, which is a controllable and personalized UV Map generative model. Unlike traditional large-scale training methods in 2D, we propose to fine-tune a pre-trained text-to-image diffusion model which is integrated with a face fusion module for achieving ID-driven customized generation. To support the finetuning strategy, we introduce a small-scale attribute-balanced training dataset, including high-quality textures with labeled text and Face ID. Additionally, we introduce some metrics to evaluate the multiple aspects of the textures. Finally, both quantitative and qualitative analyses demonstrate the effectiveness of our method in controllable and personalized UV Map generation.

UVMap-ID: A Controllable and Personalized UV Map Generative Model / Wang, W.; Zhang, J.; Liu, C.; Li, X.; Xu, X.; Shi, H.; Sebe, N.; Lepri, B.. - (2024), pp. 10725-10734. (Intervento presentato al convegno 32nd ACM International Conference on Multimedia, MM 2024 tenutosi a aus nel 2024) [10.1145/3664647.3680861].

UVMap-ID: A Controllable and Personalized UV Map Generative Model

Wang W.;Zhang J.;Liu C.;Sebe N.;Lepri B.
2024-01-01

Abstract

Recently, diffusion models have made significant strides in synthesizing realistic 2D human images based on provided text prompts. Building upon this, researchers have extended 2D text-to-image diffusion models into the 3D domain for generating human textures (UV Maps). However, some important problems about UV Map Generative models are still not solved, i.e., how to generate personalized texture maps for any given face image, and how to define and evaluate the quality of these generated texture maps. To solve the above problems, we introduce a novel method, UVMap-ID, which is a controllable and personalized UV Map generative model. Unlike traditional large-scale training methods in 2D, we propose to fine-tune a pre-trained text-to-image diffusion model which is integrated with a face fusion module for achieving ID-driven customized generation. To support the finetuning strategy, we introduce a small-scale attribute-balanced training dataset, including high-quality textures with labeled text and Face ID. Additionally, we introduce some metrics to evaluate the multiple aspects of the textures. Finally, both quantitative and qualitative analyses demonstrate the effectiveness of our method in controllable and personalized UV Map generation.
2024
MM 2024 - Proceedings of the 32nd ACM International Conference on Multimedia
New York
Association for Computing Machinery, Inc
9798400706868
Wang, W.; Zhang, J.; Liu, C.; Li, X.; Xu, X.; Shi, H.; Sebe, N.; Lepri, B.
UVMap-ID: A Controllable and Personalized UV Map Generative Model / Wang, W.; Zhang, J.; Liu, C.; Li, X.; Xu, X.; Shi, H.; Sebe, N.; Lepri, B.. - (2024), pp. 10725-10734. (Intervento presentato al convegno 32nd ACM International Conference on Multimedia, MM 2024 tenutosi a aus nel 2024) [10.1145/3664647.3680861].
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/439532
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact