During natural social gatherings, humans tend to organize themselves in the so-called free-standing conversational groups. In this context, robust head and body pose estimates can facilitate the higher-level description of the ongoing interplay. Importantly, visual information typically obtained with a distributed camera network might not sufflce to achieve the robustness sought. In this line of thought, recent advances in wearable sensing technology open the door to multimodal and richer information ows. In this paper we propose to cast the head and body pose estimation problem into a matrix completion task. We introduce a framework able to fuse multimodal data emanating from a combination of distributed and wearable sensors, taking into account the temporal consistency, the head/body coupling and the noise inherent to the scenario. We report results on the novel and challenging SALSA dataset, containing visual, auditory and infrared recordings of 18 people interacting in a regular in...

Analyzing Free-standing Conversational Groups: A Multimodal Approach / Alameda Pineda, Xavier; Yan, Yan; Ricci, Elisa; Lanz, Oswald; Sebe, Niculae. - (2015), pp. 5-14. ( 23rd ACM International Conference on Multimedia, MM 2015 Brisbane, Australia 24-31 october) [10.1145/2733373.2806238].

Analyzing Free-standing Conversational Groups: A Multimodal Approach

Alameda Pineda, Xavier;Yan, Yan;Ricci, Elisa;Sebe, Niculae
2015-01-01

Abstract

During natural social gatherings, humans tend to organize themselves in the so-called free-standing conversational groups. In this context, robust head and body pose estimates can facilitate the higher-level description of the ongoing interplay. Importantly, visual information typically obtained with a distributed camera network might not sufflce to achieve the robustness sought. In this line of thought, recent advances in wearable sensing technology open the door to multimodal and richer information ows. In this paper we propose to cast the head and body pose estimation problem into a matrix completion task. We introduce a framework able to fuse multimodal data emanating from a combination of distributed and wearable sensors, taking into account the temporal consistency, the head/body coupling and the noise inherent to the scenario. We report results on the novel and challenging SALSA dataset, containing visual, auditory and infrared recordings of 18 people interacting in a regular in...
2015
Proceedings of the 23rd ACM international conference on Multimedia
New York
ACM Press
978-1-4503-3459-4
Alameda Pineda, Xavier; Yan, Yan; Ricci, Elisa; Lanz, Oswald; Sebe, Niculae
Analyzing Free-standing Conversational Groups: A Multimodal Approach / Alameda Pineda, Xavier; Yan, Yan; Ricci, Elisa; Lanz, Oswald; Sebe, Niculae. - (2015), pp. 5-14. ( 23rd ACM International Conference on Multimedia, MM 2015 Brisbane, Australia 24-31 october) [10.1145/2733373.2806238].
File in questo prodotto:
File Dimensione Formato  
p5-alameda-pineda.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.96 MB
Formato Adobe PDF
1.96 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/125102
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 68
  • ???jsp.display-item.citation.isi??? 49
  • OpenAlex 70
social impact