We address the problem of automatic extraction of foreground objects from videos. The goal is to provide a method for unsupervised collection of samples which can be further used for object detection training without any human intervention. We use the well known Selective Search approach to produce an initial still-image based segmentation of the video frames. This initial set of proposals is pruned and temporally extended using optical flow and transductive learning. Specifically, we propose to use Dense Trajectories in order to robustly match and track candidate boxes over different frames. The obtained box tracks are used to collect samples for unsupervised training of track-specific detectors. Finally, the detectors are run on the videos to extract the final tubes. The combination of appearance-based static ”objectness” (Selective Search), motion information (Dense Trajectories) and transductive learning (detectors are forced to ”overfit” on the unsupervised data used for training) makes the proposed approach extremely robust. We outperform state-of-the-art systems by a large margin on common benchmarks used for tube proposal evaluation.

Unsupervised Tube Extraction Using Transductive Learning and Dense Trajectories / Puscas, Mihai - Marian; Sangineto, Enver; Culibrk, Dubravko; Sebe, Niculae. - ELETTRONICO. - (2015), pp. 1653-1661. (Intervento presentato al convegno ICCV 2015 tenutosi a Santiago, Chile nel 7-13 December 2015) [10.1109/ICCV.2015.193].

Unsupervised Tube Extraction Using Transductive Learning and Dense Trajectories

Puscas, Mihai - Marian;Sangineto, Enver;Culibrk, Dubravko;Sebe, Niculae
2015-01-01

Abstract

We address the problem of automatic extraction of foreground objects from videos. The goal is to provide a method for unsupervised collection of samples which can be further used for object detection training without any human intervention. We use the well known Selective Search approach to produce an initial still-image based segmentation of the video frames. This initial set of proposals is pruned and temporally extended using optical flow and transductive learning. Specifically, we propose to use Dense Trajectories in order to robustly match and track candidate boxes over different frames. The obtained box tracks are used to collect samples for unsupervised training of track-specific detectors. Finally, the detectors are run on the videos to extract the final tubes. The combination of appearance-based static ”objectness” (Selective Search), motion information (Dense Trajectories) and transductive learning (detectors are forced to ”overfit” on the unsupervised data used for training) makes the proposed approach extremely robust. We outperform state-of-the-art systems by a large margin on common benchmarks used for tube proposal evaluation.
2015
The IEEE International Conference on Computer Vision (ICCV) 2015
United States
IEEE
Puscas, Mihai - Marian; Sangineto, Enver; Culibrk, Dubravko; Sebe, Niculae
Unsupervised Tube Extraction Using Transductive Learning and Dense Trajectories / Puscas, Mihai - Marian; Sangineto, Enver; Culibrk, Dubravko; Sebe, Niculae. - ELETTRONICO. - (2015), pp. 1653-1661. (Intervento presentato al convegno ICCV 2015 tenutosi a Santiago, Chile nel 7-13 December 2015) [10.1109/ICCV.2015.193].
File in questo prodotto:
File Dimensione Formato  
Puscas_Unsupervised_Tube_Extraction_ICCV_2015_paper.pdf

accesso aperto

Descrizione: Published paper
Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.58 MB
Formato Adobe PDF
1.58 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/170044
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 20
  • ???jsp.display-item.citation.isi??? 13
social impact