/Semi-Supervised Self-Training of Object Detection Models

Semi-Supervised Self-Training of Object Detection Models

Charles Rosenberg, Martial Hebert and Henry Schneiderman
Conference Paper, Seventh IEEE Workshop on Applications of Computer Vision, January, 2005

Download Publication (PDF)

Copyright notice: This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author’s copyright. These works may not be reposted without the explicit permission of the copyright holder.


The construction of appearance-based object detection systems is time-consuming and difficult because a large number of training examples must be collected and manually labeled in order to capture variations in object appearance. Semi-supervised training is a means for reducing the effort needed to prepare the training set by training the model with a small number of fully labeled example and an additional set of unlabeled or weakly labeled examples. In this work we present a semi-supervised approach to training object detection systems based on self-training. We implement our approach as a wrapper around the training process of an existing detector and present empirical results. The key contributions of this empirical study is to demonstrate that a model trained in this manner can achieve results comparable to a model trained in the traditional manner using a much larger set of fully labeled data, and that a training data selection metric that is defined independently of the detector greatly outperforms a selection metric based on the confidence generated by the detector.

BibTeX Reference
author = {Charles Rosenberg and Martial Hebert and Henry Schneiderman},
title = {Semi-Supervised Self-Training of Object Detection Models},
booktitle = {Seventh IEEE Workshop on Applications of Computer Vision},
year = {2005},
month = {January},