Sharing visual features for multiclass and multiview object detection


Autoria(s): Torralba, Antonio; Murphy, Kevin P.; Freeman, William T.
Data(s)

08/10/2004

08/10/2004

14/04/2004

Resumo

We consider the problem of detecting a large number of different classes of objects in cluttered scenes. Traditional approaches require applying a battery of different classifiers to the image, at multiple locations and scales. This can be slow and can require a lot of training data, since each classifier requires the computation of many different image features. In particular, for independently trained detectors, the (run-time) computational complexity, and the (training-time) sample complexity, scales linearly with the number of classes to be detected. It seems unlikely that such an approach will scale up to allow recognition of hundreds or thousands of objects. We present a multi-class boosting procedure (joint boosting) that reduces the computational and sample complexity, by finding common features that can be shared across the classes (and/or views). The detectors for each class are trained jointly, rather than independently. For a given performance level, the total number of features required, and therefore the computational cost, is observed to scale approximately logarithmically with the number of classes. The features selected jointly are closer to edges and generic features typical of many natural structures instead of finding specific object parts. Those generic features generalize better and reduce considerably the computational cost of an algorithm for multi-class object detection.

Formato

17 p.

4223512 bytes

1537371 bytes

application/postscript

application/pdf

Identificador

AIM-2004-008

http://hdl.handle.net/1721.1/6736

Idioma(s)

en_US

Relação

AIM-2004-008

Palavras-Chave #AI #Object detection #sharing features #feature selection #multiclass #Boosting