An adaptation framework for head-pose classification in dynamic multi-view scenarios


Autoria(s): Rajagopal, Anoop K; Subramanian, Ramanathan; Vieriu, Radu L; Ricci, Elisa; Lanz, Oswald; Ramakrishnan, Kalpathi; Sebe, Nicu
Data(s)

2012

Resumo

Multi-view head-pose estimation in low-resolution, dynamic scenes is difficult due to blurred facial appearance and perspective changes as targets move around freely in the environment. Under these conditions, acquiring sufficient training examples to learn the dynamic relationship between position, face appearance and head-pose can be very expensive. Instead, a transfer learning approach is proposed in this work. Upon learning a weighted-distance function from many examples where the target position is fixed, we adapt these weights to the scenario where target positions are varying. The adaptation framework incorporates reliability of the different face regions for pose estimation under positional variation, by transforming the target appearance to a canonical appearance corresponding to a reference scene location. Experimental results confirm effectiveness of the proposed approach, which outperforms state-of-the-art by 9.5% under relevant conditions. To aid further research on this topic, we also make DPOSE- a dynamic, multi-view head-pose dataset with ground-truth publicly available with this paper.

Formato

application/pdf

Identificador

http://eprints.iisc.ernet.in/46554/1/PT2-9.pdf

Rajagopal, Anoop K and Subramanian, Ramanathan and Vieriu, Radu L and Ricci, Elisa and Lanz, Oswald and Ramakrishnan, Kalpathi and Sebe, Nicu (2012) An adaptation framework for head-pose classification in dynamic multi-view scenarios. In: 11th Asian Conference on Computer Vision, November 5-9, 2012, Daejeon, Korea.

Publicador

IEEE

Relação

http://dx.doi.org/10.1007/978-3-642-37444-9_51

http://eprints.iisc.ernet.in/46554/

Palavras-Chave #Electrical Engineering
Tipo

Conference Paper

PeerReviewed