1 resultado para joint range of motion
em Abertay Research Collections - Abertay University’s repository
Filtro por publicador
- ABACUS. Repositorio de Producción Científica - Universidad Europea (1)
- Aberdeen University (1)
- Abertay Research Collections - Abertay University’s repository (1)
- Acceda, el repositorio institucional de la Universidad de Las Palmas de Gran Canaria. España (2)
- AMS Tesi di Dottorato - Alm@DL - Università di Bologna (10)
- AMS Tesi di Laurea - Alm@DL - Università di Bologna (2)
- Aquatic Commons (2)
- ArchiMeD - Elektronische Publikationen der Universität Mainz - Alemanha (1)
- Archive of European Integration (21)
- Aston University Research Archive (10)
- B-Digital - Universidade Fernando Pessoa - Portugal (3)
- Biblioteca Digital da Produção Intelectual da Universidade de São Paulo (18)
- Biblioteca Digital da Produção Intelectual da Universidade de São Paulo (BDPI/USP) (2)
- Biblioteca Digital de Teses e Dissertações Eletrônicas da UERJ (1)
- Biodiversity Heritage Library, United States (2)
- BORIS: Bern Open Repository and Information System - Berna - Suiça (92)
- Boston University Digital Common (7)
- Brock University, Canada (3)
- Bucknell University Digital Commons - Pensilvania - USA (2)
- Bulgarian Digital Mathematics Library at IMI-BAS (2)
- CaltechTHESIS (4)
- Cambridge University Engineering Department Publications Database (28)
- CentAUR: Central Archive University of Reading - UK (55)
- Chinese Academy of Sciences Institutional Repositories Grid Portal (20)
- Cochin University of Science & Technology (CUSAT), India (1)
- Coffee Science - Universidade Federal de Lavras (2)
- Collection Of Biostatistics Research Archive (1)
- Comissão Econômica para a América Latina e o Caribe (CEPAL) (9)
- CORA - Cork Open Research Archive - University College Cork - Ireland (1)
- DI-fusion - The institutional repository of Université Libre de Bruxelles (1)
- Digital Commons - Michigan Tech (2)
- Digital Commons @ DU | University of Denver Research (2)
- Digital Commons at Florida International University (4)
- Digital Peer Publishing (2)
- DigitalCommons@The Texas Medical Center (5)
- DigitalCommons@University of Nebraska - Lincoln (1)
- DRUM (Digital Repository at the University of Maryland) (4)
- Duke University (4)
- eResearch Archive - Queensland Department of Agriculture; Fisheries and Forestry (16)
- Glasgow Theses Service (1)
- Greenwich Academic Literature Archive - UK (2)
- Helda - Digital Repository of University of Helsinki (7)
- Hospitais da Universidade de Coimbra (1)
- Indian Institute of Science - Bangalore - Índia (41)
- Institutional Repository of Leibniz University Hannover (2)
- Instituto Politécnico de Bragança (1)
- Instituto Politécnico do Porto, Portugal (11)
- Iowa Publications Online (IPO) - State Library, State of Iowa (Iowa), United States (1)
- Massachusetts Institute of Technology (3)
- Memoria Académica - FaHCE, UNLP - Argentina (2)
- Memorial University Research Repository (1)
- Ministerio de Cultura, Spain (1)
- National Center for Biotechnology Information - NCBI (7)
- Plymouth Marine Science Electronic Archive (PlyMSEA) (7)
- Portal de Revistas Científicas Complutenses - Espanha (3)
- Publishing Network for Geoscientific & Environmental Data (28)
- QSpace: Queen's University - Canada (3)
- QUB Research Portal - Research Directory and Institutional Repository for Queen's University Belfast (58)
- Queensland University of Technology - ePrints Archive (112)
- Repositório Científico do Instituto Politécnico de Lisboa - Portugal (1)
- Repositório Institucional da Universidade de Aveiro - Portugal (1)
- Repositório Institucional da Universidade Federal de São Paulo - UNIFESP (1)
- Repositório Institucional dos Hospitais da Universidade Coimbra (1)
- Repositório Institucional UNESP - Universidade Estadual Paulista "Julio de Mesquita Filho" (84)
- Repositorio Institucional Universidad de Medellín (1)
- Universidad de Alicante (1)
- Universidad del Rosario, Colombia (2)
- Universidad Politécnica de Madrid (13)
- Universidade Complutense de Madrid (1)
- Universidade de Lisboa - Repositório Aberto (2)
- Universidade Federal do Rio Grande do Norte (UFRN) (4)
- Universidade Técnica de Lisboa (1)
- Université de Lausanne, Switzerland (4)
- Université de Montréal (3)
- Université de Montréal, Canada (11)
- University of Canberra Research Repository - Australia (2)
- University of Michigan (77)
- University of Queensland eSpace - Australia (31)
- University of Southampton, United Kingdom (1)
- University of Washington (2)
- WestminsterResearch - UK (3)
Resumo:
Fully articulated hand tracking promises to enable fundamentally new interactions with virtual and augmented worlds, but the limited accuracy and efficiency of current systems has prevented widespread adoption. Today's dominant paradigm uses machine learning for initialization and recovery followed by iterative model-fitting optimization to achieve a detailed pose fit. We follow this paradigm, but make several changes to the model-fitting, namely using: (1) a more discriminative objective function; (2) a smooth-surface model that provides gradients for non-linear optimization; and (3) joint optimization over both the model pose and the correspondences between observed data points and the model surface. While each of these changes may actually increase the cost per fitting iteration, we find a compensating decrease in the number of iterations. Further, the wide basin of convergence means that fewer starting points are needed for successful model fitting. Our system runs in real-time on CPU only, which frees up the commonly over-burdened GPU for experience designers. The hand tracker is efficient enough to run on low-power devices such as tablets. We can track up to several meters from the camera to provide a large working volume for interaction, even using the noisy data from current-generation depth cameras. Quantitative assessments on standard datasets show that the new approach exceeds the state of the art in accuracy. Qualitative results take the form of live recordings of a range of interactive experiences enabled by this new approach.