Repository logo
 

Vision-based hand pose estimation methods for Augmented Reality in industry: Crowdsourced evaluation on HoloLens 2

Published version
Peer-reviewed

Repository DOI


Loading...
Thumbnail Image

Change log

Abstract

Gestural input based on hand pose estimation is a common interaction method for augmented reality (AR). This interaction technique has gained more popularity with the emergence of novel AR-supporting devices such as Microsoft HoloLens 2 (HL2) and advancements in computer vision research underpinning hand-tracking and gesture recognition methods. In our work, we focus on challenging cases where the AR interface is facilitated with a state-of-the-art HL2 headset for unconstrained execution of tasks requiring simultaneous hand movement and tracking. When using this headset, AR users might bimanually interact with digital and physical objects that are visible in the user's field of view (FoV) through the see-through visor. Due to the limiting in-built capabilities, we investigated a range of hand pose estimation functionalities from different domains. To ensure a fair comparison, we asked several participants to carry out tasks requiring interactions with real-world objects and record the performance of various hand-tracking solutions. Next, we evaluated the performance of these algorithms through crowdsourcing, often used to provide ground truth for machine learning training. Our results provide a guideline for AR developers in selecting appropriate hand-tracking solutions for a given deployment context.

Description

Journal Title

Computers in Industry

Conference Name

Journal ISSN

0166-3615
1872-6194

Volume Title

171

Publisher

Elsevier

Rights and licensing

Except where otherwised noted, this item's license is described as Attribution 4.0 International
Sponsorship
UK Research and Innovation (EP/V062123/1)