Abstract
Hand-tracking in Extended Reality (XR) enables moving objects in near space with direct hand gestures, to pick, drag and drop objects in 3D. In this work, we investigate the use of eye-tracking to reduce the effort involved in this interaction. As the eyes naturally look ahead to the target for a drag operation, the principal idea is to map the translation of the object in the image plane to gaze, such that the hand only needs to control the depth component of the operation. We have implemented four techniques that explore two factors: the use of gaze only to move objects in X-Y vs.\ extra refinement by hand, and the use of hand input in
the Z axis to directly move objects vs.\ indirectly via a transfer function. We compared all four techniques in a user study (N=24) against baselines of direct and indirect hand input. We detail user performance, effort and experience trade-offs and show that all eye-hand techniques significantly reduce physical effort over direct gestures, pointing toward effortless drag-and-drop for XR environments.
the Z axis to directly move objects vs.\ indirectly via a transfer function. We compared all four techniques in a user study (N=24) against baselines of direct and indirect hand input. We detail user performance, effort and experience trade-offs and show that all eye-hand techniques significantly reduce physical effort over direct gestures, pointing toward effortless drag-and-drop for XR environments.
Original language | English |
---|---|
Publication date | 2 Aug 2024 |
Number of pages | 13 |
Publication status | Accepted/In press - 2 Aug 2024 |
Event | UIST '24 - The 37th Annual ACM Symposium on User Interface Software and Technology - The Westin Pittsburgh, Pittsburgh, United States Duration: 13 Oct 2024 → 16 Oct 2024 https://uist.acm.org/2024/ |
Conference
Conference | UIST '24 - The 37th Annual ACM Symposium on User Interface Software and Technology |
---|---|
Location | The Westin Pittsburgh |
Country/Territory | United States |
City | Pittsburgh |
Period | 13/10/2024 → 16/10/2024 |
Internet address |
Keywords
- eye-tracking
- gaze interaction
- drag&drop
- object manipulation
- near space