Look over there! Investigating Saliency Modulation for Visual Guidance with Augmented Reality Glasses

Jonathan Sutton*, Tobias Langlotz, Alexander Plopski, Stefanie Zollmann, Yuta Itoh, Holger Regenbrecht

*Corresponding author for this work

Research output: Contribution to conferencePaperpeer-review

Abstract

Augmented Reality has traditionally been used to display digital overlays in real environments. Many AR applications such as remote collaboration, picking tasks, or navigation require highlighting physical objects for selection or guidance. These highlights use graphical cues such as outlines and arrows. Whilst effective, they greatly contribute to visual clutter, possibly occlude scene elements, and can be problematic for long-term use. Substituting those overlays, we explore saliency modulation to accentuate objects in the real environment to guide the user’s gaze. Instead of manipulating video streams, like done in perception and cognition research, we investigate saliency modulation of the real world using optical-see-through head-mounted displays. This is a new challenge, since we do not have full control over the view of the real environment. In this work we provide our specific solution to this challenge, including built prototypes and their evaluation.
Original languageEnglish
Number of pages15
Publication statusPublished - 28 Oct 2022

Fingerprint

Dive into the research topics of 'Look over there! Investigating Saliency Modulation for Visual Guidance with Augmented Reality Glasses'. Together they form a unique fingerprint.

Cite this