Gaze-Based Controlling a Vehicle

Diako Mardanbeigi, Dan Witzner Hansen

Research output: Contribution to conference - NOT published in proceeding or journalPaperResearchpeer-review


Research and applications of gaze interaction has mainly been conducted on a 2 dimensional surface (usually screens) for controlling a computer or controlling the movements of a robot. Emerging wearable and mobile technologies, such as google glasses may shift how gaze is used as an interactive modality if gaze trackers are embedded into the head- mounted devices. The domain of gaze-based interactive applications increases dramatically as interaction is no longer constrained to 2D displays. This paper proposes a general framework for gaze-based controlling a non- stationary robot (vehicle) as an example of a complex gaze-based task in environment. This paper discusses the possibilities and limitations of how gaze interaction can be performed for controlling vehicles not only using a remote gaze tracker but also in general challenging situations where the user and robot are mobile and the movements may be governed by several degrees of freedom (e.g. flying). A case study is also introduced where the mobile gaze tracker is used for controlling a Roomba vacuum cleaner.
Original languageEnglish
Publication date2013
Number of pages6
Publication statusPublished - 2013
EventCHI 2013: Gaze Interaction in the Post-WIMP World - Paris, France
Duration: 27 Apr 201327 Apr 2013


WorkshopCHI 2013
Internet address


  • Gaze-based interaction
  • robot
  • vehicle
  • craft
  • head- gestures
  • eye tracking
  • driving


Dive into the research topics of 'Gaze-Based Controlling a Vehicle'. Together they form a unique fingerprint.

Cite this