Gaze-Based Controlling a Vehicle

Diako Mardanbeigi, Dan Witzner Hansen

    Research output: Contribution to conference - NOT published in proceeding or journalPaperResearchpeer-review

    Abstract

    Research and applications of gaze interaction has mainly been conducted on a 2 dimensional surface (usually screens) for controlling a computer or controlling the movements of a robot. Emerging wearable and mobile technologies, such as google glasses may shift how gaze is used as an interactive modality if gaze trackers are embedded into the head- mounted devices. The domain of gaze-based interactive applications increases dramatically as interaction is no longer constrained to 2D displays. This paper proposes a general framework for gaze-based controlling a non- stationary robot (vehicle) as an example of a complex gaze-based task in environment. This paper discusses the possibilities and limitations of how gaze interaction can be performed for controlling vehicles not only using a remote gaze tracker but also in general challenging situations where the user and robot are mobile and the movements may be governed by several degrees of freedom (e.g. flying). A case study is also introduced where the mobile gaze tracker is used for controlling a Roomba vacuum cleaner.
    Original languageEnglish
    Publication date2013
    Number of pages6
    Publication statusPublished - 2013
    EventCHI 2013: Gaze Interaction in the Post-WIMP World - Paris, France
    Duration: 27 Apr 201327 Apr 2013
    http://gaze-interaction.net/

    Workshop

    WorkshopCHI 2013
    Country/TerritoryFrance
    CityParis
    Period27/04/201327/04/2013
    Internet address

    Keywords

    • Gaze-based interaction
    • robot
    • vehicle
    • craft
    • head- gestures
    • eye tracking
    • driving

    Fingerprint

    Dive into the research topics of 'Gaze-Based Controlling a Vehicle'. Together they form a unique fingerprint.

    Cite this