Measuring and modeling the perception of natural and unconstrained gaze in humans and machines

Humans are remarkably adept at interpreting the gaze direction of other individuals in their surroundings. This skill is at the core of the ability to engage in joint visual attention, which is essential for establishing social interactions. How accurate are humans in determining the gaze direction...

Full description

Bibliographic Details
Main Authors: Harari, Daniel, Gao, Tao, Kanwisher, Nancy, Tenenbaum, Joshua, Ullman, Shimon
Format: Technical Report
Language:en_US
Published: Center for Brains, Minds and Machines (CBMM), arXiv 2016
Subjects:
Online Access:http://hdl.handle.net/1721.1/105477
_version_ 1811083555939287040
author Harari, Daniel
Gao, Tao
Kanwisher, Nancy
Tenenbaum, Joshua
Ullman, Shimon
author_facet Harari, Daniel
Gao, Tao
Kanwisher, Nancy
Tenenbaum, Joshua
Ullman, Shimon
author_sort Harari, Daniel
collection MIT
description Humans are remarkably adept at interpreting the gaze direction of other individuals in their surroundings. This skill is at the core of the ability to engage in joint visual attention, which is essential for establishing social interactions. How accurate are humans in determining the gaze direction of others in lifelike scenes, when they can move their heads and eyes freely, and what are the sources of information for the underlying perceptual processes? These questions pose a challenge from both empirical and computational perspectives, due to the complexity of the visual input in real-life situations. Here we measure empirically human accuracy in perceiving the gaze direction of others in lifelike scenes, and study computationally the sources of information and representations underlying this cognitive capacity. We show that humans perform better in face-to-face conditions compared with recorded conditions, and that this advantage is not due to the availability of input dynamics. We further show that humans are still performing well when only the eyes-region is visible, rather than the whole face. We develop a computational model, which replicates the pattern of human performance, including the finding that the eyes-region contains on its own, the required information for estimating both head orientation and direction of gaze. Consistent with neurophysiological findings on task-specific face regions in the brain, the learned computational representations reproduce perceptual effects such as the Wollaston illusion, when trained to estimate direction of gaze, but not when trained to recognize objects or faces.
first_indexed 2024-09-23T12:34:56Z
format Technical Report
id mit-1721.1/105477
institution Massachusetts Institute of Technology
language en_US
last_indexed 2024-09-23T12:34:56Z
publishDate 2016
publisher Center for Brains, Minds and Machines (CBMM), arXiv
record_format dspace
spelling mit-1721.1/1054772019-04-11T09:51:54Z Measuring and modeling the perception of natural and unconstrained gaze in humans and machines Harari, Daniel Gao, Tao Kanwisher, Nancy Tenenbaum, Joshua Ullman, Shimon gaze direction social interaction human vision Humans are remarkably adept at interpreting the gaze direction of other individuals in their surroundings. This skill is at the core of the ability to engage in joint visual attention, which is essential for establishing social interactions. How accurate are humans in determining the gaze direction of others in lifelike scenes, when they can move their heads and eyes freely, and what are the sources of information for the underlying perceptual processes? These questions pose a challenge from both empirical and computational perspectives, due to the complexity of the visual input in real-life situations. Here we measure empirically human accuracy in perceiving the gaze direction of others in lifelike scenes, and study computationally the sources of information and representations underlying this cognitive capacity. We show that humans perform better in face-to-face conditions compared with recorded conditions, and that this advantage is not due to the availability of input dynamics. We further show that humans are still performing well when only the eyes-region is visible, rather than the whole face. We develop a computational model, which replicates the pattern of human performance, including the finding that the eyes-region contains on its own, the required information for estimating both head orientation and direction of gaze. Consistent with neurophysiological findings on task-specific face regions in the brain, the learned computational representations reproduce perceptual effects such as the Wollaston illusion, when trained to estimate direction of gaze, but not when trained to recognize objects or faces. This work was supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF – 1231216. 2016-11-30T17:01:05Z 2016-11-30T17:01:05Z 2016-11-28 Technical Report Working Paper Other http://hdl.handle.net/1721.1/105477 arXiv:1611.09819 en_US CBMM Memo Series;059 Attribution-NonCommercial-ShareAlike 3.0 United States http://creativecommons.org/licenses/by-nc-sa/3.0/us/ application/pdf Center for Brains, Minds and Machines (CBMM), arXiv
spellingShingle gaze direction
social interaction
human vision
Harari, Daniel
Gao, Tao
Kanwisher, Nancy
Tenenbaum, Joshua
Ullman, Shimon
Measuring and modeling the perception of natural and unconstrained gaze in humans and machines
title Measuring and modeling the perception of natural and unconstrained gaze in humans and machines
title_full Measuring and modeling the perception of natural and unconstrained gaze in humans and machines
title_fullStr Measuring and modeling the perception of natural and unconstrained gaze in humans and machines
title_full_unstemmed Measuring and modeling the perception of natural and unconstrained gaze in humans and machines
title_short Measuring and modeling the perception of natural and unconstrained gaze in humans and machines
title_sort measuring and modeling the perception of natural and unconstrained gaze in humans and machines
topic gaze direction
social interaction
human vision
url http://hdl.handle.net/1721.1/105477
work_keys_str_mv AT hararidaniel measuringandmodelingtheperceptionofnaturalandunconstrainedgazeinhumansandmachines
AT gaotao measuringandmodelingtheperceptionofnaturalandunconstrainedgazeinhumansandmachines
AT kanwishernancy measuringandmodelingtheperceptionofnaturalandunconstrainedgazeinhumansandmachines
AT tenenbaumjoshua measuringandmodelingtheperceptionofnaturalandunconstrainedgazeinhumansandmachines
AT ullmanshimon measuringandmodelingtheperceptionofnaturalandunconstrainedgazeinhumansandmachines