Eye Tracking for Everyone

From scientific research to commercial applications, eye tracking is an important tool across many domains. Despite its range of applications, eye tracking has yet to become a pervasive technology. We believe that we can put the power of eye tracking in everyone's palm by building eye tracking...

Full description

Bibliographic Details
Main Authors: Kellnhofer, Petr, Bhandarkar, Suchendra, Khosla, Aditya, Kannan, Harini D., Matusik, Wojciech, Torralba, Antonio
Other Authors: Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
Format: Article
Language:en_US
Published: Institute of Electrical and Electronics Engineers (IEEE) 2017
Online Access:http://hdl.handle.net/1721.1/111782
https://orcid.org/0000-0002-0007-3352
https://orcid.org/0000-0003-1462-2313
https://orcid.org/0000-0003-0212-5643
https://orcid.org/0000-0003-4915-0256
_version_ 1826202106361544704
author Kellnhofer, Petr
Bhandarkar, Suchendra
Khosla, Aditya
Kannan, Harini D.
Matusik, Wojciech
Torralba, Antonio
author2 Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
author_facet Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
Kellnhofer, Petr
Bhandarkar, Suchendra
Khosla, Aditya
Kannan, Harini D.
Matusik, Wojciech
Torralba, Antonio
author_sort Kellnhofer, Petr
collection MIT
description From scientific research to commercial applications, eye tracking is an important tool across many domains. Despite its range of applications, eye tracking has yet to become a pervasive technology. We believe that we can put the power of eye tracking in everyone's palm by building eye tracking software that works on commodity hardware such as mobile phones and tablets, without the need for additional sensors or devices. We tackle this problem by introducing GazeCapture, the first large-scale dataset for eye tracking, containing data from over 1450 people consisting of almost 2:5M frames. Using GazeCapture, we train iTracker, a convolutional neural network for eye tracking, which achieves a significant reduction in error over previous approaches while running in real time (10-15fps) on a modern mobile device. Our model achieves a prediction error of 1.71cm and 2.53cm without calibration on mobile phones and tablets respectively. With calibration, this is reduced to 1.34cm and 2.12cm. Further, we demonstrate that the features learned by iTracker generalize well to other datasets, achieving state-of-the-art results.
first_indexed 2024-09-23T12:01:58Z
format Article
id mit-1721.1/111782
institution Massachusetts Institute of Technology
language en_US
last_indexed 2024-09-23T12:01:58Z
publishDate 2017
publisher Institute of Electrical and Electronics Engineers (IEEE)
record_format dspace
spelling mit-1721.1/1117822022-10-01T07:44:38Z Eye Tracking for Everyone Kellnhofer, Petr Bhandarkar, Suchendra Khosla, Aditya Kannan, Harini D. Matusik, Wojciech Torralba, Antonio Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology. Media Laboratory Khosla, Aditya Kellnhofer, Petr Kannan, Harini D. Matusik, Wojciech Torralba, Antonio From scientific research to commercial applications, eye tracking is an important tool across many domains. Despite its range of applications, eye tracking has yet to become a pervasive technology. We believe that we can put the power of eye tracking in everyone's palm by building eye tracking software that works on commodity hardware such as mobile phones and tablets, without the need for additional sensors or devices. We tackle this problem by introducing GazeCapture, the first large-scale dataset for eye tracking, containing data from over 1450 people consisting of almost 2:5M frames. Using GazeCapture, we train iTracker, a convolutional neural network for eye tracking, which achieves a significant reduction in error over previous approaches while running in real time (10-15fps) on a modern mobile device. Our model achieves a prediction error of 1.71cm and 2.53cm without calibration on mobile phones and tablets respectively. With calibration, this is reduced to 1.34cm and 2.12cm. Further, we demonstrate that the features learned by iTracker generalize well to other datasets, achieving state-of-the-art results. 2017-10-04T15:28:49Z 2017-10-04T15:28:49Z 2016-12 Article http://purl.org/eprint/type/ConferencePaper 978-1-4673-8851-1 1063-6919 http://hdl.handle.net/1721.1/111782 Krafka, Kyle et al. “Eye Tracking for Everyone.” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 27-30 2016, Las Vegas, Neveda, USA, Institute of Electrical and Electronics Engineers (IEEE), December 2016: 2176-2184 © 2016 Institute of Electrical and Electronics Engineers (IEEE) https://orcid.org/0000-0002-0007-3352 https://orcid.org/0000-0003-1462-2313 https://orcid.org/0000-0003-0212-5643 https://orcid.org/0000-0003-4915-0256 en_US http://dx.doi.org/10.1109/CVPR.2016.239 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Creative Commons Attribution-Noncommercial-Share Alike http://creativecommons.org/licenses/by-nc-sa/4.0/ application/pdf Institute of Electrical and Electronics Engineers (IEEE) MIT Web Domain
spellingShingle Kellnhofer, Petr
Bhandarkar, Suchendra
Khosla, Aditya
Kannan, Harini D.
Matusik, Wojciech
Torralba, Antonio
Eye Tracking for Everyone
title Eye Tracking for Everyone
title_full Eye Tracking for Everyone
title_fullStr Eye Tracking for Everyone
title_full_unstemmed Eye Tracking for Everyone
title_short Eye Tracking for Everyone
title_sort eye tracking for everyone
url http://hdl.handle.net/1721.1/111782
https://orcid.org/0000-0002-0007-3352
https://orcid.org/0000-0003-1462-2313
https://orcid.org/0000-0003-0212-5643
https://orcid.org/0000-0003-4915-0256
work_keys_str_mv AT kellnhoferpetr eyetrackingforeveryone
AT bhandarkarsuchendra eyetrackingforeveryone
AT khoslaaditya eyetrackingforeveryone
AT kannanharinid eyetrackingforeveryone
AT matusikwojciech eyetrackingforeveryone
AT torralbaantonio eyetrackingforeveryone