OpenEyes: Eye Gaze in AR, VR, and in the Wild

ECCV 2020 Virtual Workshop


Sunday, August 23 2020
AM session: 8:00am - 11:00pm (UTC+1) [Recording of Zoom Session]
PM session: 7:00pm - 10:00pm (UTC+1) [Recording of Zoom Session]

Mini Webpage: workshopsandtutorials.eccv2020.eu

OpenEDS Challenges 2020: research.fb.com/programs/openeds-2020-challenge
EvalAI pages: Gaze Prediction  Sparse Semantic Segmentation

Playlist of Talks: Youtube


News and Updates

21st August We would like to ask all attendees to double check the times of the workshop items.
This is to avoid confusion due to the misinterpretation of timezones.

older news (hidden for brevity)

20th July We are happy to announce Prof. Thad Starner as a keynote speaker to our workshop!
28th June Due to time limitations and logistical restrictions, we have decided not to host extend abstracts at this year's OpenEyes.
4th June We are happy to announce as keynote speakers to our workshop!
30th May We have decided to extend the workshop paper submission deadline to Monday 15th June.
29th May We are happy to announce as keynote speakers to our workshop!
19th May ECCV is going virtual, and thus so is OpenEyes 2020. We will announce further details as they become available to us. In general, we will be run in the same fashion as the main conference and other ECCV workshops.
Image credits to Benh Lieu Song and Alessandro Caproni

Introduction

Full Text (hidden for brevity)

With the advent of consumer products, AR and VR as a form of immersive technology is gaining mainstream attention. However, immersive technology is still in its infancy, as both users and developers figure out the right recipe for the technology to garner mass appeal.

Eye tracking, a technology that measures where an individual is looking and can enable inference of user attention, could be a key driver of mass appeal for the next generation of immersive technologies, provided user awareness and privacy related to eye-tracking features are taken into account. As such, there is a growing interest in improving the state-of-the-art for eye tracking technology. In the past three years, investigations into gaze estimation and prediction methods produced significant improvements in robustness and accuracy by adopting increasingly unique deep neural network architectures. These improvements allow innovative applications for this technology, such as zero-shot image classification and generalized human attention and intent estimation.

Open forums of discussion provide opportunities to further improve eye tracking technology, especially in areas like scale and generalization challenges in the next generation of AR and VR systems. For that reason, Facebook organized the first challenge “Eye Tracking for VR and AR (OpenEDS)” at the ICCV 2019 and the independent GAZE committee organized a workshop titled “Gaze Estimation and Prediction in the Wild (GAZE)”.

For 2020, the Facebook and GAZE committees are partnering to host a joint workshop titled “Eye Gaze in VR, AR, and in the Wild” at the biennial ECCV conference. The workshop will host two tracks: the first focuses on gaze estimation and prediction methods, with a focus on accuracy and robustness in natural settings (in-the-wild); the second track focuses on the scale and generalization problem for eye tracking systems operating on AR and VR platforms. The second track also includes the 2020 eye tracking challenge. More information on the OpenEDS 2020 challenge can be found at https://research.fb.com/programs/openeds-2020-challenge/

The following topics are of particular interest to the joint workshop:

  • Proposal of novel eye detection, gaze estimation pipelines using deep neural networks that incorporate one or all of the following:
    • Geometric/anatomical constraints into the network in a differentiable manner.
    • Demonstration of robustness to conditions where current methods fail (illumination, appearance, low-resolution etc.).
    • Robust estimation from different data modalities such as RGB, depth, and near IR.
    • Use of additional cues, such as task context, temporal data, eye movement classification.
  • Designing new, accurate metrics to account for rapid eye movements in the real world.
  • Semi-/un-/self-supervised learning, meta-learning, domain adaptation, attention mechanisms and other related machine learning methods for gaze estimation.
  • Methods for temporal gaze estimation and prediction including Bayesian methods.
  • Unsupervised semantic segmentation of eye regions.
  • Active learning frameworks for semantic segmentation of eye images.
  • Generative models for eye image synthesis and gaze estimation.
  • Transfer learning for eye tracking from simulation data to real data.
  • Domain transfer applications for eye tracking.

This workshop will accept submissions of both published and unpublished works. We will also solicit high-quality eye tracking-related papers rejected at ECCV 2020, accompanied by the reviews and a letter of changes which clearly states the changes made to address comments by the previous reviewers. Accepted papers may be featured as spotlight talks and posters.


Call for Contributions


Full Workshop Papers

Submission: We invite authors to submit unpublished papers (14-page ECCV format) to our workshop, to be presented at a poster session upon acceptance. All submissions will go through a double-blind review process. All contributions must be submitted (along with supplementary materials, if any) at this OpenReview link.

Accepted papers will be published in the official ECCV Workshops proceedings and the Computer Vision Foundation (CVF) Open Access archive.

Note: Authors of previously rejected main conference submissions are also welcome to submit their work to our workshop. When doing so, you must submit the previous reviewers' comments (named as previous_reviews.pdf) and a letter of changes (named as letter_of_changes.pdf) as part of your supplementary materials to clearly demonstrate the changes made to address the comments made by previous reviewers. Due to potential clashes with the main conference reviewing schedule, we will accept simultaneous submissions to the ECCV main conference and OpenEyes Workshop. Simultaneous submissions are otherwise disallowed.


OpenEDS 2020 Challenges

The OpenEDS 2020 Challenges are hosted by the second track of the OpenEyes Workshop, and consist of two challenges:

  • Track 1: Gaze Prediction Challenge
  • Track 2: Sparse Temporal Semantic Segmentation Challenge

More information on the OpenEDS 2020 challenge can be found at:
https://research.fb.com/programs/openeds-2020-challenge/


Important Dates

Please note that all deadlines are at 11:59 PM (UTC -0), in line with the main conference.


Full Paper Submission Deadline Friday, 5th June 2020
Monday, 15th June 2020
Tuesday, 16th June 2020
Notification to Authors (Full Papers) Friday, 3rd July 2020
Wednesday, 8th July 2020
OpenEDS Challenges Participation Deadline Friday, 31st July 2020
Camera-Ready Deadline &
Presentation Materials Deadline
Saturday, 8th August 2020
Workshop Date Sunday, 23rd August 2020

Workshop Schedule


AM session (8:00 - 11:00 UTC+1) [Recording of Zoom Session]
Time in UTC+1 Start Time in UTC*
(probably your time zone)
Item
8:00am - 8:05am 23 Aug 2020 07:00:00 UTC Welcome and Opening Remarks
8:05am - 8:45am 23 Aug 2020 07:05:00 UTC Talk by Enkelejda Kasneci
8:45am - 9:25am 23 Aug 2020 07:45:00 UTC Talk by Mohamed Khamis
9:25am - 10:25am 23 Aug 2020 08:25:00 UTC Workshop Paper Talks
10:25am - 10:30am 23 Aug 2020 09:25:00 UTC Introduction to the OpenEDS 2020 Challenges
10:30am - 11:00am 23 Aug 2020 09:30:00 UTC Talks from the Challenge Winners

PM session (19:00 - 22:00 UTC+1) [Recording of Zoom Session]
Time in UTC+1 Start Time in UTC*
(probably your time zone)
Item
7:00pm - 7:40pm 23 Aug 2020 18:00:00 UTC Talk by Pramod Khargonekar
7:40pm - 8:20pm 23 Aug 2020 18:40:00 UTC Talk by Laurent Itti
8:20pm - 9:00pm 23 Aug 2020 19:20:00 UTC Talk by Qiang Ji
9:00pm - 9:40pm 23 Aug 2020 20:00:00 UTC Talk by Thad Starner
9:40pm - 9:50pm 23 Aug 2020 20:40:00 UTC Closing Remarks
* This time is calculated to be in your computer's reported time zone.
For example, those in Los Angeles may see UTC-7,
while those in Berlin may see UTC+2.

Please note that there may be differences to your actual time zone.

Invited Keynote Speakers

Morning Session
Enkelejda Kasneci
University of Tübingen

Eye Tracking: Ready to deliver the promises?


Biography (click to expand/collapse)

Enkelejda Kasneci is a Professor of Computer Science at the University of Tübingen, Germany, where she leads the Human-Computer Interaction Lab. As a BOSCH scholar, she received her M.Sc. degree in Computer Science from the University of Stuttgart in 2007. In 2013, she received her PhD in Computer Science from the University of Tübingen. For her PhD research, she was awarded the research prize of the Federation Südwestmetall in 2014. From 2013 to 2015, she was a postdoctoral researcher and a Margarete-von-Wrangell Fellow at the University of Tübingen. Her research evolves around the application of machine learning for intelligent and perceptual human-computer interaction. She served as academic editor for PlosOne and as a TPC member and reviewer for several major conferences and journals.

Mohamed Khamis
University of Glasgow

The interplay between Privacy, Security and Eye Tracking in the Wild: Opportunities and Challenges



Abstract & Biography (click to expand/collapse)
Abstract

Many of today’s smartphones, laptops and head-mounted displays are well equipped for accurate gaze estimation in the wild. These are important developments for security applications. For example, we will soon be able to run gaze-based security methods in the wild, such as authentication in daily scenarios, and analyze gaze behavior during security critical tasks, like when reading a phishing email. On the downside, ubiquitous eye tracking comes with privacy implications. Imagine if a neighbouring train passenger’s smartphone can track your eyes; this could potentially reveal a myriad of private information including visual interests, personality traits, mental disorders, emotional valence and more. In this talk, I will discuss how the role of eye gaze in security and privacy applications evolved in the past two decades, and how the current developments in ubiquitous eye tracking present a turning point that give rise to many opportunities and also new challenges.

Biography

Mohamed Khamis is a Lecturer (Assistant Professor) at the University of Glasgow in the UK where he leads a research team in Human-Centered Security. He received his PhD from Ludwig Maximilian University of Munich. His research is at the intersection of Ubiquitous Computing and User Privacy. He is interested in understanding the privacy implications of ubiquitous technologies, as well as developing novel systems for protecting privacy and security. He has major contributions at the intersection of security and gaze. Mohamed is a program committee member for CHI since 2019, an editorial board member of the Journal on Human Computer Studies. His work has been funded by the Royal Society of Edinburgh.

Afternoon Session
Pramod Khargonekar
University of Califonia, Irvine

Kalman Filtering, Sensor Fusion, and Eye Tracking



Biography (click to expand/collapse)

Pramod Khargonekar received B. Tech. Degree in electrical engineering in 1977 from the Indian Institute of Technology, Bombay, India, and M.S. degree in mathematics in 1980 and Ph.D. degree in electrical engineering in 1981 from the University of Florida, respectively. He has been on faculty at the University of Florida, University of Minnesota, The University of Michigan, and the University of California, Irvine. In June 2016, he assumed his current position as Vice Chancellor for Research and Distinguished Professor of Electrical Engineering and Computer Science at the University of California, Irvine. Khargonekar's current research and teaching interests include systems and control theory, machine learning, and applications to smart electric grid and neural engineering. He has authored more than 160 refereed journal publications and 200 conference publications, and has supervised 38 doctoral students. He has received numerous awards and have served in academia and industry in many roles, the details of which can be found on his webpage.

Laurent Itti
University of Southern California

Computational Modeling of Visual Attention for Virtual Reality



Biography (click to expand/collapse)

Laurent Itti received his M.S. degree in Image Processing from the Ecole Nationale Superieure des Telecommunications (Paris, France) in 1994, and his Ph.D. in Computation and Neural Systems from Caltech (Pasadena, CA) in 2000. He has since then been an Assistant, Associate, and now Full Professor of Computer Science, Psychology, and Neuroscience at USC. Dr. Itti’s research interests are in biologically-inspired computational vision, in particular in the domains of visual attention, scene understanding, control of eye movements, and surprise. This basic research has technological applications to, among others, video compression, target detection, and robotics. Dr. Itti has co-authored over 150 publications in peer-reviewed journals, books and conferences, three patents, and several open-source neuromorphic vision software toolkits.

Qiang Ji
Rensselaer Polytechnic Institute

A Bayesian Framework for Robust and Generalizable Eye Tracking



Abstract & Biography (click to expand/collapse)
Abstract

As a classical eye tracking approach, model-based eye tracking uses a geometric eye model to perform eye tracking; it generalizes well and requires no training data. Model-based eye tracking, however, requires accurate eye feature detection, which is hard to achieve in real world environments. Deep learning based methods have been employed recently to improve eye feature detection accuracy. These methods, however, do not generalize well to novel data. To achieve robust and generalizable eye tracking under different conditions, we propose a Bayesian framework to perform model-based eye tracking without explicit eye feature detection.

Specifically, the proposed method consists of a Bayesian Convolutional Neural Network (BCNN) to capture the probabilistic relationships between eye appearance and its landmarks, and a geometric model to estimate eye gaze from the eye landmarks. Under a Bayesian framework to integrate the two modules, given an eye image, the BCNN module outputs the probability distribution of the eye landmarks and their uncertainties, based on which the geometric model performs a Bayesian inference of the eye gaze by marginalizing out the eye landmarks, enabling eye gaze estimation without explicit eye landmark detection. Compared to the point-based eye landmark estimation methods, our model not only can generalize better, but is also more robust under challenging real world conditions. In addition, we extend the single-stage model to multi-stage, yielding the cascade BCNN. The cascade architecture allows feeding the uncertainty information from current stage to the next stage to progressively improve the gaze estimation accuracy. Experiments show the proposed method, while achieving comparable within-dataset performance, outperforms SoA methods in both it’s generalization capability across datasets, and in its robustness for low quality and challenging datasets.

Biography

Qiang Ji received his Ph.D degree in electrical engineering from the University of Washington. He is currently a Professor with the Department of Electrical, Computer, and Systems engineering at RPI. From January, 2009 to August, 2010, he served as a program director at the National Science Foundation, managing NSF's machine learning and computer vision programs. Prior to joining RPI in 2001, he was an assistant professor with Dept. of Computer Science, University of Nevada at Reno. He also held research and visiting positions with the Beckman Institute at University of Illinois at Urbana-Champaign, the Robotics Institute at Carnegie Mellon University, and the US Air Force Research Laboratory. Dr. Ji currently serves as the director of the Intelligent Systems Laboratory (ISL). Prof. Ji is a fellow of the IEEE and the IAPR.

Thad Starner
Georgia Tech

Eye Interfaces for Augmented Reality Eyeglasses



Abstract & Biography (click to expand/collapse)
Abstract

Eye interfaces should be reliable, scalable, consistent, discoverable, provide the user with feedback on performace, and be designed to indicate how they should be used. Combining these human computer interaction (HCI) fundamentals with machine learning and perception can point to useful and usable interfaces, in sometimes unexpected ways. This talk explores interfaces that have proved the most promising for augmented reality eyeglasses, focusing on eye gestures where tracking the relative motion of the eye is more important than absolute position. Such interfaces are surprisingly robust and have even been sensed using electrooculography (EOG) with three electrodes in the nose bridge of a Jins Meme.

Biography

Thad Starner is a Professor of Computing at Georgia Tech. He is also a Staff Research Scientist at Google, where he was one of the architects of Google Glass. In 1990, Starner coined the term "augmented reality" to describe the types of interfaces he envisioned for the future, and he has been using a head worn display as part of his daily life since 1993. He is a founder of the annual ACM International Symposium on Wearable Computers, now in its 24th year. Dr. Starner has produced over 500 papers and presentations on his research on interfaces and machine learning and has 95 issued United States utility patents. He was elected to the CHI Academy in 2017 and is always looking for a good game of table tennis.


Accepted Full Papers

Efficiency in Real-time Webcam Gaze Tracking Amogh Gudi, Xin Li, Jan van Gemert
PDF Talk
Hierarchical HMM for Eye Movement Classification Ye Zhu, Yan Yan, Oleg Komogortsev
PDF Talk
Domain Adaptation for Eye Segmentation Yiru Shen, Oleg Komogortsev, Sachin Talathi
PDF Talk
EyeSeg: Fast and Efficient Few-Shot Semantic Segmentation Jonathan Perry, Amanda Fernandez
PDF Talk

Also check out the papers from last year's GAZE 2019 and OpenEDS 2019

Awards

Best Paper Award sponsored by

Efficiency in Real-time Webcam Gaze Tracking
Amogh Gudi, Xin Li, Jan van Gemert

We would like to thank for awarding an eye tracker device for each accepted full paper.
We also congratulate the winners of the OpenEDS 2020 challenge hosted by .

Click here for the challenge website, and here for the video announcing the winners.

Gaze Prediction Challenge
First Place Team: Random_B
Members: Anton Repushko, Skoblov Svyatosiav, Vladimir Mikheyushkin
Country: Russia
Gaze Prediction Error: 3.17 Degree

Second Place Team: Eyemazing
Members: Vivek Mittal
Country: India
Gaze Prediction Error: 3.31 Degree

Third Place Team: DmitryAKonovalov
Members: Dmitry Konolov
Country: Australia
Gaze Prediction Error: 3.34 Degree

Sparse Semantic Segmentation Challenge
First Place Team: BTS Digital
Members: Almaz Zinollayev, Zhanseri Ikram
Country: Kazhakastan
mIOU: 0.952

Second Place Team: Tetelias
Members: Teternikov Ilia Anatolyevich
Country: Russia
mIOU: 0.951

Third Place Team: JHU
Members: Daniil Pakhomov, Sanchit Hira
Country: USA
mIOU: 0.950


Organizers


Gaze Estimation and Prediction in the Wild

Hyung Jin Chang
University of Birmingham
Seonwook Park
ETH Zürich
Xucong Zhang
ETH Zürich
Otmar Hilliges
ETH Zürich
Aleš Leonardis
University of Birmingham

Eye Tracking for VR and AR

Robert Cavin
Facebook Reality Labs
Cristina Palmero
Universitat de Barcelona (UB)
Jixu Chen
Facebook
Alexander Fix
Facebook Reality Labs
Elias Guestrin
Facebook Reality Labs
Oleg Komogortsev
Texas State University

Kapil Krishnakumar
Facebook
Abhishek Sharma
Facebook Reality Labs
Yiru Shen
Facebook Reality Labs
Tarek Hefny
Facebook Reality Labs
Karsten Behrendt
Facebook
Sachin S. Talathi
Facebook Reality Labs

Program Committee

We thank our reviewers for their time and care in reviewing the submissions to OpenEyes 2020.

Mihai Bâce
ETH Zurich
Marcel Bühler
ETH Zurich
Hyung Jin Chang
University of Birmingham
Jixu Chen
Facebook
Eunji Chong
Georgia Tech
Tobias Fischer
Queensland University of Technology
Alexander Fix
Facebook Reality Labs
Otmar Hilliges
ETH Zürich
Nora Horanyi
University of Birmingham
Oleg Komogortsev
Texas State University
Aleš Leonardis
University of Birmingham
Seonwook Park
ETH Zürich
Abhishek Sharma
Facebook Reality Labs
Yiru Shen
Facebook Reality Labs
Sachin Talathi
Facebook Reality Labs
Xucong Zhang
ETH Zürich
Yufeng Zheng
ETH Zürich

Workshop sponsored by:


Gaze Estimation and Prediction in the Wild


Eye Tracking for VR and AR