PhD Defense by Eunji Chong

Event Details
  • Date/Time:
    • Thursday January 9, 2020
      3:30 pm - 5:30 pm
  • Location: TSRB 222
  • Phone:
  • URL:
  • Email:
  • Fee(s):
    N/A
  • Extras:
Contact
No contact information submitted.
Summaries

Summary Sentence: Computational Methods for Measurement of Visual Attention from Videos towards Large-scale Behavioral Analysis

Full Summary: No summary paragraph submitted.

Title: Computational Methods for Measurement of Visual Attention from Videos towards Large-scale Behavioral Analysis

 

Eunji Chong

School of Computer Science

College of Computing

Georgia Institute of Technology

 

Date:  Thursday, January 9th, 2020

Time: 3:30 - 5:30 PM (EST)

Location: TSRB 222

 

Committee:

Dr. James M. Rehg (Advisor), School of Computer Science, Georgia Institute of Technology

Dr. Agata Rozga, School of Computer Science, Georgia Institute of Technology

Dr. Gregory D. Abowd, School of Computer Science, Georgia Institute of Technology

Dr. Irfan Essa, School of Computer Science, Georgia Institute of Technology

Dr. Yaser Sheikh, Robotics Institute, Carnegie Mellon University

 

Abstract:

Visual attention is a critically-important aspect of human social behavior, visual navigation, and interaction with the 3D environment, and where and what people are paying attention to reveals a lot of information about their social, cognitive, and affective states. While monitor-based and wearable eye trackers are widely available, they are not sufficient to support the large-scale collection of naturalistic gaze data in contexts such as face-to-face social interactions or object manipulation in 3D environments. Wearable eye trackers are burdensome to participants and bring issues of calibration, compliance, cost, and battery life.

 

This thesis investigates different ways to measure real-world human visual attention using computer vision from plain videos and its use for identifying meaningful social behaviors. Specifically, three methods are investigated. First, I present methods for detection of looks to camera in first-person view and its use for eye contact detection. Experimental results show that the presented method can achieve the first human expert-level detection of eye contact. Second, I develop a method for tracking heads in a 3d space for measuring attentional shifts. Lastly, I propose spatiotemporal deep neural networks for detecting time-varying attention targets in video and present its application for the detection of shared attention and joint attention. The final method achieves state-of-the-art results on different benchmark datasets on attention measurement as well as the first empirical result on clinically-relevant gaze shift classification.

 

Presented approaches have the benefit of linking gaze estimation to the broader tasks of action recognition and dynamic visual scene understanding, and bears potential as a useful tool for understanding attention in various contexts such as human social interactions, skill assessments, and human-robot interactions.

 

Additional Information

In Campus Calendar
No
Groups

Graduate Studies

Invited Audience
Faculty/Staff, Public, Graduate students, Undergraduate students
Categories
Other/Miscellaneous
Keywords
Phd Defense
Status
  • Created By: Tatianna Richardson
  • Workflow Status: Published
  • Created On: Jan 2, 2020 - 2:02pm
  • Last Updated: Jan 2, 2020 - 2:02pm