Skip to content

The objective is to take as data input the images of the road, the bounding boxes of the surrounding objects, provided by IBEO, and the gaze data provided by a gaze tracker, and to output the classes of the objects looked at in real time

Notifications You must be signed in to change notification settings

Raffael1998/Semantic-Gaze-Tracking

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

53 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Semantic Gaze Tracking

This repository contains an explaination and demonstration of a proof of concept we created to merge signals from the eye tracker and the sensors seeing the vehicle environment. The data used is provided by :

  • A camera, on which we apply a semantic segmentation specific to road objects and environment
  • A gaze tracker, that provides information about the head position and the gaze direction
  • An advanced lidar system (here IBEO), that provides position and size of bounding boxes of the surrounding objects
  • Another lidar (here Velodyne), that provides a dense point cloud of the surroundings

The image segmentation used is inspired by the CSAILVision Github.
This project was conducted during a 5 months intership in the Centre of Accident Research and Road Safety (CARRS-Q) of the Queensland University of Technology (Australia).

Context :

The problem

  • Driver’s distraction
  • Level 2 of vehicle autonomy : the driver must maintain their vigilance on the road
  • Level 3 : the driver is the fallback solution if the automation understands that it will fail to operate in a near future
  • Level 4 : the automation needs to understand the readiness of the driver if they decide to take over
  • Current driver monitoring systems do not analyse the driver visual attention with respect to the environment Proposed solution
  • Analyse the road and the driver’s focus to infer the danger rating of the situation

Targets

  • Identify the position of the objects looked at by the driver
  • Identify the class of the objects looked at Problems currently
  • The gaze tracker gives information about the direction of the gaze but the information about the distance of the focus point is not precise enough (intersection of the 2 gaze lines)
  • In order to compute the distance of the objects, we need a 3D representation of the environment
  • Solution 1 : Stereo depth map images (we need a stereo camera system or a machine learning solution to recreate this image)
  • Solution 2 : Using a lidar (IBEO / Lidar)
  • Finding the class of the objects

Processing Pipeline :

image

About

The objective is to take as data input the images of the road, the bounding boxes of the surrounding objects, provided by IBEO, and the gaze data provided by a gaze tracker, and to output the classes of the objects looked at in real time

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages