Skip to content

A human-computer interaction system that combines eye tracking with Segment Anything Model (SAM), and it enables users to segment object they are looking at in real-time. [NeurIPSW 2023]

ukaukaaaa/GazeSAM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GazeSAM: What You See is What You Segment

Bin Wang, Armstrong Aboah, Zheyuan Zhang, Ulas Bagci

[Paper] [Demo (Youtube)] [Demo (bilibili)] [BibTeX]

framework

The GazeSAM is a human-computer interaction system that combines eye tracking technology with Segment Anything Model (SAM), and it enables users to segment object they are looking at in real-time. This system is designed specifically for the radiologist to record segmentation mask during image reading by simply looking at the desired regions, which can boost the daily clinical workflow. Besides, eye-gaze data can also be easily recorded for the further eye-tracking research. This system supports both 2D and 3D images.

GIF 1 GIF 2
GIF 1 GIF 2

Experiment Setting

A user interface is provided as shown in the left image below and experiment setting is illustrated as the right image below.

Installation

This code requires requires python=3.8.0, as well as pytorch>=1.7 and torchvision>=0.8. For your convenience, we already install the package of segment-anything in our repo but you still need to install some dependencies.

pip install opencv-python pycocotools matplotlib onnxruntime onnx

Besides, you need to download the model checkpoint provided by SAM as follows

And put the checkpoint under "./model/".

Getting Started

In this work, we use Tobii Pro Nano as the eye tracker. If you have the device and want to repeat our result. Here is some steps you need to do.

First, download Tobii Pro Eye Tracker Manager, then open the open it to install one Tobii Pro Nano device in your PC. After that, finish the calibration procedure to make sure the eye movement is recorded accurately.

Second, install some dependencies

pip install tobii-research, PyQt5, SimpleITK

Run the user interface by

python ui.py

Visualization

Here is few examples of GazeSAM.

performance

Citing

@article{wang2023gazesam,
  title={GazeSAM: What You See is What You Segment},
  author={Wang, Bin and Aboah, Armstrong and Zhang, Zheyuan and Bagci, Ulas},
  journal={arXiv preprint arXiv:2304.13844},
  year={2023}
}

About

A human-computer interaction system that combines eye tracking with Segment Anything Model (SAM), and it enables users to segment object they are looking at in real-time. [NeurIPSW 2023]

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages