• Stars
    star
    103
  • Rank 333,046 (Top 7 %)
  • Language
    Python
  • License
    MIT License
  • Created about 6 years ago
  • Updated over 3 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Code for ACCV2018 paper 'Believe It or Not, We Know What You Are Looking at!'

Gaze following

PyTorch implementation of our ACCV2018 paper:

'Believe It or Not, We Know What You Are Looking at!' [paper] [poster]

Dongze Lian*, Zehao Yu*, Shenghua Gao

(* Equal Contribution)

Prepare training data

GazeFollow dataset is proposed in [1], please download the dataset from http://gazefollow.csail.mit.edu/download.html. Note that the downloaded testing data may have wrong label, so we request test2 provided by author. I do not know whether the author update their testing set. If not, it is better for you to e-mail authors in [1]. For your convenience, we also paste the testing set link here provided by authors in [1] when we request. (Note that the license is in [1])

Download our dataset

OurData is in Onedrive Please download and unzip it

OurData contains data descriped in our paper.

OurData/tools/extract_frame.py

extract frame from clipVideo in 2fps. Different version of ffmpeg may have different results, we provide our extracted images.

OurData/tools/create_video_image_list.py

extract annotation to json.

Testing on gazefollow data

Please download the pretrained model manually and save to model/

cd code
python test_gazefollow.py

Evaluation metrics

cd code
python cal_min_dis.py
python cal_auc.py

Test on our data

cd code
python test_ourdata.py

Training scratch

cd code
python train.py

Inference

simply run python inference.py image_path eye_x eye_y to infer the gaze. Note that eye_x and eye_y is the normalized coordinate (from 0 - 1) for eye position. The script will save the inference result in tmp.png.

cd code
python inference.py ../images/00000003.jpg 0.52 0.14

Reference:

[1] Recasens*, A., Khosla*, A., Vondrick, C., Torralba, A.: Where are they looking? In: Advances in Neural Information Processing Systems (NIPS) (2015).

Citation

If this project is helpful for you, you can cite our paper:

@InProceedings{Lian_2018_ACCV,
author = {Lian, Dongze and Yu, Zehao and Gao, Shenghua},
title = {Believe It or Not, We Know What You Are Looking at!},
booktitle = {ACCV},
year = {2018}
}

More Repositories

1

impersonator

PyTorch implementation of our ICCV 2019 paper: Liquid Warping GAN: A Unified Framework for Human Motion Imitation, Appearance Transfer and Novel View Synthesis
Python
1,725
star
2

PlanarReconstruction

[CVPR'19] Single-Image Piece-wise Planar 3D Reconstruction via Associative Embedding
Python
362
star
3

FastMVSNet

[CVPR'20] Fast-MVSNet: Sparse-to-Dense Multi-View Stereo With Learned Propagation and Gauss-Newton Refinement
Python
253
star
4

PPGNet

Source code for our CVPR 2019 paper - PPGNet: Learning Point-Pair Graph for Line Segment Detection
Python
173
star
5

Indoor-SfMLearner

[ECCV'20] Patch-match and Plane-regularization for Unsupervised Indoor Depth Estimation
Python
151
star
6

HRNet-for-Fashion-Landmark-Estimation.PyTorch

[DeepFashion2 Challenge] Fashion Landmark Estimation with HRNet
Cuda
129
star
7

AS-MLP

[ICLR'22] This is an official implementation for "AS-MLP: An Axial Shifted MLP Architecture for Vision".
Python
124
star
8

PlaneDepth

[CVPR2023] This is an official implementation for "PlaneDepth: Self-supervised Depth Estimation via Orthogonal Planes".
Python
100
star
9

CIDNN

CIDNN: Encoding Crowd Interaction with Deep Neural Network
Python
74
star
10

IVOS-W

[CVPR'21] Learning to Recommend Frame for Interactive Video Object Segmentation in the Wild
Python
49
star
11

MLEP

Python
48
star
12

LBYLNet

[CVPR2021] Look before you leap: learning landmark features for one-stage visual grounding.
Python
47
star
13

Weekly_Group_Meeting_Paper_List

42
star
14

RGBD-Counting

RGBD crowd counting
Python
38
star
15

WeakSVR

(CVPR 2023) Official implemention of the paper "Weakly Supervised Video Representation Learning with Unaligned Text for Sequential Videos"
Python
27
star
16

Locating_Counting_with_a_Depth_Prior

[TPAMI] Locating and Counting Heads in Crowds With a Depth Prior
Python
25
star
17

RGBD-Gaze

RGBD Based Gaze Estimation via Multi-task CNN
Python
22
star
18

SVIP-Sequence-VerIfication-for-Procedures-in-Videos

[CVPR2022] SVIP: Sequence VerIfication for Procedures in Videos
Python
19
star
19

ShanghaiTechRGBDSyn

[TPAMI] Locating and Counting Heads in Crowds With a Depth Prior
C++
11
star
20

Medical-Image-CodeBase-SVIP-Lab

Useful and frequently used code for computer vision
Python
9
star
21

Saliency-Detection-in-360-Videos

Saliency-Detection-in-360-Videos
Python
8
star
22

svip-lab.github.io

HTML
8
star
23

SvipLab-ChatGPT-Web-Share

7
star
24

CrowdCountingPAL

Python
7
star
25

SphericalDNNs

Python
6
star