• Stars
    star
    107
  • Rank 323,587 (Top 7 %)
  • Language
    Shell
  • License
    Apache License 2.0
  • Created over 4 years ago
  • Updated about 4 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Code for “Disentangled Non-local Neural Networks”

Disentangled Non-Local Neural Networks

By Minghao Yin, Zhuliang Yao, Yue Cao, Xiu Li, Zheng Zhang, Stephen Lin, Han Hu.

We provide related codes and configuration files to reproduce the "Disentangled Non-Local Neural Networks" on Cityscapes and ADE20K semantic segmentation. Our code is adapted from torchcv.

Introduction

The non-local block models long-range dependency in many computer vision tasks. Our paper first studies the non-local block in depth, where we find that its attention computation can be split into pairwise term and unary term. However, the two terms are tightly coupled in the non-local block, which hinders the learning of each. Based on that, we developed the disentangled non-local block, where the two terms are decoupled to facilitate learning for both terms. Disentangled non-local block outperforms non-local block in various tasks.

The disentangled Non-local block composes of a whitened pairwise term accounting for the relationship between two pixels and a unary term representing the saliency of every pixel. The two terms are clearly decoupled by using independent Softmax functions and embedding matrices. As shown in the attention maps, the whitened pairwise term learns clear within-region clues while the unary term learns salient boundaries.

Citation

@article{yin2020disentangled,
  title={Disentangled Non-Local Neural Networks},
  author={Yin, Minghao and Yao, Zhuliang and Cao, Yue and Li, Xiu and Zhang, Zheng and Lin, Stephen and Hu, Han},
  journal={arXiv preprint arXiv:2006.06668},
  year={2020}
}

Install

Requirements

  • Linux or macOS (Windows is not currently officially supported)
  • Python 3.6+
  • PyTorch 1.0+
  • CUDA 9.0+

a. Create a conda virtual environment and activate it.

conda create -n dnlnet python=3.6 -y
conda activate dnlnet

b. Install PyTorch and torchvision following the official instructions, e.g.,

conda install pytorch torchvision -c pytorch

c. Clone the repository.

git clone https://github.com/yinmh17/DNL-Networks.git
cd DNL-Networks

d.Install build requirements

pip3 install -r requirements.txt
cd exts
sh make.sh

Train

Before training, you need to preprocess your datasets following the instruction in datasets/seg/preprocess. Download ImageNet pretrained model 3x3-Resnet101

You can use following commands to train your model.

bash ./scripts/seg/{DATASET}/{SCRIPT_FILE_NAME}.sh train {TAG} {DATA_DIR}  {PRETRAINED_MODEL} 

Optional arguments:

  • DATASET: Name of the dataset, for example: "cityscapes", "ade20k".
  • SCRIPT_FILE_NAME: Name of the bash file you are going to run. Training settings are contained in the bash file.
  • DATA_DIRECTION: Directory of the preprocessed dataset.
  • PRETRAINED_MODEL: Directory of your pretrained model.

Take DNL network training on Cityscapes dataset for example, you need to run the following commands.

bash ./scripts/seg/cityscapes/run_fs_res101_nonlocalnowd_ln_cityscapes_seg.sh train tag  /torchcv/data/cityscapes   ./pretrained_models/3x3resnet101-imagenet.pth 

Inference

After training, you can use following commands to validate your model.

bash ./scripts/seg/{DATASET}/{SCRIPT_FILE_NAME}.sh val {TAG} {DATA_DIR} 

Take DNL network validation on Cityscapes dataset for example:

bash ./scripts/seg/cityscapes/run_fs_res101_nonlocalnowd_ln_cityscapes_seg.sh val tag  /torchcv/data/cityscapes  

Models will come soon!!